Skip to main content

Resolved Issues

View the resolved issues for a specific Platform release.

Go to download resolved issues by patch release.

Browse release notes for a selected Pega Version.

NOTE: Enter just the Case ID number (SR or INC) in order to find the associated Support Request.

Please note: beginning with the Pega Platform 8.7.4 Patch, the Resolved Issues have moved to the Support Center.

SR-C74784 · Issue 428820

Added file truncation to HCClient.txt to ensure data integrity

Resolved in Pega Version 8.1.4

When a work object model contained more than two embedded lists, simple save corrupted the JSON in the local storage work object browser. Investigation traced the issue to using the same file without truncating its content. Because the HCClient.txt file in Rule: webwb - pzpega_mobile_settings - js was reused while displaying items in DocumentViewer, writing a smaller amount of data into it than the previous write left orphaned content in the file that corrupted the new data. To resolve this, the file content will be truncated before writing new data into it.

INC-211655 · Issue 712350

Added handling for Malaysian locale date/time

Resolved in Pega Version 8.6.4

When using the "ms_MY" locale for Malaysia, entering ‘14/02/2022 13:00’ in the displayed DateTime input field generated the error "14/02/2022 13:00 is not a valid date / time value." This was due to differences in the underlying Java version: in Java 8, the API PRDateFormat.getShortDateTimeFormat returns "dd/MM/yyyy h:mm" and the PRDateFormat.getAmPmStrings returns [AM, PM], but in Java 11, the API returns "d/MM/yyyy h:mm a" with PRDateFormat.getAmPmStrings returning [PG, PTG]. For the Malaysian locale ms_MY, the clock format is 12 hour and AM/PM Strings are PG, PTG. This caused environments running JDK11 to fail client side validation for date time when PTG (a 3-char AM/PM string) was selected. This has been resolved by adding handling for this usecase.

INC-138354 · Issue 594353

Third party cookies reminder added for Mashup in incognito browsers

Resolved in Pega Version 8.5.2

When working in incognito/private browser windows, attempting to use Mashup displayed a message indicating login credentials were not recognized ("Unknown password and/or username"). This was due to recent changes originating with the browsers regarding blocking third-party cookies which are necessary for Mashup to run. For informational purposes, a new API has been added which will check for the PegaRules cookie; if the cookie is not present, the message "Third party cookies must be enabled for Mashup to function." will be displayed. This message may be customized by configuring the setting "pega.web.config.xCookiesDisabled" with the required message.

INC-136563 · Issue 607418

Tenant Switching updated

Resolved in Pega Version 8.5.2

When working on new “tenant switching” implementations in a multi-tenant environment, the API itself seemed to be working as expected (case details from another tenant were shown correctly in review mode), but clicking on the assignment unexpectedly opened the case in a new window that contained the same review view of the case rather than the perform view. Clicking the assignment link again caused yet another review window to open with the same issue, over and over. This has been resolved by adding a UI-side update to send the requests on proper sub-domain URL instead of using the parent domain, and by ensuring the document threads are properly deleted for multitenant usecases.

SR-C93079 · Issue 430562

Action queue explicitly cleared in closeAllDocuments method

Resolved in Pega Version 8.1.4

After logging in to the app studio and performing some changes in the case designer, attempting to switch to data explorer or other tabs generated an alert indicating "Remaining tabs have unsaved work". After the alert was generated and closed, a "Maximum limit of opens tabs is reached" popup was then displayed even when there were no open tabs. This was traced to a change in the event handler for "DesktopAction" which caused the system to immediately move from the alert to triggering createNewWork, which then used the same dynamic document as the original alert. To resolve this, the system will clear the action queue pega.c.actionSequencer.clearQueue() in the closeAllDocuments method.

SR-C86952 · Issue 428817

Viewing screen capture attchments no longer requires IE/Silverlight

Resolved in Pega Version 8.1.4

Previous versions of the Pega Platform required the use of the IE browser with Silverlight to access old attachments added using the screen capture feature. In order to support multibrowser compatibility, the system has been updated to use javascript functions to display these images so IE/Silverlight are not required.

SR-C75099 · Issue 413433

Improvements made for the stability and reliability of the stream tier

Resolved in Pega Version 8.1.4

Several enhancements and fixes have been made to improve the stability and reliability of the stream tier with a focus on rolling restart. Issue 410567: When a stream node shuts down, it takes up to a minute for all client nodes to recognize the event and failover to other nodes due to large default timeouts. Timeouts have now been reduced to allow nodes to failover much earlier. Issue 413375: When a stream node shuts down, other stream nodes are notified and attempt to discover the new cluster topology. Since the cluster is still in flux, this operation may take a long time blocking all producers until a new topology is determined. This procedure has been changed to be less heavy and make the discovery only if required. Issue 401326: Kafka (running as a separate process as part of the Stream services) generates log files, but does not use log rotation with a maximum amount of log files. This means that the log files will eventually fill up the disk, leading to failures of the service and PRPC instance running on the same disk. To resolve this, the system will now limit number of generated log files to prevent high disk usage by using the parameters of file size = 2 MB and a max nr of size = 10. Issue 410280: A large number of ""Processing override for entityPath"" entries were seen in the Kafka logs. This has been addressed by ensuring message.timestamp.difference.max.ms is set to equal to retention.ms. In addition, the system will properly compare the Kafka Configuration created for the entire topic and the configuration retrieved for Partitions so the configuration is not unnecessarily updated. Issue 414765: During stream node restart, decision hub nodes may block for 15-30 seconds when writing to a stream data under certain conditions. This can cause high response times for a small period. This has been addressed by updating stream dataset logic to handle stream tier restarts properly and not cause response time degradation. Issue 401093: HttpMessageDecoder was throwing an exception regarding “TooLongFrameException” causing a 502 on HTTP stream tier requests going over 8192. The default max header size used by stream data set REST service was 8KB. In this particular case, the HTTP header contains a lot of cookies and other meta-info and exceeds 8KB. To resolve this, the default max header size has been increased to 16KB. A new prconfig setting to control max header size has also been introduced.

SR-C77625 · Issue 414766

Improvements made for the stability and reliability of the stream tier

Resolved in Pega Version 8.1.4

Several enhancements and fixes have been made to improve the stability and reliability of the stream tier with a focus on rolling restart. Issue 410567: When a stream node shuts down, it takes up to a minute for all client nodes to recognize the event and failover to other nodes due to large default timeouts. Timeouts have now been reduced to allow nodes to failover much earlier. Issue 413375: When a stream node shuts down, other stream nodes are notified and attempt to discover the new cluster topology. Since the cluster is still in flux, this operation may take a long time blocking all producers until a new topology is determined. This procedure has been changed to be less heavy and make the discovery only if required. Issue 401326: Kafka (running as a separate process as part of the Stream services) generates log files, but does not use log rotation with a maximum amount of log files. This means that the log files will eventually fill up the disk, leading to failures of the service and PRPC instance running on the same disk. To resolve this, the system will now limit number of generated log files to prevent high disk usage by using the parameters of file size = 2 MB and a max nr of size = 10. Issue 410280: A large number of "Processing override for entityPath" entries were seen in the Kafka logs. This has been addressed by ensuring message.timestamp.difference.max.ms is set to equal to retention.ms. In addition, the system will properly compare the Kafka Configuration created for the entire topic and the configuration retrieved for Partitions so the configuration is not unnecessarily updated. Issue 414765: During stream node restart, decision hub nodes may block for 15-30 seconds when writing to a stream data under certain conditions. This can cause high response times for a small period. This has been addressed by updating stream dataset logic to handle stream tier restarts properly and not cause response time degradation. Issue 401093: HttpMessageDecoder was throwing an exception regarding “TooLongFrameException” causing a 502 on HTTP stream tier requests going over 8192. The default max header size used by stream data set REST service was 8KB. In this particular case, the HTTP header contains a lot of cookies and other meta-info and exceeds 8KB. To resolve this, the default max header size has been increased to 16KB. A new prconfig setting to control max header size has also been introduced.

SR-C85146 · Issue 423597

Improvements made for the stability and reliability of the stream tier

Resolved in Pega Version 8.1.4

Several enhancements and fixes have been made to improve the stability and reliability of the stream tier with a focus on rolling restart. Issue 410567: When a stream node shuts down, it takes up to a minute for all client nodes to recognize the event and failover to other nodes due to large default timeouts. Timeouts have now been reduced to allow nodes to failover much earlier. Issue 413375: When a stream node shuts down, other stream nodes are notified and attempt to discover the new cluster topology. Since the cluster is still in flux, this operation may take a long time blocking all producers until a new topology is determined. This procedure has been changed to be less heavy and make the discovery only if required. Issue 401326: Kafka (running as a separate process as part of the Stream services) generates log files, but does not use log rotation with a maximum amount of log files. This means that the log files will eventually fill up the disk, leading to failures of the service and PRPC instance running on the same disk. To resolve this, the system will now limit number of generated log files to prevent high disk usage by using the parameters of file size = 2 MB and a max nr of size = 10. Issue 410280: A large number of ""Processing override for entityPath"" entries were seen in the Kafka logs. This has been addressed by ensuring message.timestamp.difference.max.ms is set to equal to retention.ms. In addition, the system will properly compare the Kafka Configuration created for the entire topic and the configuration retrieved for Partitions so the configuration is not unnecessarily updated. Issue 414765: During stream node restart, decision hub nodes may block for 15-30 seconds when writing to a stream data under certain conditions. This can cause high response times for a small period. This has been addressed by updating stream dataset logic to handle stream tier restarts properly and not cause response time degradation. Issue 401093: HttpMessageDecoder was throwing an exception regarding “TooLongFrameException” causing a 502 on HTTP stream tier requests going over 8192. The default max header size used by stream data set REST service was 8KB. In this particular case, the HTTP header contains a lot of cookies and other meta-info and exceeds 8KB. To resolve this, the default max header size has been increased to 16KB. A new prconfig setting to control max header size has also been introduced.

SR-C61477 · Issue 418266

HTML conversion added to email IVA

Resolved in Pega Version 8.1.4

When sending through emails that are in HTML format, HTML tags were visible within the Analysis of the email under Entities. These tags also potentially had an impact on topic detection, as the same email sent through in HTML would be classified under a different topic from one sent through as Plain Text.This issue originated because Email IVA received text as HTML while it expected plain text without HTML tags, and was traced to the use of an IMAP setting which sends only HTML to the listener. This has now been fixed such that If any HTML is received by email IVA, it will be converted to plain text via JSoup APIs in the service method as the first step.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us