Skip to main content

Resolved Issues

View the resolved issues for a specific Platform release.

Go to download resolved issues by patch release.

Browse release notes for a selected Pega Version.

NOTE: Enter just the Case ID number (SR or INC) in order to find the associated Support Request.

Please update your bookmarks. This site will be discontinued in Dec 2024.

Pega Platform Resolved Issues for 8.1 and newer are now available on the Support Center.

INC-196447 · Issue 684644

Enhancements added for external Kafka Stream Service

Resolved in Pega Version 8.7

To ensure data privacy when using multi-tenant Stream Service hosted on a single Kafka cluster, access will be authorized based on ACLs when a tenant sends direct requests to Kafka. In addition, all Kafka resources (topics and consumer groups) are now able to contain a prefix naming convention which can be used for tenants. This is handled through using a <env name="services/stream/name/pattern" value="{tenant.name}-{environment}-{stream.name}"/> prconfig setting to set the stream name pattern. For example, if the tenant.name is resolved into "companyname", environment into "prod1", and the stream dataset name is pyFTSIncrementalIndexer, then the Topic name created on the external Kafka will be companyname-prod1-pyFTSIncrementalIndexer.

INC-196842 · Issue 683766

APIs updated for sending large attachments with MSGraph

Resolved in Pega Version 8.7

Sending mail using MSGraph was failing with the exception "Error during http request" when attachments were greater than 3MB. This was traced to changes made by Microsoft in the Graph API implementation which caused it to throw a runtime exception instead of the expected 413 status code and relevant message for the attachment size. To resolve this, an update has been made to ensure that any attachment larger than 3MB is sent via the Microsoft Outlook API.

INC-201338 · Issue 690896

Restored local blocking queue cache

Resolved in Pega Version 8.7

After update, it was not possible to bring up secondary VBD nodes after restarting. Investigation traced this to earlier work done to resolve a memory leak issue, in which stale entries for local blocking queues were removed from cache. This resulted in modifying the queue listener logic to use "cache.getQueueIfPresent(jobId)" instead of "cache.getQueue(jobId)". Because the listener was not creating the cache if it was not present and the cache which held the local blocking queue didn't have the entry for the current remote execution job ID, the caller of the remote execution on Node2 ended up in blocking state forever, waiting on the local blocking queue. To resolve this, the code has been updated to ensure the blocking queue is created and stored in the local queue cache before publishing the remote job message.

INC-202878 · Issue 713792

Handling updated for social messaging items with very large embedded images

Resolved in Pega Version 8.6.5

Job scheduler was running out of memory while handling large/multiple files, causing nodes to crash while handling Social Messaging items. Inline images were also not being sent in the ACK email. Investigation traced this to a custom job scheduler and activity used to create PDFs from email communication and send the PDFs to a third-party system. This used the standard Email Triage/PegaSocial functionality for the emails, which were stored as instances of PegaSocial-Message. The custom Job scheduler and activity then leveraged the Data Page D_pxEmailPosts to retrieve the emails in order to harvest the text from the emails. The issue occurred when a number of the emails had multiple or large embedded images in them, meaning the PegaSocial-Message instances are very large. This resulted in the node running out of memory and crashing. To resolve this, code has been added to support inline images in ACK email and the complete message from Pulse will be used instead of the original message as the email body for the ACK email.

INC-211974 · Issue 706616

UpgradeInteractionCase modified to update work page on save

Resolved in Pega Version 8.6.5

After update, service cases were failing to unpend when using a custom data transform that accessed the D_CorrespondenceCase data page. Investigation showed the D_CorrespondenceCase data page was internally using the D_pxGetInteractionCaseDetails data page; this call upgraded and saved and committed the case but this commit released the lock which the custom data transform had acquired. Because this happened before accessing the CorrespondenceCase datapage, it led to an error while committing the service case details. To resolve this, the postprocessing activity "pyUpgradeInteractionCase" of D_pxGetInteractionCaseDetails has been modified to write the update on the work page as soon as a save is performed.

INC-215937 · Issue 713773

Added exception handling for PageGroup alerts

Resolved in Pega Version 8.6.5

Queue items were going to the broken queue if there was an issue fetching the alert configuration from the Queue Processor rule. The error "java.lang.IllegalArgumentException: Alert id cannot be blank" was seen. This has been resolved by adding exception handling while gathering alerts from PageGroup so that a malformed alert configuration will not cause overall failure of a processed message, but instead an empty alert will be returned if configuration-data is corrupted.

INC-217781 · Issue 714185

JobScheduler updated to better handle DST change

Resolved in Pega Version 8.6.5

If a job scheduler was set to run on a weekly basis between 1 AM CET and 3 AM CET, the DST time change caused the job scheduler to skip that week. During DST, there is one 23-hour day in the year, and if execution time is set to that missing hour the system was throwing an IllegalArgumentException for the non-existent date. This has been resolved by adding a check that verifies whether a given date does exist; if it does not exist, the system will postpone execution time by one hour.

INC-218001 · Issue 719922

Error text revised for parameterized data page used for token generation

Resolved in Pega Version 8.6.5

While trying to add a claim in the header of a Token Generation Profile instance, selecting Map From as "Clipboard" and trying to give any DataPage(parameterized) as the source property failed to be saved and the error "JWS Alias— Please provide correct algorithm key with correct key length." appeared. Changing the "Map From" to a Constant and giving a dummy value worked as expected. Tracer showed the error "declare page parameters not supported by PropertyReference", indicating the actual issue: at this time, the Token profile does not support using a parameterized data page. This has been addressed by ensuring an appropriate error message is shown on save of the token profile rule form when a parameterized data page reference is configured. The error will now read "The reference D_pzPreferenceStore[PreferenceOperatorID:"[email protected]"].pxObjClass is not valid. Reason: Parameterized data page reference is not supported." Support for a parameterized data page used with Map From will be taken as an enhancement for a future release.

INC-218340 · Issue 714663

Override added to delete records for a stream dataset after processing

Resolved in Pega Version 8.6.5

Kafka data was accumulating for a Stream data set due to huge volume of inbound calls. This has been resolved by adding support to override pyDeletedProcessed through a DASS in order to remove the records for a particular stream dataset (topic) as soon as they are processed by Pega.

INC-218909 · Issue 715282

Override added to delete records for a stream dataset after processing

Resolved in Pega Version 8.6.5

Kafka data was accumulating for a Stream data set due to huge volume of inbound calls. This has been resolved by adding support to override pyDeletedProcessed through a DASS in order to remove the records for a particular stream dataset (topic) as soon as they are processed by Pega.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us