SR-D94019 · Issue 0
Alerting features added for Listeners
Resolved in Pega Version 8.4.2
An enhancement has been added for Email Listener alerts. At the end of each listener's processing cycle, the system will determine the average processing time per message, the total processing time of the listener, the total unread messages count, and total number of messages processed. In each AESRemote agent run, this data will be published to the PDC, where it will be analyzed and any necessary alerts raised.
SR-D94505 · Issue 553018
Toggles added to allow file listener performance improvement
Resolved in Pega Version 8.4.2
In order to offer flexibility in improving the performance of file listener CPU usage while processing intake files, rsf.setSaveIntermediateState( true ) has been replaced with rsf.setSaveIntermediateState( listener.getListenerProperties().getAttemptRecovery() ) in FileActionImpl. This change allows the option of not selecting the setting for attempting recovery on the file listener rule-form, which then would skip the intermediate save state normally performed per record/per file in case recovery is needed, thereby speeding up the processing. A toggle has also been added to see whether the system should bypass the ListenerState check at the end of each record. To use this, set the Dynamic System Setting "listener/skipListenerStateCheck" in the "Pega-IntegrationEngine" ruleset to true.Additional Installation Instructions Please create/set the Dynamic System Setting "listener/skipListenerStateCheck" in the "Pega-IntegrationEngine" ruleset as true.
INC-125136 · Issue 575999
Topic name prefix support added for Kafka
Resolved in Pega Version 8.4.2
An enhancement has been added to support the use of site-specific standards for the naming of queues. This can be used to add clarification between Kafka Topics created by Pega and those created for internal topics when using a shared Kafka Cluster. This is done using the new setting: <env name=”services/stream/name/pattern” value=”prefix{stream.name}postfix”/> .As an example, if tenant.name is resolved into "myco" and the environment into "prod1", value="{tenant.name}-{environment}-{stream.name}" will result in stream myco-prod1-{stream.name} By default, when stream.provider = EXTERNAL, the default will be “pega-{stream.name}” .
INC-120326 · Issue 567450
Landing page refresh modified to avoid frequent reloads
Resolved in Pega Version 8.4.2
Approximately every two weeks, clients were able to login but it was not possible to work as the landing page refreshed constantly until all of the nodes were rebooted and the issue was cleared. Investigation indicated the frequent reloads were related to the cacheing of the operator details, and this has been resolved by updating the datapage reload strategy based on 'when' so frequent reloads will be avoided.
INC-120600 · Issue 565200
Handling added to avoid primary key exception in pxExtractDataWithArgs
Resolved in Pega Version 8.4.2
While running the activity pxExtractDataWithArgs with -c, the "Error while getting the sequenceNumberERROR: duplicate key value violates unique constraint pr_log_pk" occurred if both Parent and Child classes were pointing to different table and extraction started from the child class. In this case, because the Parent and Child classes were pointing to different tables while extracting with the -c option, the class list order caused a primary key constraint violation exception while generating the process sequence number. To resolve this, while adding the classes to the instance map, the order will be changed to parent followed by child classes.
INC-127635 · Issue 570057
Handling added for BIX count mismatch in manifest and extracted CSV data
Resolved in Pega Version 8.4.2
A mismatch in the counts in the manifest file when compared to the data extracted in the CSV files was traced to a corrupted or wrong pzinskey in the source database. If the BLOB is corrupted while extracting to output type and manifest type to CSV, an invalidstream exception was generated and resulted in the manifest file and extracted CSV file total instance count mismatch. This has been resolved by adding handling to balance the counts.
INC-133173 · Issue 573443
Landing page refresh modified to avoid frequent reloads
Resolved in Pega Version 8.4.2
Approximately every two weeks, clients were able to login but it was not possible to work as the landing page refreshed constantly until all of the nodes were rebooted and the issue was cleared. Investigation indicated the frequent reloads were related to the cacheing of the operator details, and this has been resolved by updating the datapage reload strategy based on 'when' so frequent reloads will be avoided.
SR-D95626 · Issue 563585
Added handling for download of attachment with special characters
Resolved in Pega Version 8.4.2
If the name of an uploaded attachment contained a non-ascii character, the upload was successful but the downloaded file had the wrong name without any suffix. This was an issue on WAS or JBoss servers where filenames contained non-ascii characters, and the filename parameter in the content disposition header also contained these characters. To resolve this, the filename parameter in the content disposition header will be encoded as well along with the filename* parameter to handle non-ascii characters in the filename.
INC-121460 · Issue 569118
IH Summaries correctly loads aggregated count
Resolved in Pega Version 8.4.2
After the IH Aggregated DataFlow has processed more than 34 million records from an Oracle database, the IH Fact table contained only ~120,000 records. Running a browse on the pxInteractionHistory DS resulted in an infinite loop. Investigation showed that this was caused by the handling of Oracle timestamps: using Pega DateTime in query parameters generated Oracle DATE instead of TIMESTAMP, causing all query parameters to be rounded to the second and interfering with the IH pre-aggregation. This has been resolved by updating the handling.
INC-122180 · Issue 563270
Null check added for Entity Extraction result
Resolved in Pega Version 8.4.2
Training data for an Entity Extraction model was not getting distributed correctly, as per the given % data. This was due to the result of the Entity Extraction result being null, and has been resolved by adding a check that will consider detected types as primary candidate for grouping if the result column is null.