INC-141579 · Issue 591852
Enhancement added for node heartbeat recovery process
Resolved in Pega Version 8.3.5
Nodes were not showing up in the admin portal even though they were up and running and could be seen in the pr_sys_statusnodes table. The exception "An exception was encountered while invoking the cluster membership listener callback" was seen. All nodes became visible again after multiple restarts. The root cause was traced back to a temporary database connectivity problem. The database itself was fine according to database monitoring reports, but a network problem, a slow database query, or another issue prevented Pega from establishing a connection for more than a minute. An enhancement has been made to resolve this: if a node becomes unhealthy due to the service registry missing due to a failed heartbeat, the heartbeat will try to recover after 60 seconds and keep trying every 30 seconds until it succeeds.
INC-142674 · Issue 594496
Enhancement added for node heartbeat recovery process
Resolved in Pega Version 8.3.5
Nodes were not showing up in the admin portal even though they were up and running and could be seen in the pr_sys_statusnodes table. The exception "An exception was encountered while invoking the cluster membership listener callback" was seen. All nodes became visible again after multiple restarts. The root cause was traced back to a temporary database connectivity problem. The database itself was fine according to database monitoring reports, but a network problem, a slow database query, or another issue prevented Pega from establishing a connection for more than a minute. An enhancement has been made to resolve this: if a node becomes unhealthy due to the service registry missing due to a failed heartbeat, the heartbeat will try to recover after 60 seconds and keep trying every 30 seconds until it succeeds.
INC-137255 · Issue 584910
Resolved Cassandra query read/write latency
Resolved in Pega Version 8.3.5
Read/write latency was seen when certain Cassandra queries were triggered by the system. This was traced to an unusual case where a flush was not done for an exceptionally long time period, and has been resolved.
INC-121460 · Issue 569118
IH Summaries correctly loads aggregated count
Resolved in Pega Version 8.4.2
After the IH Aggregated DataFlow has processed more than 34 million records from an Oracle database, the IH Fact table contained only ~120,000 records. Running a browse on the pxInteractionHistory DS resulted in an infinite loop. Investigation showed that this was caused by the handling of Oracle timestamps: using Pega DateTime in query parameters generated Oracle DATE instead of TIMESTAMP, causing all query parameters to be rounded to the second and interfering with the IH pre-aggregation. This has been resolved by updating the handling.
INC-122180 · Issue 563270
Null check added for Entity Extraction result
Resolved in Pega Version 8.4.2
Training data for an Entity Extraction model was not getting distributed correctly, as per the given % data. This was due to the result of the Entity Extraction result being null, and has been resolved by adding a check that will consider detected types as primary candidate for grouping if the result column is null.
INC-125838 · Issue 561484
Ignore Case feature updated
Resolved in Pega Version 8.4.2
In Pega Marketing 8.4 on the Action rule form, configuring an eligibility criteria and selecting the option for the comparison to ignore case did not copy this part of the criteria to the back proposition filter rule, so the case of the comparison was not ignored at runtime. This has been corrected.
INC-125900 · Issue 560788
ResolveFTCR flow Draft mode turned off
Resolved in Pega Version 8.4.2
When using the FastTrack Change Request functionality of Revision Manager 8.3 the submission was going to Pending packaging instead of Resolved-Completed with the error "flow is in draft mode; it cannot be executed". This was due to pxResolveFTCR flow having been checked in with Draft mode turned on, and has been corrected by turning Draft mode off.
INC-126129 · Issue 569668
PropertyToColumnMap made more robust
Resolved in Pega Version 8.4.2
The DF_ProcessEmails dataflow was intermittently failing with a StageException error. This was traced to schema changes being propagated asynchronously by system pulse, which seem to have caused PropertyToColumnMap to cache stale schema. To resolve this, if the property mapping is not found the first time, the system will make a second attempt to get the mapping. Additional logging has also been added for better diagnostics.
INC-126187 · Issue 564340
Report Definition replaced with a data set for training the data and build model
Resolved in Pega Version 8.4.2
An issue with training the data and build model via an email or chat channel was traced in an exception in logs indicating the inability to execute a Report Definition. This can happen when UDF are removed/disabled in the database or in environments where functions are not allowed to read the blob for performance reasons. To resolve this, the Report Definition has been replaced with a data set.
INC-126750 · Issue 564155
DSS added to configure Dataset-Execute page handling
Resolved in Pega Version 8.4.2
When Kafka Data Set pages were saved to a data set with the Dataset-Execute method, there was is feedback if any of the pages were not successfully saved. Instead, the step always completes as successful. In addition, if any properties are added or modified by the save operation itself, those changes are not visible. This is due to the data set execute save operation saving pages as DSM pages to the data set. Due to the conversion of the pages, copies of the pages are used which do not reflect back any changes on the input pages. DSM pages are used by default because they are more lightweight than regular clipboard pages and therefore have potentially better performance. In order to allow the use of DSM pages to be customized, a new Dynamic System Setting: dataset/execute/save/statusFailOnError has been added. This can be enabled by setting it to true; it is disabled by default for greater backwards compatibility. By removing the DSM page conversion in the generated save code, changes to input pages will be reflected back if any are performed by the data set save operation, and the system will report back which pages are saved or failed by adding messages to the pages that failed to save. Performance may be affected with this change as regular clipboard pages are in general slower than DSM pages, however, that may be offset by removing the conversion to DSM pages process and will depend on the site configuration.