INC-141579 · Issue 591852
Enhancement added for node heartbeat recovery process
Resolved in Pega Version 8.3.5
Nodes were not showing up in the admin portal even though they were up and running and could be seen in the pr_sys_statusnodes table. The exception "An exception was encountered while invoking the cluster membership listener callback" was seen. All nodes became visible again after multiple restarts. The root cause was traced back to a temporary database connectivity problem. The database itself was fine according to database monitoring reports, but a network problem, a slow database query, or another issue prevented Pega from establishing a connection for more than a minute. An enhancement has been made to resolve this: if a node becomes unhealthy due to the service registry missing due to a failed heartbeat, the heartbeat will try to recover after 60 seconds and keep trying every 30 seconds until it succeeds.
INC-142674 · Issue 594496
Enhancement added for node heartbeat recovery process
Resolved in Pega Version 8.3.5
Nodes were not showing up in the admin portal even though they were up and running and could be seen in the pr_sys_statusnodes table. The exception "An exception was encountered while invoking the cluster membership listener callback" was seen. All nodes became visible again after multiple restarts. The root cause was traced back to a temporary database connectivity problem. The database itself was fine according to database monitoring reports, but a network problem, a slow database query, or another issue prevented Pega from establishing a connection for more than a minute. An enhancement has been made to resolve this: if a node becomes unhealthy due to the service registry missing due to a failed heartbeat, the heartbeat will try to recover after 60 seconds and keep trying every 30 seconds until it succeeds.
INC-137255 · Issue 584910
Resolved Cassandra query read/write latency
Resolved in Pega Version 8.3.5
Read/write latency was seen when certain Cassandra queries were triggered by the system. This was traced to an unusual case where a flush was not done for an exceptionally long time period, and has been resolved.
SR-D37163 · Issue 505477
Corrected Decision Data import filter behavior
Resolved in Pega Version 8.2.4
After upgrade, if a Decision data import referred another component to filter out the import, it did not work. This has been corrected.
SR-D12733 · Issue 488666
Code fragment removed to eliminate Fortify false positive
Resolved in Pega Version 8.2.4
A code remnant related to Boolean.getBoolean(..) in Rule-Declare testConsistency was causing a false positive in a Fortify scan. This piece of code is obselete and is not used anywhere, and has been removed.
SR-D26976 · Issue 507217
Filter added to ensure correct context for proposition strategy rules
Resolved in Pega Version 8.2.4
Given two applications (ex App1 and App2) hosted on the same domain where App2 was built on App1, trying to create a strategy rule in App1 and do a test run strategy using the propositional data component which internally uses App2 propositions generated the error: Failed to find a 'RULE-DECISION-DECISIONPARAMETERS' with the name 'GROUP_2'. There were 1 rules with this name in the rulebase, but none matched this request." Investigation showed the strategy was using the PropositionNoCacheUtils and PropositionTools java classes to load the propositions during run time. In these classes, the group classes were browsed from the db irrespective of the application context, causing the strategy run to fail as it was not able to access the decision data rules in other applications which shared the same SR class as the current application. To resolve this, a filter has been added to the PropositionNoCacheUtils and PropositionTools java classes to filter out the groups that are not in the current application context.
SR-D36591 · Issue 507537
Kafka producer message size made configurable
Resolved in Pega Version 8.2.4
Kafka producer was using a default max message size setting of 1.3Mb while the Kafka broker max message size was set to 5Mb. This caused large processing queues to eventually throw errors indicating a scheduler.JobExecutionException related to "There was a problem saving an instance of class System-Message-QueueProcessor-DelayedItem". To correct this, a producer message size configuration option has been added along with additional logging for the KafkaSaveOperation.
SR-D41730 · Issue 508143
TTL value correctly passed for Adaptive Event store
Resolved in Pega Version 8.2.4
The ADM table was growing due to the Time to Live (TTL) for entries in the Adaptive Event Store not being propagated to clean them out. This was traced to the TTL field on the data flow not being checked, causing the TTL value to be supplied as zero so there was no expiration. This has been corrected.
SR-D37945 · Issue 506799
Server node cache refresh will use remote execution timeout
Resolved in Pega Version 8.2.4
A campaign was failing due to VBD remote ping timeout with a stacktrace that indicated a StageException. Investigation showed that when the cluster is heavily loaded, calls to the remote execution API could time out. If this occurred when the VBD client was refreshing its cache of VBD server nodes, then the insert failed and the error was propagated up the calling data flow. To resolve this, the system will use the remote execution timeout when refreshing node cache, extend the timeout to 60 seconds, and ensure timeouts are retried during inserts.
SR-D38415 · Issue 507995
Resolved Transfer to Queue duplicate assignments
Resolved in Pega Version 8.2.4
The Transfer to a Queue option was creating duplicate assignments after accepting the Chat. Once the chat was accepted, an instance was created\maintained in both Assign-Workbasket and Assign-Worklist tables. This happened after adding the "Transfers" queue to an Agent; if that queue was not added, the transfer to a queue option gave an error to the Agent receiving\accepting the chat. The Work-.ReassignDefaults activity is an extension that was customized in the Pega-DecisionManager ruleset for a certain use case in revision management. The customization is no longer required and has become redundant and has therefore been removed to resolve this issue.