Skip to main content

Resolved Issues

View the resolved issues for a specific Platform release.

Go to download resolved issues by patch release.

Browse release notes for a selected Pega Version.

NOTE: Enter just the Case ID number (SR or INC) in order to find the associated Support Request.

Please note: beginning with the Pega Platform 8.7.4 Patch, the Resolved Issues have moved to the Support Center.

SR-D90579 · Issue 550769

Real Time data flow exception during shutdown corrected

Resolved in Pega Version 8.2.7

A Real Time DataFlow run was failing with a java.util.concurrent.RejectedExecutionException. This occurred when the node was shutting down while picking up a partition, and was due to the life cycle events not being able to update the partition status. To resolve this, a check has been added to prevent the exception form occurring by evaluating whether the Executor is shutdown while distributing messages.

SR-D96836 · Issue 555748

Refinements made to MarkerNode memory use

Resolved in Pega Version 8.2.7

Significant memory usage can be observed when data join in strategies are missing the join conditions. E.g., when the primary source in the strategy can have up to 300 or even more propositions, the join without the join condition would perform a cartesian product with possibly up to 400 or even more records returned. This may cause performance degradation. To guard against this, an update has been made that seeks to prevent repetitive markers being accumulated under a MarkerNode while trying not to rely on any implementation of equality and hashCode for individual marker implementations.

INC-181148 · Issue 663164

Documentation updated for altering data type to Timestamp to avoid duplicate data in Decision Hub

Resolved in Pega Version 8.5.6

By default, the pxOutComeTime column has a Date data type in the pr_data_ih_fact table. This may cause duplicate data issues. To resolve this, the 8.5 and 8.6 installation and upgrade guides for Pega Customer Decision Hub have been updated to include a new procedure "Modifying column definitions to use the TIMESTAMP datatype in an Oracle database". Performing this procedure will allow users to avoid the duplicate data issue. All documents are available from the Pega Customer Decision Hub product page.

INC-172785 · Issue 662335

Adaptive model retry mechanism enabled

Resolved in Pega Version 8.5.6

Adaptive models were missing from the Model Management page as well as in the Prediction studio while similar models for the same proposition, only differing by the Channel name, were visible. This was traced to data not being synchronized between the database and Cassandra. The pegadata.pr_data_adm_factory database table did not contain the record of the missing channel, but Cassandra did. Since the current Cassandra adm_scoringmodel contained model information, the system still believed the model was present. In order to ensure Cassandra and the database table are in sync, an update has been made to enable the retry mechanism "SyncFactoryKeysTask" to create the ADM model in factory table by periodically looking for scoring models without factories or an entry in adm_meta.

INC-173596 · Issue 660221

Google OAuth and Spring versions updated

Resolved in Pega Version 8.5.6

The Google-oauth-client jar has been upgraded to version 1.31.1, and SpringFramework libraries have been updated to version 5.3.9 .

INC-174661 · Issue 678821

Handling added for clearing node killed between assignment and processing

Resolved in Pega Version 8.5.6

An Offer flow was not resuming after it expired according to the wait shape. Investigation traced this to partitions which were assigned to a dead node in NEW state where they were not picked up by the dataflow. The problem was only encountered in the unusual situation when a dataflow node was killed in the brief period of time between the assignment and the processing, and has been resolved by adding an update which will clear unknown new assigned partitions for the batch run health task.

INC-174781 · Issue 655120

Kerberos authentication added for external Cassandra

Resolved in Pega Version 8.5.6

Support has been added for Kerberos authentication with Cassandra.

INC-175207 · Issue 655553

Added handling for DSM Services stuck in leaving status after database outage

Resolved in Pega Version 8.5.6

During a database outage, the heartbeat would fail and DSM services would eventually try to enter safe mode and stop. As the first step they would try to change the state to LEAVING, but because the database was down saving the LEAVING state failed and the exception was not handled correctly. This resulted in the rest of the stop operation logic not being executed and the service being stuck in LEAVING. To resolve this, an update has been made to ensure the service goes to LEAVING_FAILED if anything fails during the stop operation including when setting state to LEAVING_FAILED. The state LEAVING_FAILED will get flushed to the database eventually when it comes back up. This will allow the aggregation service to start from the LEAVING_FAILED state and recover by itself after a database outage.

INC-177031 · Issue 664856

Adaptive model retry mechanism enabled

Resolved in Pega Version 8.5.6

Adaptive models were missing from the Model Management page as well as in the Prediction studio while similar models for the same proposition, only differing by the Channel name, were visible. This was traced to data not being synchronized between the database and Cassandra. The pegadata.pr_data_adm_factory database table did not contain the record of the missing channel, but Cassandra did. Since the current Cassandra adm_scoringmodel contained model information, the system still believed the model was present. In order to ensure Cassandra and the database table are in sync, an update has been made to enable the retry mechanism "SyncFactoryKeysTask" to create the ADM model in factory table by periodically looking for scoring models without factories or an entry in adm_meta.

INC-179727 · Issue 704546

Modified batch requestor handling to ensure cleanup

Resolved in Pega Version 8.5.6

A large buildup of batch requestors was seen, and restarting the node did not clear it. Investigation showed that in the case of one class reading a DSS value, a Pega requestor was being created when it was unable to retrieve any Pega context, and this requestor was not cleaned up afterwards. To resolve this, an update has been made to the way the requestor and Pega context is being created along with ensuring it will be cleaned up properly after use.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us