Skip to main content

Resolved Issues

View the resolved issues for a specific Platform release.

Go to download resolved issues by patch release.

Browse release notes for a selected Pega Version.

NOTE: Enter just the Case ID number (SR or INC) in order to find the associated Support Request.

Please note: beginning with the Pega Platform 8.7.4 Patch, the Resolved Issues have moved to the Support Center.

INC-142084 · Issue 599876

Support added for expression in strategy scorecards

Resolved in Pega Version 8.2.8

When invoking REST against a dataflow that had a strategy containing a scorecard that used an expression and the "Include model explanations option" was enabled in the Strategy configuration, the system failed with the error "PropertyValueInvalid .pxMaxScore Cannot cast the value (unknown) to double". This was traced to the Scorecard explanations failing during serialization when an expression was used, and has been corrected.

INC-143927 · Issue 599491

Oracle database performance improvements

Resolved in Pega Version 8.2.8

When the IH Summary was enabled and materialized on an ADM model, updating the ADM model was very slow on large sites. This has been resolved by adding several performance improvements for working with Oracle databases, including Oracle pre/post processing steps.

SR-D93777 · Issue 565692

Handling added for Oracle Aggregate IH Summaries

Resolved in Pega Version 8.2.8

When using using (Non Materialized) IH Summaries to aggregate IH data, the data returned by the IH summary did not include all the expected records. If the same criteria was executed on the database via SQL or by using a strategy to process raw IH data then the results were as expected. This was due to a difference in handling of Oracle vs postgres which causes an order by clause not to be generated in the query: the postgres column name is lower case, while in Oracle it is upper case. This has been resolved by updating the system to get the column name correctly from the propertytocolumn map so IH records are returned in correct order by Browse By Keys operation.

SR-D95605 · Issue 565486

Data Flow correctly saved to Database table Dataset

Resolved in Pega Version 8.2.8

After setting up a dataflow with Report definition as a source and Database table dataset as destination with the option "Insert new records and override existing", a data transform was used to modify a few of the values and write to the same database table. This table was mapped to pegadata schema and didn't have a pzpvstream column. Running the data transform generated an exception stating "DataStoreSaveStatementWithoutStream(PageDatabaseMapperImpl). The error was not seen when the pzpvstream column was added or if pzInskey was also removed along with pzpvstream, or if "only insert new records" was selected. This was traced to pzInsKey and pxInsName being null in the query formed while writing. There are two execution paths for the database dataset save operation, one for internal tables and another one for external tables. For a table with pzInskey but no blob, the system was incorrectly using the external table logic. This has been corrected.

INC-202111 · Issue 710105

Logging extended for PRPCPropertyInfoProvider

Resolved in Pega Version 8.6.5

In order to assist with diagnosing issues with Kafka and JSON, additional logging has been added for PRPCPropertyInfoProvider.

INC-208976 · Issue 719166

Enhanced SSA metrics made available

Resolved in Pega Version 8.6.5

In order to better diagnose delays related to the time when a Campaign is scheduled to start and the time when the Dataflow actually starts to run, an update has been made which will generate detailed metrics to cover some of the strategy execution key performance intensive areas. Additional lower level internal metrics related to SSA engine execution have also been made available by way of a DSS to collect more runtime insight for diagnosis. To enable the collection of these Level 2 SSA internal metrics, set the dataflow/shape/strategy/detailed_metrics/level2 DSS in the Pega-DecisionEngine rule set to 'true'. A comprehensive set of enhanced metrics will be available in Pega 8.8.

INC-213217 · Issue 714194

Updated context handling for executing data transforms in dataflows

Resolved in Pega Version 8.6.5

When performing a sort operation in a data transform on a page list and invoking the Data transform through a Data flow, a java.lang.UnsupportedOperationException was generated on the sort step even though the data was correct. Investigation showed that the dataflows were using a different execution context than the context used for regular activity execution. An update has been made to ensure the correct context is being used while executing data transforms in dataflows.

INC-215601 · Issue 721524

Explicit Kerberos renewal added for HDFS

Resolved in Pega Version 8.6.5

When using WebSphere/IBM with Hadoop connectivity, the Kerberos credentials were expiring after 7 days and causing connectivity failures. Investigation showed the Apache Hadoop HDFS Client was not renewing the ticket even when there were operation calls every few hours. To resolve this, an enhancement has been added to the Pega HDFS client which will explicitly tell Hadoop authentication layer to check for Kerberos ticket renewal instead of relying on Hadoop built-in logic.

INC-217290 · Issue 721376

Added support for creating predictive models in Production

Resolved in Pega Version 8.6.5

While creating a new predictive model rule in Prediction studio, the case was going into broken process after selecting the template with the error message "Error loading D_ProjectList , Reason : No databases defined in properties file:/databases.properties". This was an unexpected use case for creating models in Production level, and has been resolved by updating the flows to turn off the draft mode in this scenario.

INC-217728 · Issue 712733

Made CommitLog messages more resilient to class change

Resolved in Pega Version 8.6.5

Persistent class mismatch exceptions were seen while refreshing the ADM models after updating from Pega 8.5. This was caused by changes to the Cassandra adm_commitlog message schema which caused the commit log messages generated to be incompatible with the deserialisation logic present in higher versions of Pega, but as the commit log messages had a time to live of 24 hours the issue self-resolved. However, this has been resolved by setting serialVersionUID for the ModelParameters class so CommitLog messages are now compatible with future minor and patch releases.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us