SR-A24962 · Issue 250080
DataSet-Execute updated to correct time-to-live error
Resolved in Pega Version 7.2.2
When using the time to live option of a Decision Data Store dataset, the java generation for the DataSet-Execute method was causing java compilation errors. This was due to an inconsistency between the code generation for DataSet-Execute method in an activity and the settings provided. To correct this, the algorithm used to generate the code for DataSet-Execute method has been updated to include the time to live setting and to generate correct and compilable code.
SR-A24962 · Issue 248678
DataSet-Execute updated to correct time-to-live error
Resolved in Pega Version 7.2.2
When using the time to live option of a Decision Data Store dataset, the java generation for the DataSet-Execute method was causing java compilation errors. This was due to an inconsistency between the code generation for DataSet-Execute method in an activity and the settings provided. To correct this, the algorithm used to generate the code for DataSet-Execute method has been updated to include the time to live setting and to generate correct and compilable code.
SR-A80784 · Issue 254197
Enhancement for Facebook connector retrieval
Resolved in Pega Version 7.2.2
Whenever the Facebook connector was restarted, historical posts on the monitored page were being fetched and duplicate documents were getting harvested into the DB. To resolve this, a new search parameter has been added for the Facebook connector that will receive a date as parameter for how many period/days of historical data to fetch.
SR-B6425 · Issue 275629
Scrum Board chicklet error resolved
Resolved in Pega Version 7.2.2
After collapsing and expanding a user story chicklet in the Scrum Board, adding a new note to the user story would cause the user story chicklet to disappear. This was a rendering issue due to an error in the type property syncing with the level property, and has been fixed.
SR-A75956 · Issue 250547
D_OfflineWorkIDs DP performance improvements
Resolved in Pega Version 7.2.2
The logic of D_OfflineWorkIDs DP has been updated to improve performance for the SA PROD App.
INC-203994 · Issue 698853
DSS added to handle merges with lower versions of Postgres
Resolved in Pega Version 8.7.1
After update, executing the batch campaign with volume constraint resulted in the second data flow DF_Wait failing with error message "ERROR: number of columns (1844) exceeds limit (1664)". This was due to the database set’s change (in 8.5) to use the database layer’s merge statement. Prior to that, the logic used "deletes and inserts". Depending on the version of Postsgres, the generated SQL statement for a merge statement is different. The “INSERT … ON CONFLICT … UPDATE” syntax is generated for Postgres 9.5+ AND when there is a PK constraint defined for the DB table. Otherwise, the complex UPSERT statement (old syntax) is generated, as was the case in this issue. This is a known issue in the Postgres server software where it mis-interprets the number of columns involved. i.e., it mistakenly counts the number of columns twice. As a result, the actual maximum columns allowed is only half of the official limit (1664). The same UPSERT statement does not cause the “exceeds limit” exception if there are 832 or fewer columns in the statement. To resolve this, an option has been provided to select between the “original logic” (deletes and inserts) and the “merge statements” logic by way of the DSS “decision/datasets/db/useMergeStatementForUpdates”. Setting “true” will use the merge statement logic, and setting “false” will use deletes and inserts. When the DSS is not defined, the default is "true" and the system will use merge statements in the form preferred by Postgres 9.5+.
INC-180246 · Issue 699700
Support for apostrophe added to keyword tokenization
Resolved in Pega Version 8.7.1
A keyword containing an apostrophe was not detected properly in Text extraction model. This has been resolved by updating the annotator used in the tokenization.
INC-193399 · Issue 688115
DSS added to handle merges with lower versions of Postgres
Resolved in Pega Version 8.7.1
After update, executing the batch campaign with volume constraint resulted in the second data flow DF_Wait failing with error message "ERROR: number of columns (1844) exceeds limit (1664)". This was due to the database set’s change (in 8.5) to use the database layer’s merge statement. Prior to that, the logic used "deletes and inserts". Depending on the version of Postsgres, the generated SQL statement for a merge statement is different. The “INSERT … ON CONFLICT … UPDATE” syntax is generated for Postgres 9.5+ AND when there is a PK constraint defined for the DB table. Otherwise, the complex UPSERT statement (old syntax) is generated, as was the case in this issue. This is a known issue in the Postgres server software where it mis-interprets the number of columns involved. i.e., it mistakenly counts the number of columns twice. As a result, the actual maximum columns allowed is only half of the official limit (1664). The same UPSERT statement does not cause the “exceeds limit” exception if there are 832 or fewer columns in the statement. To resolve this, an option has been provided to select between the “original logic” (deletes and inserts) and the “merge statements” logic by way of the DSS “decision/datasets/db/useMergeStatementForUpdates”. Setting “true” will use the merge statement logic, and setting “false” will use deletes and inserts. When the DSS is not defined, the default is "true" and the system will use merge statements in the form preferred by Postgres 9.5+.
INC-193632 · Issue 679172
Cassandra driver metrics exposed for performance troubleshooting
Resolved in Pega Version 8.7.1
By default Cassandra driver metrics are now enabled. Metrics can be disabled by setting the dnode/disable_driver_metrics prconfig parameter.
INC-193847 · Issue 695974
DSS added to allow masking of subjectID in alerts
Resolved in Pega Version 8.7.1
In order to allow customizing whether or not a subjectID is included in alerts, a DSS has been added to conditionally mask the subjectID from being logged. To use this, set the "alerts/maskIHsubjectID" DSS in the Pega-DecisionEngine ruleset to true to hide the pySubjectID.