SR-A96852 · Issue 264159
Resolved OOM error when adding VDB DNode
Resolved in Pega Version 7.2.2
A memory leak was found relating to JBDC PreparedStatement objects (char arrays allocated to hold database query results) used when VBD loads its data partitions. This has been fixed with the addition of code to close the PreparedStatement objects created when loading Actuals partition data.
SR-A99782 · Issue 266250
XSS vulnerability closed in Warning Justification Text field
Resolved in Pega Version 7.2.2
A potential XSS vulnerability has been addressed in the Warning Justification Text field.
SR-A96515 · Issue 267504
Added handling for Qualsys security scan
Resolved in Pega Version 7.2.2
A stack overflow error was being generated when running a Qualsys vulnerability security scan on servers hosting IBM WebSphere and Pega, resulting in two of the three Pega nodes in production becoming unresponsive and requiring a restart. This was caused by the third party tool connecting to port 7003, which is used by Stream data set server, and sending unknown requests. To avoid this, the system has been updated to not send back a response if the channel is closed
SR-A79502 · Issue 253855
Corrected ability to add DNode port via prconfig
Resolved in Pega Version 7.2.2
Adding DNode Port via prconfig was not taking effect due to the configuration values being stored as strings, but the code which reads the storage port configuration treats the setting as integer. This has been fixed.
SR-A87150 · Issue 256602
Dynamic setting added to override ADM JMS default
Resolved in Pega Version 7.2.2
ADM (a part of Pega DSM) was detecting JMS and switching over to JMS when available and using HttpInvoker in other cases. In order to allow customization, it is now possible to override the ADM JMS configuration and allow EAR deployments by way of a dynamic system setting to disable JMS at will.
SR-A102723 · Issue 270420
Pulse will not fire ADM model update until node is ready
Resolved in Pega Version 7.2.2
ADMExceptions were seen during ADM node restart due to ADM clients triggered by Pulse trying to update models on the server when the server is not yet in NORMAL state. To correct this, Pulse operation has been modified to not to fire the model update process when the node is not yet in NORMAL state.
SR-A95909 · Issue 265781
PegaMarketing custom shape RUF compilation fixed
Resolved in Pega Version 7.2.2
After upgrade, the PegaMarketing custom data flow shape was failing while calling Rule-Utility-Function libraries. This was caused by an error in the compilation process, and the data flow generation code has been modified to use a different class loader that takes libraries into account.
SR-B1679 · Issue 272770
Dataflow pre- and post-activities can be run across nodes
Resolved in Pega Version 7.2.2
An enhancement has been added to allow the execution of pre- and/or post-activities on all data flow nodes by way of the following properties. Please note these are not available in the UI and must be used in programmatically-created runs. Data-Decision-DDF-RunOptions.pyPreActivity.pyRunOnAllNodes = true Data-Decision-DDF-RunOptions.pyPostActivity.pyRunOnAllNodes = true
SR-B5134 · Issue 274372
Smoothed data flow terminations
Resolved in Pega Version 7.2.2
An issue was found with data flows not correctly terminating when encountering an error. This has been fixed by ensuring a work object is marked as stopped when an assignment fails. In addition, when performing a commit, a dataflow failed with "Interrupted by unexpected service shutdown". This has been corrected by ensuring the system always provides a regular clipboard page for assignments and work object and not a DSM clipboard when loading/saving an assignment or work object from/to the database.
SR-A87606 · Issue 257332
VDB start waits for Cassandra node with keyspace data
Resolved in Pega Version 7.2.2
An issue was found with restarting a DNode if there were no copies of the VBD keyspace data on the available nodes in the Cassandra cluster when VBD started up. VBD uses replica factor 3 on its keyspace, meaning there will be at most 3 nodes in the cluster with full copies of the VBD data. If when restarting the cluster only Cassandra nodes that happen to not have VBD's data were brought up, starting up VBD would hang trying to read partition summary data. As a solution, If VBD is started and its Cassandra data is not yet available, the process to load partition summary data will wait until the next time there is VBD activity. Once the Cassandra node comes up with VBD's data, it will be loaded and VBD functionality will be enabled. Additionally, an intermittent deadlock was discovered when starting 2 VBD nodes in parallel. The deadlock appeared when VBD was initializing its persistence at the same time another thread checked to see if persistence was initialized (triggered by remote request from second node). The first thread owned a Hazelcast distributed lock and attempted to use a Supplier to get an instance of an object. The second thread was already calling the Supplier and was waiting for the Hazelcast distributed lock. This deadlock has been fixed.