INC-220642 · Issue 717032
Updated context handling for executing data transforms in dataflows
Resolved in Pega Version 8.8
When performing a sort operation in a data transform on a pagelist and invoking the Data transform through a Data flow, a java.lang.UnsupportedOperationException was generated on the sort step even though the data was correct. Investigation showed that the dataflows were using a different execution context than the context used for regular activity execution. An update has been made to ensure the correct context is being used while executing data transforms in dataflows.
INC-213217 · Issue 714192
Updated context handling for executing data transforms in dataflows
Resolved in Pega Version 8.7.2
When performing a sort operation in a data transform on a page list and invoking the Data transform through a Data flow, a java.lang.UnsupportedOperationException was generated on the sort step even though the data was correct. Investigation showed that the dataflows were using a different execution context than the context used for regular activity execution. An update has been made to ensure the correct context is being used while executing data transforms in dataflows.
INC-220642 · Issue 717033
Updated context handling for executing data transforms in dataflows
Resolved in Pega Version 8.7.2
When performing a sort operation in a data transform on a page list and invoking the Data transform through a Data flow, a java.lang.UnsupportedOperationException was generated on the sort step even though the data was correct. Investigation showed that the dataflows were using a different execution context than the context used for regular activity execution. An update has been made to ensure the correct context is being used while executing data transforms in dataflows.
INC-218145 · Issue 715679
DSS introduced to control DSM clipboard page serialization
Resolved in Pega Version 8.8
When using a Kafka dataset to consume a message from an external topic that had an attribute name with a special character contained in a page list structure, using a JSON data transform for the mapping in a realtime dataflow resulted in the error "Exception in stage: KafkaDS; LegacyModelAspectInvokableRuleContainer.invoke-Exception encountered a :java.lang.UnsupportedOperationException." To resolve this, a new DSS dataset/CLASS_NAME/DATASET_NAME/JSONDataTransform/deserialization/useDSMPage has been introduced. When the value is set to true, the process will follow the previous behavior of DSM clipboard pages being generated when Kafka records are deserialized using JSON data transform. When the value is set to false, the JSON data transform will generate regular clipboard pages and convert them later to DSM clipboard pages. This would avoid errors when a JSON data transform calls methods from the Clipboard API that are not implemented by DSM pages. This DSS is set per data set instance. CLASS_NAME and DATASET_NAME are placeholders which should be replaced by data set's pyClassName and pyPurpose property values. In addition, a similar DSS, dataset/CLASS_NAME/DATASET_NAME/JSONDataTransform/serialization/useDSMPage, has been introduced for serialization.
INC-175707 · Issue 655347
Multi-level strategy gives correct rules during Batch run
Resolved in Pega Version 8.7
When using a nested page structure with multi-level strategies that loop through each page and process the customer, a customer with two linked customers A and B had the results of A overridden by B during Batch Processing while single case canvas testing showed the correct results. This was caused by the DSM clipboard page method of carrying the parent page info not supporting the re-use of pages, and has been resolved by adding an enhancement that will sanitize the strategy stage processor to ensure the correct page reference is used.
INC-210126 · Issue 708784
Timeout added for Cassandra writes
Resolved in Pega Version 8.8
When using Customer Decision Hub with external C* deployment, attempting to stop a data flow run which was reporting stale threads for some partitions resulted in some of the partitions continuing to run while threads were waiting on the Cassandra save operation. This has been resolved by adding configuration for a timeout when waiting for the results of the save operation, with a default of 300 seconds.
INC-180762 · Issue 666291
Calendar gadget correctly displays event of >31 days
Resolved in Pega Version 8.8
When using the pxCalendar gadget to display a hotel reservation of duration > 31 days, the entire bar was not displayed on the PEGA calendar. Investigation showed that eventslist was not considering the month value, and this has been resolved by updating the pzEventPreProcessing activity to check for month.
INC-186889 · Issue 674014
Unused DSS compiledExpressions now defaults to false
Resolved in Pega Version 8.7
After updating from Pega 8.2 to Pega 8.5, executing a test campaign failed with an error in logs. This occurred when running Distribution tests with "Include funnel analysis reports" checked, and was traced to the DSS setting "decision/strategy/ssaVM/compiledExpressions". As this was previously used to toggle a feature which is no longer in use, the default setting for this DSS has now been changed to false to resolve this issue.
INC-198232 · Issue 685700
Tombstone cleanup modified to handle very large volume environments
Resolved in Pega Version 8.7
Cassandra nodes were periodically suffering from an out of memory condition. This was traced to tumbling windows creating too many tombstones for Cassandra to clean up effectively when running in a very high volume environment with a short tumbling time. This has been resolved by modifying the procedure away from using individual tombstones to instead use range tombstones which will continuously supersede the previous tombstone so there is never an opportunity for them to build up.
INC-181401 · Issue 665936
Added handling for MS SQL datestamp format
Resolved in Pega Version 8.7
When using an Microsoft SQL Server, the browse operation was failing with an SQL error for the pyIHSummary dataset. Investigation showed there were some datasets that generated SQL statements with a date timestamp format that SQL was not able to handle implicitly, affecting browsing but not functionality. This has been resolved.