INC-218145 · Issue 715680
DSS introduced to control DSM clipboard page serialization
Resolved in Pega Version 8.6.5
When using a Kafka dataset to consume a message from an external topic that had an attribute name with a special character contained in a page list structure, using a JSON data transform for the mapping in a realtime dataflow resulted in the error "Exception in stage: KafkaDS; LegacyModelAspectInvokableRuleContainer.invoke-Exception encountered a :java.lang.UnsupportedOperationException." To resolve this, a new DSS dataset/CLASS_NAME/DATASET_NAME/JSONDataTransform/deserialization/useDSMPage has been introduced. When the value is set to true, the process will follow the previous behavior of DSM clipboard pages being generated when Kafka records are deserialized using JSON data transform. When the value is set to false, the JSON data transform will generate regular clipboard pages and convert them later to DSM clipboard pages. This would avoid errors when a JSON data transform calls methods from the Clipboard API that are not implemented by DSM pages. This DSS is set per data set instance. CLASS_NAME and DATASET_NAME are placeholders which should be replaced by data set's pyClassName and pyPurpose property values. In addition, a similar DSS, dataset/CLASS_NAME/DATASET_NAME/JSONDataTransform/serialization/useDSMPage, has been introduced for serialization.
INC-218757 · Issue 714553
AESRemote updated to be asynchronous
Resolved in Pega Version 8.6.5
The Autonomic Event Services (AES) agent PushCDHMetrics became stuck and was not pushing metrics to the console. This has been resolved by updating AESRemote to be asynchronous.
INC-219186 · Issue 720470
Prefix allowed for aggregates Keyspace
Resolved in Pega Version 8.6.5
Attempting to connect multiple Pega environments to an external Cassandra cluster with limited database access using the prconfig setting dnode/keyspaces_prefix resulted in the error "Unable to determine if cassandra table [aggregates.config] exists" while connecting to the aggregates Keyspace. This was traced to an incomplete implementation: the system allowed the keyspace to be created with the prefix, but queries used the keyspace name without the prefix. In order to support this use, an update has been made to ensure keyspace prefixes are used if present.
INC-220174 · Issue 716384
Improved cleanup for data from joining nodes
Resolved in Pega Version 8.6.5
A Cassandra node that is down for longer than the grace period (default 10 days) can introduce zombie data that creates instability when it returns to the cluster. This can include VBD partition summary data that was deleted and can break loading rules and cause the service to fail to start up. To resolve this, additional logic has been implemented to detect zombie summary records, including summary records without field descriptors and summary records without dictionaries having already been provided, and to read dictionaries from the latest summary record in case there are preceding zombie records with dictionaries.
INC-222561 · Issue 721043
Check added for destination type for distribution test reports
Resolved in Pega Version 8.6.5
When there were two output destinations in the system, one of type VBD and another of type Database table and both had the same name, an incorrect class was set for distribution test reports and an error was generated when trying to open the report. Investigation showed the system was only checking for the name of the destination and not its type; this has been resolved by adding a pzSetSimulationOutputClass data transform to check for the destination type in addition to the destination name when setting the class for reports.
INC-224038 · Issue 723258
Performance improvement for serviceregistry table queries
Resolved in Pega Version 8.6.5
The expression filter ($1 - "SSR"."pytimeout") in the queries for service registry table caused full table scans which might result in performance issue and even risk of contention/locking. This has been resolved by replacing the pyTimeout column reference in filters to the default value = 9000ms so that performance for these queries can be improved.
INC-202111 · Issue 710106
Logging extended for PRPCPropertyInfoProvider
Resolved in Pega Version 8.7.3
In order to assist with diagnosing issues with Kafka and JSON, additional logging has been added for PRPCPropertyInfoProvider.
INC-208976 · Issue 719165
Enhanced SSA metrics made available
Resolved in Pega Version 8.7.3
In order to better diagnose delays related to the time when a Campaign is scheduled to start and the time when the Dataflow actually starts to run, an update has been made which will generate detailed metrics to cover some of the strategy execution key performance intensive areas. Additional lower level internal metrics related to SSA engine execution have also been made available by way of a DSS to collect more runtime insight for diagnosis. To enable the collection of these Level 2 SSA internal metrics, set the dataflow/shape/strategy/detailed_metrics/level2 DSS in the Pega-DecisionEngine rule set to 'true'. A comprehensive set of enhanced metrics will be available in Pega 8.8.
INC-217290 · Issue 721375
Added support for creating predictive models in Production
Resolved in Pega Version 8.7.3
While creating a new predictive model rule in Prediction studio, the case was going into broken process after selecting the template with the error message "Error loading D_ProjectList , Reason : No databases defined in properties file:/databases.properties". This was an unexpected use case for creating models in Production level, and has been resolved by updating the flows to turn off the draft mode in this scenario.
INC-218145 · Issue 715678
DSS introduced to control DSM clipboard page serialization
Resolved in Pega Version 8.7.3
When using a Kafka dataset to consume a message from an external topic that had an attribute name with a special character contained in a page list structure, using a JSON data transform for the mapping in a realtime dataflow resulted in the error "Exception in stage: KafkaDS; LegacyModelAspectInvokableRuleContainer.invoke-Exception encountered a :java.lang.UnsupportedOperationException." To resolve this, a new DSS dataset/CLASS_NAME/DATASET_NAME/JSONDataTransform/deserialization/useDSMPage has been introduced. When the value is set to true, the process will follow the previous behavior of DSM clipboard pages being generated when Kafka records are deserialized using JSON data transform. When the value is set to false, the JSON data transform will generate regular clipboard pages and convert them later to DSM clipboard pages. This would avoid errors when a JSON data transform calls methods from the Clipboard API that are not implemented by DSM pages. This DSS is set per data set instance. CLASS_NAME and DATASET_NAME are placeholders which should be replaced by data set's pyClassName and pyPurpose property values. In addition, a similar DSS, dataset/CLASS_NAME/DATASET_NAME/JSONDataTransform/serialization/useDSMPage, has been introduced for serialization.