Skip to main content

Published Release Notes

Find release notes for the selected Pega Version and Capability

Browse resolved issues for Platform releases.

This documentation is for non-current versions of Pega Platform. For current release notes, go here.

Use Kinesis data sets in Pega Decision Management

Valid from Pega Version 7.4

You can create Kinesis data set instances to connect to Amazon Kinesis Data Streams and use this data set in decision management for processing real-time streaming data. Integrating Kinesis data streams into Pega® Platform in the cloud provides a fault-tolerant and scalable solution for processing IT infrastructure log data, application logs, social media, market data feeds, and web clickstream data.

For more information, see Creating a Kinesis data set.

Additional configuration options for File data sets

Valid from Pega Version 8.2

You can now create File data sets for more advanced scenarios by adding custom Java classes for data encryption and decryption, and by defining a file set in a manifest file.

Additionally, you can improve data management by viewing detailed information in the dedicated meta file for every file that is saved, or by automatically extending the filenames with the creation date and time.

For more information, see Creating a File data set for files on repositories and Requirements for custom stream processing in File data sets.

Store and scale the processing of Stream data records on multiple nodes

Valid from Pega Version 7.4

You can configure the Stream service on Pega® Platform to ingest, route, and deliver high volumes of low-latency data such as web clicks, transactions, sensor data, and customer interaction history. You can store streams of records in a fault-tolerant way and process stream records as they occur. Add or remove Stream nodes to increase or decrease the use of the Stream service and optimize data processing.

For more information, see Stream service overview.

Label changes for text analytic models

Valid from Pega Version 7.4

The classification analysis label has changed to topic detection and the entity extraction label has changed to text extraction. Also, the sentiment analysis, topic detection, and intent detection labels are now located under Text Categorization in the Analytics Center. These name changes reflect industry standards and provide a clearer distinction between different types of text analytics models in Pega® Platform.

For more information, see Text analytics models.

Enhanced adaptive model reporting

Valid from Pega Version 7.4

The new Model report replaces the Behavior and the Performance overview reports to improve report usability and provide consistent information. You can export your Model reports into PDF or Excel files to view or share them outside the Pega® Platform. The Model report also includes information on the groups of correlated predictors where the best performing predictor from each group is active in the model and other remain inactive; this information helps you understand why predictors are active or inactive.

For more information, see Generating a model report.

Manage propositions in multitenant environments

Valid from Pega Version 7.4

Proposition Management is now supported in multitenant environments of Pega® Platform. The DisablePropCache Dynamic System Setting switches off the use of a proposition cache. When you import propositions in strategies, you can import only versioned propositions (Decision Data rules). Unversioned propositions are not supported under this setting, but these can be converted to versioned propositions.

For more information, see Unversioned propositions, Synchronization of the proposition cache, and Converting groups with unversioned propositions.

Decisioning services now use default node classification

Valid from Pega Version 7.4

Decisioning services have been integrated with default node classification on Pega® Platform to provide a unified way of creating and initializing services. As a result of the integration, the Data Flow service has been divided into Batch and Real Time services to better handle different types of data flow runs. You can now specify separate subsets of Data Flow nodes for batch data flow runs and real-time data flow runs to divide the workload between these two subsets.

For more information, see Node classification, Data Flows landing page, and Services landing page.

New interaction history attribute

Valid from Pega Version 7.4

Pega® Platform 7.4 introduces the pySubjectType attribute that is used in interaction history aggregations. This attribute is populated for interaction history records that were created in release 7.4. For records that originated in earlier releases, the attribute must be set in the following scenarios:

  • Single-level decisioning frameworks that use interaction history.
  • Multi-level decisioning frameworks where interaction history is used by two or more levels of strategies that are defined on different classes.

For the single-level scenario, configure the Dynamic System Setting that sets the pySubjectType attribute when your framework reads interaction history records. The value of this Dynamic System Setting becomes the name of the customer class.

For the multi-level scenario, update the database table for all strategy levels manually. For each level, make sure that the value in the Subject Type column is set to the name of the class for the corresponding strategy. For example, the value for the top level strategy should be set to the name of the class of that strategy.

For more information about interaction history aggregations, see Data Sources landing page

For more information about multi-line strategies and contexts, see Strategy components - Embedded strategy

Kafka custom serializer

Valid from Pega Version 8.2

In Kafka data sets, you can now create and receive messages in your custom formats, as well as in the default JSON format. To use custom logic and formats for serializing and deserializing ClipboardPage objects, create and implement a Java class. When you create a Kafka data set, you can choose to apply JSON or your custom format that uses a PegaSerde implementation.

For more information, see Creating a Kafka data set and Kafka custom serializer/deserialized implementation.

Data flow life cycle monitoring

Valid from Pega Version 8.2

You can now generate a report from the Run details section of a Data Flow rule that provides information about run events. The report includes reasons for specific events which you can analyze to troubleshoot and debug issues more quickly. You can export the report and share it with others, such as Global Customer Support.

For more information about accessing event details, see Creating a real-time run for data flows and Creating a batch run for data flows.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us