Creating a Stream data set
Process a continuous data stream of events (records) by creating a Stream
data set.
You can test how data flow processing is distributed across Data
Flow service nodes in a multinode Decision Management environment by specifying the
partition keys for Stream data set and by using the load balancer provided by Pega. For
example, you can test whether the intended number and type of partitions negatively affect
the processing of a Data Flow rule that references an event strategy.
- In the header of Dev Studio, click Create > Data Model > Data Set.
-
In the Data Set Record Configuration section of the
Create Data Set tab, define the data set by performing
the following actions:
-
In the Label field, enter the data set
label.
Result: The identifier is automatically created based on the data set label.
- Optional: To change the automatically created identifier, click Edit, enter an identifier name, and then click OK.
- From the Type list, select Stream.
-
In the Label field, enter the data set
label.
- In the Context section, specify the ruleset, applicable class, and ruleset version of the data set.
- Click Create and open.
- Optional:
To create partition keys for testing purposes, in the
Stream tab, in the Partition
key(s) section, perform the following actions:
Note: Create partition keys for Stream data sets only in application environments where the production level is set to 1 - Sandbox, 2 - Development, or 3 - Quality assurance. For more information, see Specifying the production level.
- Click Add key.
-
In the Key field, press the Down arrow key, and
then select a property to use as a partition key.
The available properties are based on the applicable class of the data set which you defined in step 3.
- To add more partition keys, repeat steps 5.a through 5.b.
For more information on when and how to use partition keys in a Stream data set, see Partition keys for Stream data sets. - Optional:
To specify additional settings for your Stream data set, in the
Settings tab, perform the following actions:
Note: After saving the data set, you will not be able to change these settings.
-
In the Authentication section, specify if you
want the data set to require basic authentication.
The REST and WebSocket endpoints are secured by using the Pega Platform common authentication scheme. Each post to the stream requires authenticating with your user name and password. By default, the Enable basic authentication check box is selected.
-
In the Maximum storage field, enter the maximum
amount of data to store in GB.
The default amount is 20 GB.
-
In the Retention period field, specify for how
long you want to keep records in this data set.
The default value is 1 day.
-
In the Authentication section, specify if you
want the data set to require basic authentication.
- Click Save.
- Optional:
To populate the Stream data set with external data, perform one of the
following actions:
Choice Action Use an existing Pega REST service - In the navigation panel of Dev Studio, click Records > Integration-Connectors > Connect REST.
- Select a Pega REST service.
-
Configure the settings in the Methods tab.
For more information, see Mapping the request and response data for Connect REST rules.
Create a Pega REST service -
Create a Connect REST rule.
For more information, see Creating Connect REST rules.
-
Configure the settings in the Methods tab.
For more information, see Mapping the request and response data for Connect REST rules.