Skip to main content

         This documentation site is for previous versions. Visit our new documentation site for current releases.      

Changing the number of retries for SAVE operations in batch and real-time data flow runs

Updated on May 17, 2024

Control how many times batch and real-time data flow runs retry SAVE operations on records. With automatic retries, when a SAVE operation fails, the run can still successfully complete if the resources that were initially unavailable become operational. The run fails only when all the retries are unsuccessful.

You can control the global number of retries for SAVE operations through a dedicated dynamic system setting. If you want to change that setting for an individual batch or real-time data flow run, update a property in the integrated API.
Note: If a single record fails for Merge and Compose shapes, the entire batch run fails.
Retries trigger lifecycle events. For more information, see Event details in data flow runs on Pega Community.
  1. In the navigation pane of Dev Studio, click RecordsSysAdminDynamic System Settings.
  2. In the list of instances, search for and open the dataflow/shape/maxRetries dynamic system setting.
  3. In the dynamic system setting editing tab, in the Value field, enter the number of retries that you want to run when a SAVE operation on a record fails during a data flow run.
    The default value is 5.
What to do next: If you want to change that setting for a single batch data flow run, update the pyResilience.pyShapeMaxRetries property in the RunOptions page for the run through the integrated API. For more information, see Pega APIs and services.

Have a question? Get answers now.

Visit the Support Center to ask questions, engage in discussions, share ideas, and help others.

Did you find this content helpful?

Want to help us improve this content?

We'd prefer it if you saw us at our best. is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us