The case archiving process
Understand the case archiving and data expunging processes so you can manage your
historical case data and thereby optimise your application database performance. Before you can begin archiving case data, plan and set up
your case archiving process. In Pega Cloud deployments, Pega Platform automaticaly uses Pega Cloud File storage (PCFS) to store archived data; in
virtual-machine-based deployments, you must configure the
secondary storage repository before you can archive data.
For details, see Secondary storage repository for archived data. Planning your case archiving process Configure the case archiving process Your
case archiving policy configures the number of days that
cases must be resolved after which they are eligible for
both archiving and deletion. Schedule case archiving jobs. Search, review, recover, or clone the case data whenever you
need. The archival process archives certain artifacts within a case, such as work history
and attachments. The following table shows the artifacts that Pega Platform
archives. The pyPegaArchiver Job Scheduler (default
short description: Archival_Copier) copies
files to secondary storage by using the following steps: Exclude cases from the archival process to prevent them from being moved to the
secondary storage repository and subsequently deleted from the secondary repository. For
example, exclude a case that should not be archived due to a legal hold. For more information about excluding cases from the archive and purge process, see Excluding cases from archival and expunge.Case archiving process
The case archiving
process in Pega Platform consists of a set of jobs that move case
data to an external repository as defined in your case archiving policy. After the
system completes the case archiving process, you can still search and review cases in
the external repository as illustrated in the following diagram.Case archiving process details
Step Task or action More information Configure the case archiving process Define the case archiving policy for each case type. Defining the archiving policy for case types. Define dynamic system settings to configure: Settings that control case archiving processes Schedule case archiving jobs. Configure the pyPegaArchiverpyPegaPurger and
pyPegaIndexer job schedulers for case
archiving as needed for your business purposes. Run case archiving jobs Monitor process. Archival and expunge job statistics Review log files. Archival jobs log file entries Complete case archiving The system copies eligible cases to the secondary storage
repository. Schedule case archiving jobs Elasticsearch indexes the copied cases. The system removes the copied cases removed from the Pega Platform database. Access archived case data Retrieving archived case data Archived data
Archived case artifacts
Archived artifacts Non-archived artifacts Case archival jobs
Case archival pipeline explained
Job or activity Events pyPegaArchiver pyPegaIndexer The pyPegaIndexer Job Scheduler (default
short description: Archival_Indexer) indexes
the copied files into Elasticsearch. The index keeps the association
between an archived case and its archived file in the secondary
storage. pyPegaPurger The pyPegaPurger Job Scheduler (default
short description: Archival_Purger) deletes
cases and their associated data from the primary database. The job
also integrates a SQL VACUUM command to process deleted space and
reclaim the irrelevant empty database tables. Optional: pyArchival_ReIndexer The pyArchival_ReIndexer (default short
description: Archival_ReIndexer) Job Scheduler
fixes corrupted Elasticsearch indexes. This job follows a case
archival and purge job when trying to fix case archives. Case exclusions
Previous topic Archiving and expunging case data Next topic Secondary storage repository for archived data