INC-174475 · Issue 650110
Removed unnecessary commit in pyGetAttachmentsByCids
Resolved in Pega Version 8.4.5
Attempting to reply to inbound emails generated the error "This action is not allowed as it is outside the current transaction". Investigation showed a commit was occurring for pygetattachmentsbycids even when there was no change in attachment, and this interfered with the subsequent transactions and resulted in the Transaction id mismatch issue. To resolve this, the commit has been removed and the system has been updated to save only when needed for backwards compatibility.
INC-179722 · Issue 659281
Corrected race condition that caused JobScheduler error
Resolved in Pega Version 8.4.5
A race condition between the pzInitializeQueueProcessors and pyPersistNodeState job schedulers that both run at platform's start was causing a null pointer exception with the error "com.pega.platform.executor.jobscheduler.scheduler.JobExecutionException: Activity execution failed with [PRRuntimeException]". This has been resolved by moving the QueueProcessor initialization process from JobScheduler to StartupTask.
INC-143622 · Issue 602981
Improved privacy for email listener alerts
Resolved in Pega Version 8.4.5
To improve privacy, the customer 'from' address will be removed from PEGA0011 alerts for email listener processing.
INC-143668 · Issue 617487
Performance improvements for Attachment Migration Utility
Resolved in Pega Version 8.4.5
The Attachment Migration Utility was showing a zero count for attachments in the Pega Database. This was traced to the underlying report definition that retrieves data for the Attachment Migration Utility timing out, and has been resolved by updating the activities to improve performance for applications having a large number of attachments without customization.
INC-145802 · Issue 636322
Oracle temporary space explicitly freed after writing LOBs
Resolved in Pega Version 8.4.5
Oracle temporary space used to write the LOB was not freed up after writing records to Log-Service-File when using OJDBC8 or OJDBC10 versions and the table involved had inline BLOB storage such as small blob tables like Log-service-file (2-4kB blobs). This did not occur when using OJDBC7, but investigation with Oracle could not determine why the older version did not have this issue. This issue has been resolved by explicitly freeing the LOB space with a call to free() / freeTemporary() , a utility for LOB objects that was introduced in JDBC 4.
INC-150610 · Issue 627425
Argument number exception handled for SERVICES-PAL log configuration
Resolved in Pega Version 8.4.5
An issue with being unable to configure the SERVICES-PAL log file has been resolved by adding handling for an unlogged "wrong number of arguments" exception.
INC-152057 · Issue 621208
S3 attachment migration handles LInk-Attachment with multiple instances
Resolved in Pega Version 8.4.5
After S3 attachment migration, some attachments were intermittently not opening and displaying an error relating to being unable to load the file. Investigation showed that the attachments that failed to open did not have a pxStorageType tag in the XML of the work item. After the migration is done, the corresponding Data-WorkAttach-File and Link-Attachment instances are updated to point them to the repository. In this case, multiple LInk-Attachment instances were pointing to the same Data-WorkAttach-File instances, so only one Link-Attachment was updated and all of the other instances pointing to the same Data-WorkAttach-File instance were rendered unusable. While there was a workaround of manually updating the storage type in the database, this has been resolved by updating "pzgetattachmentcontent","pzupdatesourcereferences", and "pzuploadcontenttoexternalstorage" to check whether attachment is already migrated or not using the boolean parameter "isAttachmentAlreadyMigrated" to ensure all of the other Link-Attachment instances for a particular attachment are updated. If a migration was done and all Data-WorkAttach-File instances are pointing to a repository with some Link-Attachment instances not updated, those will be updated by running the migration again.
INC-153014 · Issue 625695
Handling added for missing archival class definitions
Resolved in Pega Version 8.4.5
Handling has been added to avoid suspending the archival process when a class definition no longer exists in the system. If the system does not find a class corresponding to a configured case type, the exception generated will be logged and processing will continue with the next case type.
INC-154026 · Issue 631040
Resolved exception from expanding deserialized Data Page
Resolved in Pega Version 8.4.5
When using a data page backed by a report definition where the data page data type was Rule- and the virtual page data type was Rule-Obj-Property, serializing and deserializing the data page and then expanding the deserialized data page generated the exception "the reference .pzInsKey is not valid. Reason: .pzInsKey is mapped to class Rule-Obj-Property, which is not the primary class of the list: Rule-; must specify a group name in mapping". The issue occurred only if "Run on Data Page" on the data page was checked and "Report on descendant class instances" on the report definition was unchecked, and has been resolved by updating the runLoadActivity() of DeclarativePageDirectoryImpl to skip setting pyObjClass for the report definition if it already present, i.e. if already set by the loader activity.
INC-155294 · Issue 626661
Enhancement added for attachment size handling with Kafka
Resolved in Pega Version 8.4.5
Attempting to send an email attachment larger than 2 MB resulted in the error "Caused by: org.apache.kafka.common.errors.RecordTooLargeException: The message is 8101592 bytes when serialized which is larger than the maximum request size you have configured with the max.request.size configuration." This has been resolved by updating the Stream SPI size to 2.0.5-14 to support custom producer configurations. The settings can be passed as environment parameters, for example 'Dstream.producer.max.request.size=500990'.