INC-211100 · Issue 708300
QueueProcessor timeout made configurable
Resolved in Pega Version 8.6.4
Queue Processor timeouts were causing thread interruptions, causing items to be processed more than once. To resolve this, the QueueProcessor kafka-write timeout has been made configurable via the DASS Pega-Engine:queueprocessor/dataset/write/timeoutMs
INC-211917 · Issue 709322
Scheduled-item operations performed in shared context
Resolved in Pega Version 8.6.4
Scheduled items were not getting processed either by DelayedItemsDataFlowService data flow or by reverting to the pzDelayedQueueProcessorSchedule Job Scheduler, and the number of scheduled items in the pr_sys_delayed_queue table continued to grow even though DelayedItemsDataFlowService showed the processed record count to be increasing. Because the Admin Studio portal was not equipped to show the scheduled record count for the tenant layer in a multi-tenant environment, only the scheduled record count for the shared layer was shown. This has been resolved by updating the system to perform scheduled-item operations in shared context.
INC-213610 · Issue 708113
Job Scheduler explicitly unlocked after nodes restart
Resolved in Pega Version 8.6.4
If a job scheduler was in running state and the utility nodes were restarted, the background processing nodes were not coming up and resuming as expected and PersistentJobCleanupFactory:PresentMembersJobCleanup was throwing a stale execution exception. Investigation showed locks were not being removed from the database for the job scheduler when the restart was performed (noted in pegadata.pr_sys_locks table), preventing further runs. This has been resolved by adding an explicit lock removal process for this condition.
INC-213763 · Issue 708109
Job Scheduler explicitly unlocked after nodes restart
Resolved in Pega Version 8.6.4
If a job scheduler was in running state and the utility nodes were restarted, the background processing nodes were not coming up and resuming as expected and PersistentJobCleanupFactory:PresentMembersJobCleanup was throwing a stale execution exception. Investigation showed locks were not being removed from the database for the job scheduler when the restart was performed (noted in pegadata.pr_sys_locks table), preventing further runs. This has been resolved by adding an explicit lock removal process for this condition.
SR-132637 · Issue 203591
Added path checking to FirstLogCreationTime.log
Resolved in Pega Version 7.1.9
FirstLogCreationTime.log was being written to a temp folder using hard-coded logic to use a relative path within the install directory. This caused errors if the installer was trying to write the logs in an absolute path. FirstLogCreationTime.log should be written to the same directory as the rest of the logs, and this issue has been fixed by adding a condition in LogDeleteUtil.getLogCreationTimeFile() method to check for absolute paths.
SR-133747 · Issue 203813
Corrected hashtag handling for Oracle SQL generation
Resolved in Pega Version 7.1.9
Given a column name with a hashtag in an Oracle environment, the system was compiling SQL statements which caused ORA-00904 errors due to replacing the # in one variable name with an _ but leaving it as a # in another. This has been corrected.
SR-133057 · Issue 205124
System updated to better handle multi-node rule resolution
Resolved in Pega Version 7.1.9
In an upgraded multi-node environment, the error 'com.pega.pegarules.pub.database.MultipleRuleVersionException: Rule resolution identified 2 versions of the rule' was sporadically appearing. Once the issue occurred, it would persist and the only way to resolve it was to bounce the server. This was an issue with an instance of one rule being found in one node and not found in other node, and was related to the trigger logic. This should not occur with the current Pega7 version, as trigger logic has been moved into the engine and there is no longer a database trigger to fix. In order to ensure continued smooth trigger resolution, tests have been added that will check for this condition in the updates cache table to ensure that this scenario does not occur again
SR-A1479 · Issue 205900
UpgradeDataSchema target sets resume.action correctly
Resolved in Pega Version 7.1.9
Attempting to resume an upgrade after correcting a local error failed due to resume.properties having the default resume.action=install instead of =upgrade, causing the system to attempt to create tables in the data schema which already existed. The default of "install" that is set for resume mode has been removed, and the install process will now specifically set it when appropriate. A check has also been added to ensure the flag is set properly for the work being done.
SR-A1479 · Issue 205896
UpgradeDataSchema target sets resume.action correctly
Resolved in Pega Version 7.1.9
Attempting to resume an upgrade after correcting a local error failed due to resume.properties having the default resume.action=install instead of =upgrade, causing the system to attempt to create tables in the data schema which already existed. The default of "install" that is set for resume mode has been removed, and the install process will now specifically set it when appropriate. A check has also been added to ensure the flag is set properly for the work being done.
SR-A1479 · Issue 205895
UpgradeDataSchema target sets resume.action correctly
Resolved in Pega Version 7.1.9
Attempting to resume an upgrade after correcting a local error failed due to resume.properties having the default resume.action=install instead of =upgrade, causing the system to attempt to create tables in the data schema which already existed. The default of "install" that is set for resume mode has been removed, and the install process will now specifically set it when appropriate. A check has also been added to ensure the flag is set properly for the work being done.