Skip to main content

Resolved Issues

View the resolved issues for a specific Platform release.

Go to download resolved issues by patch release.

Browse release notes for a selected Pega Version.

NOTE: Enter just the Case ID number (SR or INC) in order to find the associated Support Request.

Please note: beginning with the Pega Platform 8.7.4 Patch, the Resolved Issues have moved to the Support Center.

SR-128285 · Issue 193082

Added handling for column size mismatch on Oracle import

Resolved in Pega Version 7.1.8

When upgrading versions of CMF, errors were generated when importing the Pega jar file. The exported schema has pyAccessibility and pyBrowserSupportLevel datatypes as "integer" with size 10, but on the target Oracle database the same column was defined as Number with default precision and scale (0, -127). While importing, the logic checks for column compatibility and was generating an alter statement based on this scale mismatch. This has been resolved by modifying the system to detect that the metadata is returning the new style of NUMBER column metadata handling it correctly.

SR-128475 · Issue 194266

DB Trigger logic refactored for better vw table syncing

Resolved in Pega Version 7.1.8

There are many times in customer environments where for whatever reason the database triggers are disabled for a period of time and thus the vw table gets out of sync. Typically in that case the triggers are reenabled and a statement like "update pr4_rule_file set pzinskey=pzinskey" is run to get the vw table populated correctly. However, if the system was being upgraded the statement would update an existing row in the vw table but it would not add any missing rows to the vw table during migration. To ensure that all rows are handled correctly, trigger logic has been completely refactored for 7.1.8.

SR-128932 · Issue 194789

Added support for -x and -y parameters to ApplicationBundle import

Resolved in Pega Version 7.1.8

Previously, the use of the -x and -y parameters in the command line utility prpcUtils only worked with PegaArchives and not ApplicationBundle archives. In order to facilitate automation, the app bundle command line import command now honors the -x (override) and -y (skip) import parameters, and these options will be passed along to any nested par files.

SR-131199 · Issue 201548

Smoothed editing of data tables mapped to external DB

Resolved in Pega Version 7.1.8

It was not possible to edit a data table mapped to an external database table with column name different from that of the property. To resolve this, the system will ensure that the external mappings take precedence while trying to update the data table.

SR-131266 · Issue 199637

Cleared exception during import of RAP by prpcUtils

Resolved in Pega Version 7.1.8

When attempting to import a RAP file through command line by using prpcUtils.sh, the build was successful and all the rules were imported but an SQL exception was generated related to the execution of Stored Procedure. The RAP file did not contain any code to execute stored procedure, so the execution was found to have been done internally by prpcUtils.sh. To prevent this, prpcUtils has been updated to use the data schema as the default schema for the connection while running against either of the DB2 dbs.

SR-131347 · Issue 198661

Changed Dependency Checker table-based cataloging to be case insensitive

Resolved in Pega Version 7.1.8

After updating, the Dependency Checker and Packager functions for v5 items in the table-base catalog features were not finding existing cataloged items that contained mixed case IDs. To resolve this, the dependency checking code has been modified to be case insensitive.

SR-132265 · Issue 200632

Improved hotfix detection

Resolved in Pega Version 7.1.8

Migrating from ML6 to ML7 was failing with an error reporting the system contained hotfixes that were generated after the release of ML7. This was due to the ML Readiness step incorrectly identifying pre-7.1 rules as having been updated after the upgrade to 7.1.6. This happened when the revalidate and save was run after the release of ML7, which updated the pxUpdateDateTime for the rules in question and led to them being caught by the validation. This was meant to catch hotfixes, and the handling has been improved to avoid this error.

INC-130304 · Issue 567924

Retry logic added for downloading upgraded rules

Resolved in Pega Version 8.1.9

A Rules upgrade failed while downloading applications from the maintenance server due to an SFTP server connection failure. This has been resolved by adding logic to retry if the first connection attempt fails.

INC-130695 · Issue 587659

Enhancements for upgrading in multi-tenant environment

Resolved in Pega Version 8.1.9

Some muti-tenant installations use the same applications or rule instances with the same pzInsKeys for different tenants. This can cause upgrades to time out due to the system fetching all pzInsKeys (which will have duplicates) and working with them in a default batch size of 500 each over 4 threads. This led to the same keys potentially being allocated and processed in different threads, resulting in duplicate processing and timeouts. This has been resolved by updating the select query to fetch the tentantid and pzInskeys in the MT system to avoid duplicate work in multiple threads. In addition, running Generate Declarative indexes fetches the pzinskeys and generates indexes for each record, but before generating, the existing index for the record is deleted and then inserted. Because the delete query to generate the index was not tenant aware, all of the records for the key were deleted for the tenants for that key, but the new index was created only in one tenant. This has been resolved by enhancing the DELETE query to be tenant aware, which will avoid deleting the indexes for all the tenants given an index key.

INC-132218 · Issue 573359

Resolved buffer overflow for Migration loadDatabase

Resolved in Pega Version 8.1.9

A Rules upgrade failed in the Migration step at loadDatabase stage, involving the move of all the table records from old schema to new schema. This was traced to the inability of Migration to load blob of sizes more than 100 MB, and has been resolved by updating Migration to use byte[] to read the blob content with the help of metadata that contains blob length.

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice
Contact us