SR-D38492 · Issue 505490
Batching now enabled when using IH Summary shape in substrategies
Resolved in Pega Version 8.4
Batching was not being enabled when using IH Summary shape in substrategies. This was traced to the IH_SUMMARY_CACHE field not automatically being passed to sub-strategies, and has been corrected.
SR-D96836 · Issue 555749
Refinements made to MarkerNode memory use
Resolved in Pega Version 8.5
Significant memory usage can be observed when data join in strategies are missing the join conditions. E.g., when the primary source in the strategy can have up to 300 or even more propositions, the join without the join condition would perform a cartesian product with possibly up to 400 or even more records returned. This may cause performance degradation. To guard against this, an update has been made that seeks to prevent repetitive markers being accumulated under a MarkerNode while trying not to rely on any implementation of equality and hashCode for individual marker implementations.
SR-D39956 · Issue 511638
Corrected method IF use with shortcut function
Resolved in Pega Version 8.4
After upgrade, method IF was not working as expected when used in an expression like "@if(.totalorders_120days>0,(.remakeorders_120days/.totalorders_120days)<0.3,false)". This was caused by a missed use case for the combination of an exception-generating function in combination with a shortcut function (i.e. ternary, and, or), and has been resolved.
INC-195387 · Issue 681673
HandleResponses updated for use with optimized campaign
Resolved in Pega Version 8.7
After updating from Pega 8.4, responses were not processed in the pxHandleResponses Data flow if the "pyCampaignOptimization" flag was set to true. This was traced to a missed use case for an optimized campaign, and has been resolved by modifying the condition in HandleResponses.
SR-D42451 · Issue 518068
ExecuteRDB call updated to use NativeSQL for blob
Resolved in Pega Version 8.4
After creating a test activity to clear data set records that used the DataSet-Execute method and passed the data set name and truncate operation, only 51 records were deleted in a single run when the data set had more than 51 records. Investigation showed that for blob tables, the database truncate operation was using executeRDB with an empty results page, i.e. it didn't specify pyMaxRecords, which on some databases might have limited the number affected records. To resolve this, the executeRDB call in the database truncate operation has been modified to use NativeSQL for blob tables.
INC-156818 · Issue 628465
Materialization uses time limit boundary for query
Resolved in Pega Version 8.4.5
After turning on Materialization for pyIHSummary and OfferOutcomesForPast45Days datasets, an SQL query was taking an excessive amount of time and causing multiple alerts in the logs. Investigation traced the issue to database partitioning, specifically that running a query where the pyOutcomeTime range spanned multiple partitions was causing the indexes for all partitions in the range to be opened. To resolve this, the query has been updated with a DSS to support a partition size of min(pxOutcomeTime) to limit the time range to querying day by day, or hour by hour, or any other chronology unit specified. If there are no records for the current limit, then it will look at the next partition. This should prevent the query from needing to open more than 1 or 2 partitions.
SR-D89428 · Issue 550392
Data Flow StartTime uses locale timezone
Resolved in Pega Version 8.5
The start time of the dataflow was displayed in GMT instead of the operator locale timezone. This has been corrected.
SR-D77157 · Issue 544472
DataSet preview will use date instead of datetime
Resolved in Pega Version 8.5
While using a DataSet preview functionality, the date appeared as reduced by one day. This has been resolved by parsing date as 'date' instead of 'datetime' to avoid issues with timezone interactions.
INC-158813 · Issue 629484
Updated report handling for simulations using database output
Resolved in Pega Version 8.4.5
When running a simulation with a database table as the output destination, the pxObjClass property was not populating with a value in the results. This caused sub-reports to not be populated with data. Investigation showed that issue happened when the simulation output destination database table was inferred as an external table due to not having an exposed column for pzinskey. To resolve this, the pxObjClass and pxCreateDateTime properties, which were added to simulation output destination tables for compatibility with Report browser, will not be added to those tables when they are created. Instead, to address compatibility issues of simulation output destination classes with Report browse in the Customer Decision Hub, the pyDefaultSummaryReport has been brought up into the Data-Decision-StrategyExecution-ResultOutput class from the @baseclass.
INC-197530 · Issue 686027
Value Finder updated for use with external Cassandra nodes
Resolved in Pega Version 8.7
Attempting to use the Value Finder feature resulted in the error "Running Simulation is not possible because the required services are not available. Contact your System Administrator to enable the data flow and decisioning data store services." Analysis traced this to a check which identifies whether there is more than one internal node available for CDH/ DSS node. Since there were only external nodes available in this scenario and no internal nodes, the method returned false and returned the error when the CDH / VF page was launched. To resolve this, the check has been modified to allow for external Cassandra.