Data Quality message queue now supports batching

Data Quality message queue now supports batching

Until the current release Data Quality message queue requests were processed one at a time. In order to have better performance and resource utilization, these requests can now be batched within the same message queue. This leads to an performance improvement up to factor of 5, especially in scenarios where many single requests to DQ are sent in a very short period of time. E.g. an item change event that was executed on thousands of single objects.
  • Without batching, a sample data set of 20k item records was taking approximately 25 minutes to complete the execution for a rich set of Data Quality rules.
  • With batching, the same data set of 20k item records executing the same set of Data Quality rules, completed in approximately 5 minutes.
Results from the test scenario above:
Operation
10.0 HotFix 1 no batching
10.0 HotFix 1 with batching
Improvement
Sample data set 20k item records
25 minutes
5 minutes
5x
There is no change to the structure of the Data Quality request and response message. Batching can be disabled by setting
queue.default.message.batch.threshold = 1
.

0 COMMENTS

We’d like to hear from you!