Table of Contents

Search

  1. Preface
  2. Transformations
  3. Source transformation
  4. Target transformation
  5. Access Policy transformation
  6. Aggregator transformation
  7. B2B transformation
  8. Cleanse transformation
  9. Data Masking transformation
  10. Data Services transformation
  11. Deduplicate transformation
  12. Expression transformation
  13. Filter transformation
  14. Hierarchy Builder transformation
  15. Hierarchy Parser transformation
  16. Hierarchy Processor transformation
  17. Input transformation
  18. Java transformation
  19. Java transformation API reference
  20. Joiner transformation
  21. Labeler transformation
  22. Lookup transformation
  23. Machine Learning transformation
  24. Mapplet transformation
  25. Normalizer transformation
  26. Output transformation
  27. Parse transformation
  28. Python transformation
  29. Rank transformation
  30. Router transformation
  31. Rule Specification transformation
  32. Sequence transformation
  33. Sorter transformation
  34. SQL transformation
  35. Structure Parser transformation
  36. Transaction Control transformation
  37. Union transformation
  38. Velocity transformation
  39. Verifier transformation
  40. Web Services transformation

Transformations

Transformations

Advanced properties

Advanced properties

You can configure advanced properties for an Aggregator transformation. The advanced properties control settings such as the tracing level for session log messages, whether the transformation uses sorted input, cache settings, and whether the transformation is optional or required.
The properties that are available vary based on the mapping mode.
You can configure the following properties:
Property
Description
Tracing Level
Detail level of error and status messages that
Data Integration
writes in the session log. You can choose terse, normal, verbose initialization, or verbose data. Default is normal.
Sorted Input
Indicates that input data is pre-sorted by groups. Select this option only if the mapping passes sorted data to the Aggregator transformation.
Cache Directory
Local directory where
Data Integration
creates the index and data cache files.
By default,
Data Integration
uses the directory entered in the Secure Agent $PMCacheDir property for the Data Integration Server. If you enter a new directory, make sure that the directory exists and contains enough disk space for the aggregate caches.
Data Cache Size
Data cache size for the transformation. Select one of the following options:
  • Auto.
    Data Integration
    sets the cache size automatically. If you select Auto, you can also configure a maximum amount of memory for
    Data Integration
    to allocate to the cache.
  • Value. Enter the cache size in bytes.
Default is Auto.
Index Cache Size
Index cache size for the transformation. Select one of the following options:
  • Auto.
    Data Integration
    sets the cache size automatically. If you select Auto, you can also configure a maximum amount of memory for
    Data Integration
    to allocate to the cache.
  • Value. Enter the cache size in bytes.
Default is Auto.
Transformation Scope
Specifies how
Data Integration
applies the transformation logic to incoming data. Select one of the following options:
  • Transaction. Applies the transformation logic to all rows in a transaction. Choose Transaction when a row of data depends on all rows in the same transaction, but does not depend on rows in other transactions.
  • All Input. Applies the transformation logic on all incoming data. When you choose All Input,
    Data Integration
    drops incoming transaction boundaries. Choose All Input when a row of data depends on all rows in the source.
Optional
Determines whether the transformation is optional. If a transformation is optional and there are no incoming fields, the
mapping
task can run and the data can go through another branch in the data flow. If a transformation is required and there are no incoming fields, the task fails.
For example, you configure a parameter for the source connection. In one branch of the data flow, you add a transformation with a field rule so that only Date/Time data enters the transformation, and you specify that the transformation is optional. When you configure the
mapping
task, you select a source that does not have Date/Time data. The
mapping
task ignores the branch with the optional transformation, and the data flow continues through another branch of the mapping.

0 COMMENTS

We’d like to hear from you!