Table of Contents

Search

  1. Preface
  2. Introduction to Intelligent Streaming
  3. Intelligent Streaming Configuration
  4. Connections
  5. Sources and Targets in a Streaming Mapping
  6. Intelligent Streaming Mappings
  7. Window Transformation
  8. Data Type Reference

Big Data Streaming User Guide

Big Data Streaming User Guide

Mapping Configurations

Mapping Configurations

To configure a mapping, configure the connection and run-time properties for the mapping.
When you configure the mapping, configure the following run-time properties:
Validation Environment
The environment in which the validations are done. Select Hadoop in the validation environment and select the Spark engine. The Data Integration Service pushes the mapping logic to the Spark engine..
Execution Environment
The environment in which the mappings are executed. Select Hadoop as the execution environment.
Hadoop
Specify the following properties for the Spark engine:
  • Connection. Select the connection to the Spark engine used for pushdown of processing. Select
    Connection
    and browse for a connection or select a connection parameter.
  • Runtime Parameters. An optional list of configuration parameters to apply to the Spark engine. You can change the default Spark configuration properties values, such as
    spark.executor.memory
    or
    spark.driver.cores
    .
    Use the following format:
    <property1>=<value>
    • <property1> is a Spark configuration property.
    • <value> is the value of the property.
    To enter multiple properties, separate each name-value pair with the following text:
    &:
    .
    If you use a JMS source or Amazon Kinesis Streams source in the mapping, configure two or more executors for the mapping. For example, use the following configuration:
    spark.executor.instances=2 &: spark.executor.cores=2 &: spark.driver.cores=1
Source Configuration
Specify the following properties to configure how the data is processed:
  • Maximum Rows Read. Specify the maximum number of rows that are read before the mapping stops running. Default is
    Read All Rows
    .
  • Maximum Runtime Interval. Specify the maximum time to run the mapping before it stops. If you set values for this property and the Maximum Rows Read property, the mapping stops running after one of the criteria is met. Default is
    Run Indefinitely
    . A value of
    Run Indefinitely
    enables the mapping to run without stopping.
  • State Store. Specify the HDFS location on the cluster to store information about the state of the Spark Job. Default is
    <Home Directory>/stateStore
    You can configure the state store as part of the configuration of execution options for the Data Integration Service.
You can use these properties to test the mapping.
Streaming Properties
Specify the following streaming properties:
  • Batch interval. The Spark engine processes the streaming data from sources and publishes the data in batches. The batch interval is number of seconds after which a batch is submitted for processing.
  • Cache refresh interval. You can cache a large lookup source or small lookup tables. When you cache the lookup source, the Data Integration Service queries the lookup cache instead of querying the lookup source for each input row. You can configure the interval for refreshing the cache used in a relational Lookup transformation.
The following image shows the connection and run-time properties:
The Run-time properties shows the Validation Environment and the Execution Environment properties.