Table of Contents


  1. Preface
  2. Introduction to Informatica Big Data Management
  3. Connections
  4. Mappings in the Hadoop Environment
  5. Mapping Objects in the Hadoop Environment
  6. Processing Hierarchical Data on the Spark Engine
  7. Stateful Computing on the Spark Engine
  8. Monitoring Mappings in the Hadoop Environment
  9. Mappings in the Native Environment
  10. Profiles
  11. Native Environment Optimization
  12. Data Type Reference
  13. Complex File Data Object Properties
  14. Function Reference
  15. Parameter Reference

Configuring a Mapping to Run in a Hadoop Environment

Configuring a Mapping to Run in a Hadoop Environment

You can configure a mapping to run in a Hadoop environment. To configure a mapping, you must select the Hadoop validation environment and a Hadoop connection.
  1. Select a mapping from a project or folder from the
    Object Explorer
    view to open in the editor.
  2. In the
    view, select the
  3. Select
    as the value for the validation environment.
    The Hive on MapReduce, the Blaze, and the Spark engines are selected by default. To only use the Hive on MapReduce engine, clear the other engines. If you use the Blaze engine, you cannot clear the Hive on MapReduce engine.
  4. In the execution environment, select
  5. In the Hadoop environment, select
    and use the drop down in the value field to browse for a connection or create a connection parameter:
    • To select a connection, click
      and select a connection.
    • To create a connection parameter, click
      Assign Parameter
  6. Optionally, select
    Execution Parameters
    to override a Hadoop custom property or a Spark default configuration parameter.
  7. Right-click an empty area in the editor and click
    The Developer tool validates the mapping.
  8. View validation errors on the
    Validation Log
  9. Click the
    Data Viewer
  10. Click
    Show Execution Plan
    to view the execution plan for the mapping.

Updated December 13, 2018