Table of Contents

Search

  1. Preface
  2. Introduction to Informatica Big Data Management
  3. Mappings in the Hadoop Environment
  4. Mapping Sources in the Hadoop Environment
  5. Mapping Targets in the Hadoop Environment
  6. Mapping Transformations in the Hadoop Environment
  7. Processing Hierarchical Data on the Spark Engine
  8. Configuring Transformations to Process Hierarchical Data
  9. Processing Unstructured and Semi-structured Data with an Intelligent Structure Model
  10. Stateful Computing on the Spark Engine
  11. Monitoring Mappings in the Hadoop Environment
  12. Mappings in the Native Environment
  13. Profiles
  14. Native Environment Optimization
  15. Cluster Workflows
  16. Connections
  17. Data Type Reference
  18. Function Reference
  19. Parameter Reference

Cluster Environment Variables

Cluster Environment Variables

Cluster Environment Variables property lists the environment variables that the cluster uses. Each environment variable contains a name and a value. You can add environment variables or edit environment variables.
To edit the property in the text box, use the following format with &: to separate each name-value pair:
<name1>=<value1>[&:<name2>=<value2>…&:<nameN>=<valueN>]
Configure the following environment variables in the
Cluster Environment Variables
property:
HADOOP_NODE_JDK_HOME
Represents the directory from which you run the cluster services and the JDK version that the cluster nodes use. Required to run the Java transformation in the Hadoop environment and Sqoop mappings on the Blaze engine. You must use JDK version 1.7 or later. Default is /usr/java/default. The JDK version that the Data Integration Service uses must be compatible with the JRE version on the cluster.
Set to <cluster JDK home>/jdk<version>.
For example,
HADOOP_NODE_JDK_HOME=<cluster JDK home>/jdk<version>
.
DB2_HOME
Specifies the DB2 home directory. Required to run mappings with DB2 sources and targets on the Hive engine.
Set to /databases/db2<version>.
For example,
DB2_HOME=/databases/db2V10.5_64BIT
.
DB2INSTANCE
Specifies the DB2 database instance name. Required to run mappings with DB2 sources and targets on the Hive engine.
Set to <DB2 instance name>.
For example,
DB2INSTANCE=db10inst
.
DB2CODEPAGE
Specifies the code page configured in the DB2 instance. Required to run mappings with DB2 sources and targets on the Hive engine.
Set to <DB2 instance code page>.
For example,
DB2CODEPAGE="1208"
.
GPHOME_LOADERS
Represents the directory to the Greenplum libraries. Required to run Greenplum mappings on the Hive engine.
Set to <Greenplum libraries directory>.
For example,
GPHOME_LOADERS=opt/thirdparty/
.
PYTHONPATH
Represents the directory to the Python path libraries. Required to run Greenplum mappings on the Hive engine.
Set to <Python path libraries directory>.
For example,
PYTHONPATH=$GPHOME_LOADERS/bin/ext
.
NZ_HOME
Represents the directory that contains the Netezza client libraries. Required to run Netezza mappings on the Hive or Blaze engine.
Set to <Netezza client library directory>.
For example,
NZ_HOME=/opt/thirdparty/netezza
.
NZ_ODBC_INI_PATH
Represents the directory that contains the odbc.ini file. Required to run Netezza mappings on the Hive or Blaze engine.
Set to <odbc.ini file path>.
For example,
NZ_ODBC_INI_PATH=/opt/ODBCINI
.
ODBCINI
Represents the path and file name of the odbc.ini file.
  • Required to run Netezza mappings on the Hive or Blaze engine.
    Set to <odbc.ini file path>/<file name>.
    For example,
    ODBCINI=/opt/ODBCINI/odbc.ini
    .
  • Required to run mappings with ODBC sources and targets on the Hive engine.
    Set to <odbc.ini file path>/<file name>.
    For example,
    ODBCINI=$HADOOP_NODE_INFA_HOME/ODBC7.1/odbc.ini
    .
ODBC_HOME
Specifies the ODBC home directory. Required to run mappings with ODBC sources and targets on the Hive engine.
Set to <odbc home directory>.
For example,
ODBC_HOME=$HADOOP_NODE_INFA_HOME/ODBC7.1
.
ORACLE_HOME
Specifies the Oracle home directory. Required to run mappings with Oracle sources and targets on the Hive engine.
Set to <Oracle home directory>.
For example,
ORACLE_HOME=/databases/oracle12.1.0_64BIT
.
TNS_ADMIN
Specifies the directory to the Oracle client
tnsnames.ora
configuration files. Required to run mappings with Oracle sources and targets on the Hive engine.
Set to <tnsnames.ora config files directory>.
For example,
TNS_ADMIN=/opt/ora_tns
.
HADOOP_CLASSPATH
Represents the directory to the TDCH libraries. Required to run Teradata mappings through TDCH on the Hive engine.
Set to <TDCH libraries directory>.
For example,
/opt/cloudera/parcels/CDH-5.13.0-1.cdh5.13.0.p0.29/lib/hive/conf /opt/cloudera/parcels/CDH-5.13.0-1.cdh5.13.0.p0.29/lib/hive/lib/* /usr/lib/tdch/1.5/lib/*


Updated October 23, 2019