Data Engineering Integration
- Data Engineering Integration 10.5.5
- All Products
Property
| Description
|
---|---|
Cluster Staging Directory
| The directory on the cluster where the Data Integration Service pushes the binaries to integrate the native and non-native environments and to store temporary files during processing. Default is
/tmp .
|
Hadoop Staging User
| The HDFS user that performs operations on the Hadoop staging directory. The user requires write permissions on Hadoop staging directory. Default is the operating system user that starts the Informatica daemon.
|
Custom Hadoop OS Path
| The local path to the Informatica server binaries compatible with the Hadoop operating system. Required when the Hadoop cluster and the Data Integration Service are on different supported operating systems. The Data Integration Service uses the binaries in this directory to integrate the domain with the Hadoop cluster. The Data Integration Service can synchronize the following operating systems:
SUSE and Redhat
Include the source directory in the path. For example,
<Informatica server binaries>/source .
Changes take effect after you recycle the Data Integration Service.
When you install an Informatica EBF, you must also install it in this directory.
|
Hadoop Kerberos Service Principal Name
| Service Principal Name (SPN) of the Data Integration Service to connect to a Hadoop cluster that uses Kerberos authentication.
Not required for the MapR distribution.
|
Hadoop Kerberos Keytab
| The file path to the Kerberos keytab file on the machine on which the Data Integration Service runs.
Not required for the MapR distribution.
|
Custom Properties
| Properties that are unique to specific environments.
You can configure run-time properties for the Hadoop environment in the Data Integration Service, the Hadoop connection, and in the mapping. You can override a property configured at a high level by setting the value at a lower level. For example, if you configure a property in the Data Integration Service custom properties, you can override it in the Hadoop connection or in the mapping. The Data Integration Service processes property overrides based on the following priorities:
When a mapping uses Hive Server 2 to run a job or parts of a job, you cannot override properties that are configured on the cluster level in preSQL or post-SQL queries or SQL override statements.
Workaround: Instead of attempting to use the cluster configuration on the domain to override cluster properties, pass the override settings to the JDBC URL. For example:
beeline -u "jdbc:hive2://<domain host>:<port_number>/tpch_text_100" --hiveconf hive.execution.engine=tez
|