Table of Contents

Search

  1. Preface
  2. Command Line Programs and Utilities
  3. Installing and Configuring Command Line Utilities
  4. Using the Command Line Programs
  5. Environment Variables for Command Line Programs
  6. Using infacmd
  7. infacmd as Command Reference
  8. infacmd aud Command Reference
  9. Infacmd bg Command Reference
  10. infacmd cms Command Reference
  11. infacmd dis Command Reference
  12. Infacmd es Command Reference
  13. infacmd ipc Command Reference
  14. infacmd isp Command Reference
  15. infacmd mrs Command Reference
  16. infacmd ms Command Reference
  17. infacmd oie Command Reference
  18. infacmd ps Command Reference
  19. infacmd pwx Command Reference
  20. infacmd rms Command Reference
  21. infacmd rtm Command Reference
  22. infacmd sch Command Reference
  23. infacmd search Command Reference
  24. infacmd sql Command Reference
  25. infacmd tdm Command Reference
  26. infacmd wfs Command Reference
  27. infacmd ws Command Reference
  28. infacmd xrf Command Reference
  29. infacmd Control Files
  30. infasetup Command Reference
  31. pmcmd Command Reference
  32. pmrep Command Reference
  33. Working with pmrep Files

Hive Connection Options

Hive Connection Options

Use connection options to define a Hive connection.
Enter connection options in the following format:
... -o option_name='value' option_name='value' ...
To enter multiple options, separate them with a space.
The following table describes Hive connection options for infacmd isp CreateConnection and UpdateConnection commands that you configure when you want to use the Hive connection:
Option
Description
connectionType
Required. Type of connection is HIVE.
name
The name of the connection. The name is not case sensitive and must be unique within the domain. You can change this property after you create the connection. The name cannot exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
relationalSourceAndTarget
Hive connection mode. Set this option to true if you want to use the connection to access the Hive data warehouse. If you want to access Hive target, you need to enable the same connection or another Hive connection to run the mapping in the Hadoop cluster.
If you enable relational source and target, you must provide the metadataDatabaseString option.
pushDownMode
Hive connection mode. Set this option to true if you want to use the connection to run mappings in the Hadoop cluster.
If you enable the connection for pushdown mode, you must provide the options to run the Informatica mappings in the Hadoop cluster.
environmentSQL
SQL commands to set the Hadoop environment. In native environment type, the Data Integration Service executes the environment SQL each time it creates a connection to Hive metastore. If the Hive connection is used to run mappings in the Hadoop cluster, the Data Integration Service executes the environment SQL at the beginning of each Hive session.
The following rules and guidelines apply to the usage of environment SQL in both the connection modes:
  • Use the environment SQL to specify Hive queries.
  • Use the environment SQL to set the classpath for Hive user-defined functions and then use either environment SQL or PreSQL to specify the Hive user-defined functions. You cannot use PreSQL in the data object properties to specify the classpath. The path must be the fully qualified path to the JAR files used for user-defined functions. Set the parameter hive.aux.jars.path with all the entries in infapdo.aux.jars.path and the path to the JAR files for user-defined functions.
  • You can also use environment SQL to define Hadoop or Hive parameters that you intend to use in the PreSQL commands or in custom queries.
If the Hive connection is used to run mappings in the Hadoop cluster, only the environment SQL of the Hive connection is executed. The different environment SQL commands for the connections of the Hive source or target are not executed, even if the Hive sources and targets are on different clusters.
enableQuotes
Encloses all database reserved words in quotes. Default is false.

Properties to Access Hive as Source or Target

The following table describes the mandatory options for infacmd isp CreateConnection and UpdateConnection commands that you configure when you want to use the Hive connection to access Hive data:
Property
Description
metadataConnString
The JDBC connection URI used to access the metadata from the Hadoop server.
The connection string uses the following format:
jdbc:hive://<hostname>:<port>/<db>
Where
  • hostname
    is name or IP address of the machine on which the Hive server is running.
  • port
    is the port on which the Hive server is listening.
  • db
    is the database to which you want to connect. If you do not provide the database details, the Data Integration Service uses the default database details.
bypassHiveJDBCServer
JDBC driver mode. Enable this option to use the embedded JDBC driver (embedded mode).
To use the JDBC embedded mode, perform the following tasks:
  • Verify that Hive client and Informatica Services are installed on the same machine.
  • Configure the Hive connection properties to run mappings in the Hadoop cluster.
If you choose the non-embedded mode, you must configure the Data Access Connection String.
The JDBC embedded mode is preferred to the non-embedded mode.
connectString
The connection string used to access data from the Hadoop data store. The non-embedded JDBC mode connection string must be in the following format:
jdbc:hive://<hostname>:<port>/<db>
Where
  • hostname
    is name or IP address of the machine on which the Hive server is running.
  • port
    is the port on which the Hive server is listening. Default is 10000.
  • db
    is the database to which you want to connect. If you do not provide the database details, the Data Integration Service uses the default database details.

Properties to Run Mappings in the Hadoop Cluster

The following table describes the mandatory options for infacmd isp CreateConnection and UpdateConnection commands that you configure when you want to use the Hive connection to run Informatica mappings in the Hadoop cluster:
Property
Description
databaseName
Namespace for tables. Use the name
default
for tables that do not have a specified database name.
defaultFSURI
The URI to access the default Hadoop Distributed File System.
The FS URI must be in the following format:
hdfs://<node name>:<port>
Where
  • node name
    is the host name or IP address of the NameNode.
  • port
    is the port on which the NameNode listens for remote procedure calls (RPC).
jobTrackerURI
The service within Hadoop that submits the MapReduce tasks to specific nodes in the cluster.
JobTracker URI must be in the following format:
<jobtrackername>:<port>
Where
  • jobtrackername
    is the host name or IP address of the JobTracker.
  • port
    is the port on which the JobTracker listens for remote procedure calls (RPC).
hiveWarehouseDirectoryOnHDFS
The absolute HDFS file path of the default database for the warehouse, which is local to the cluster. For example, the following file path specifies a local warehouse:
/user/hive/warehouse
metastoreExecutionMode
Controls whether to connect to a remote metastore or a local metastore. For a local metastore, you must specify the Metastore Database URI, Driver, Username, and Password. For a remote metastore, you must specify only the
Remote Metastore URI
.
metastoreDatabaseURI
The JDBC connection URI used to access the data store in a local metastore setup. The URI must be in the following format:
jdbc:<datastore type>://<node name>:<port>/<database name>
where
  • node name
    is the host name or IP address of the data store.
  • data store type
    is the type of the data store.
  • port
    is the port on which the data store listens for remote procedure calls (RPC).
  • database name
    is the name of the database.
For example, the following URI specifies a local metastore that uses MySQL as a data store:
jdbc:mysql://hostname23:3306/metastore
metastoreDatabaseDriver
Driver class name for the JDBC data store. For example, the following class name specifies a MySQL driver:
com.mysql.jdbc.Driver
metastoreDatabaseUserName
The metastore database user name.
metastoreDatabasePassword
The password for the metastore user name.
remoteMetastoreURI
The metastore URI used to access metadata in a remote metastore setup. For a remote metastore, you must specify the Thrift server details.
The URI must be in the following format:
thrift://<hostname>:<port>
Where
  • hostname
    is name or IP address of the Thrift metastore server.
  • port
    is the port on which the Thrift server is listening.


Updated April 22, 2019