Table of Contents

Search

  1. Preface
  2. Introduction to Informatica Big Data Management
  3. Connections
  4. Mappings in the Hadoop Environment
  5. Mapping Objects in the Hadoop Environment
  6. Processing Hierarchical Data on the Spark Engine
  7. Stateful Computing on the Spark Engine
  8. Monitoring Mappings in the Hadoop Environment
  9. Mappings in the Native Environment
  10. Profiles
  11. Native Environment Optimization
  12. Data Type Reference
  13. Complex File Data Object Properties
  14. Function Reference
  15. Parameter Reference

JDBC Connection Properties

JDBC Connection Properties

You can use a JDBC connection to access tables in a database. You can create and manage a JDBC connection in the Administrator tool, the Developer tool, or the Analyst tool.
The order of the connection properties might vary depending on the tool where you view them.
The following table describes JDBC connection properties:
Property
Description
Database Type
The database type.
Name
Name of the connection. The name is not case sensitive and must be unique within the domain. The name cannot exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID
String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It must be 255 characters or less and must be unique in the domain. You cannot change this property after you create the connection. Default value is the connection name.
Description
The description of the connection. The description cannot exceed 765 characters.
User Name
The database user name.
If you configure Sqoop, Sqoop uses the user name that you configure in this field. If you configure the --username argument in a JDBC connection or mapping, Sqoop ignores the argument.
Password
The password for the database user name.
If you configure Sqoop, Sqoop uses the password that you configure in this field. If you configure the --password argument in a JDBC connection or mapping, Sqoop ignores the argument.
JDBC Driver Class Name
Name of the JDBC driver class.
The following list provides the driver class name that you can enter for the applicable database type:
  • DataDirect JDBC driver class name for Oracle:
    com.informatica.jdbc.oracle.OracleDriver
  • DataDirect JDBC driver class name for IBM DB2:
    com.informatica.jdbc.db2.DB2Driver
  • DataDirect JDBC driver class name for Microsoft SQL Server:
    com.informatica.jdbc.sqlserver.SQLServerDriver
  • DataDirect JDBC driver class name for Sybase ASE:
    com.informatica.jdbc.sybase.SybaseDriver
  • DataDirect JDBC driver class name for Informix:
    com.informatica.jdbc.informix.InformixDriver
  • DataDirect JDBC driver class name for MySQL:
    com.informatica.jdbc.mysql.MySQLDriver
For more information about which driver class to use with specific databases, see the vendor documentation.
Connection String
Connection string to connect to the database. Use the following connection string:
jdbc:<subprotocol>:<subname>
The following list provides sample connection strings that you can enter for the applicable database type:
  • Connection string for DataDirect Oracle JDBC driver:
    jdbc:informatica:oracle://<host>:<port>;SID=<value>
  • Connection string for Oracle JDBC driver:
    jdbc:oracle:thin:@//<host>:<port>:<SID>
  • Connection string for DataDirect IBM DB2 JDBC driver:
    jdbc:informatica:db2://<host>:<port>;DatabaseName=<value>
  • Connection string for IBM DB2 JDBC driver:
    jdbc:db2://<host>:<port>/<database_name>
  • Connection string for DataDirect Microsoft SQL Server JDBC driver:
    jdbc:informatica:sqlserver://<host>;DatabaseName=<value>
  • Connection string for Microsoft SQL Server JDBC driver:
    jdbc:sqlserver://<host>;DatabaseName=<value>
  • Connection string for Netezza JDBC driver:
    jdbc:netezza://<host>:<port>/<database_name>
  • Connection string for Pivotal Greenplum driver:
    jdbc:pivotal:greenplum://<host>:<port>;/database_name=<value>
  • Connection string for Postgres Greenplum driver:
    jdbc:postgressql://<host>:<port>/<database_name>
  • Connection string for Teradata JDBC driver:
    jdbc:teradata://<host>/database_name=<value>,tmode=<value>,charset=<value>
For more information about the connection string to use with specific drivers, see the vendor documentation.
Environment SQL
Optional. Enter SQL commands to set the database environment when you connect to the database. The Data Integration Service executes the connection environment SQL each time it connects to the database.
If you enable Sqoop, Sqoop ignores this property.
Transaction SQL
Optional. Enter SQL commands to set the database environment when you connect to the database. The Data Integration Service executes the transaction environment SQL at the beginning of each transaction.
If you enable Sqoop, Sqoop ignores this property.
SQL Identifier Character
Type of character that the database uses to enclose delimited identifiers in SQL queries. The available characters depend on the database type.
Select (None) if the database uses regular identifiers. When the Data Integration Service generates SQL queries, the service does not place delimited characters around any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration Service generates SQL queries, the service encloses delimited identifiers within this character.
If you enable Sqoop, Sqoop ignores this property.
Support Mixed-case Identifiers
Enable if the database uses case-sensitive identifiers. When enabled, the Data Integration Service encloses all identifiers within the character selected for the
SQL Identifier Character
property.
When the
SQL Identifier Character
property is set to none, the
Support Mixed-case Identifiers
property is disabled.
If you enable Sqoop, Sqoop honors this property when you generate and execute a DDL script to create or replace a target at run time. In all other scenarios, Sqoop ignores this property.
Use Sqoop Connector
Enables Sqoop connectivity for the data object that uses the JDBC connection. The Data Integration Service runs the mapping in the Hadoop run-time environment through Sqoop.
You can configure Sqoop connectivity for relational data objects, customized data objects, and logical data objects that are based on a JDBC-compliant database.
Select
Sqoop v1.x
to enable Sqoop connectivity.
Default is
None
.
Sqoop Arguments
Enter the arguments that Sqoop must use to connect to the database. Separate multiple arguments with a space.
If you want to use Teradata Connector for Hadoop (TDCH) specialized connectors for Sqoop and run the mapping on the Blaze engine, define the TDCH connection factory class in the Sqoop arguments. The connection factory class varies based on the TDCH Sqoop Connector that you want to use.
  • To use the Cloudera Connector Powered by Teradata, configure the following Sqoop argument:
    -Dsqoop.connection.factories=com.cloudera.connector.teradata.TeradataManagerFactory
  • To use the Hortonworks Connector for Teradata (powered by the Teradata Connector for Hadoop), configure the following Sqoop argument:
    -Dsqoop.connection.factories=org.apache.sqoop.teradata.TeradataManagerFactory
    You do not need to define the TDCH connection factory class in the Sqoop arguments if you run the mapping on the Spark engine.
If you do not enter Sqoop arguments, the Data Integration Service constructs the Sqoop command based on the JDBC connection properties.
On the Hive engine, to run a column profile on a relational data object that uses Sqoop, set the Sqoop argument m to 1. Use the following syntax:
-m 1


Updated December 13, 2018