To accurately load data to the target database, review the configuration considerations for your target type and the requirements of the target load utility.
Considerations That Apply to Multiple Target Types
Greenplum, Netezza, Teradata, and Vertica targets require data files to use a new-line character (\n) as a row separator. If the new-line character occurs in the Oracle data as a data value, the load fails. To prevent this problem, configure Fast Clone to replace each new-line character with the space character in one of the following ways:
In the Fast Clone Console, select the
Replace new line with space
option on the
Runtime Settings
>
Format Settings
view.
Manually enter the replace_new_line_with_space=true parameter in the configuration file.
At the command line, enter the replace_newline_with_space=true parameter.
Replacement of the new-line character with the space character can increase the overhead of unload processing.
For Teradata targets, if you unload data in binary format, Fast Clone does not use a row separator in the output data files.
Greenplum and Vertica targets require the column separator and the enclosure character to be a single ASCII character. You can also specify a non-printable separator character by using the \x
hh
escape sequence in hexadecimal notation.
Greenplum, Teradata, and Vertica targets require data files to not use a column separator at the end of each row. For Teradata and Vertica targets, you must manually configure Fast Clone to produce data files without a column separator at the end of each row in one of the following ways:
In the Fast Clone Console, clear the
Trailing column separator
option on the
Runtime Settings
tab >
Format Settings
view.
Manually enter the trailing_column_separator parameter=false parameter in the configuration file.
At the command line, enter the trailing_column_separator=false command-line parameter.
For Greenplum targets, Fast Clone never writes a trailing column separator at the end of each row, regardless of the trailing_column_separator setting.
Considerations for Amazon Redshift Targets
Fast Clone connects to and writes source data to the Amazon Simple Storage Service (Amazon S3). Amazon S3 is a storage service that can copy data from a source and simultaneously move the data to Amazon Redshift clusters. After the source data is in Amazon S3 storage, Fast Clone issues a copy command that copies the data to an Amazon Redshift target table.
Fast Clone does not delete the temporary files that contain the source data from the Amazon S3 bucket if the Fast Clone executable ends with an error. You can delete these files manually, or use Amazon S3 Object Lifecycle Management to automatically delete the temporary files from the bucket when the lifetime of the temporary files expires.
Fast Clone cannot load data from source columns that have binary datatypes to Amazon Redshift targets.
Considerations for Flat File Targets
Fast Clone does not generate load scripts and control files for flat file targets.
Considerations for
Cloudera and Hortonworks Targets
To unload data to
Cloudera and Hortonworks targets, your Fast Clone license must include Hadoop target support.
Fast Clone does not generate load scripts and control files for Hadoop targets.
Considerations for Hive Targets
To unload data to Hive targets, your Fast Clone license must include Hadoop target support.
To unload data to remote Hive targets, the underlying Hadoop software must allow remote connections.
In the
Column separator
field on the
Runtime Settings
tab >
Format Settings
view of the Fast Clone Console, enter the field delimiter that the Hive target uses in data files. By default, Hive data warehouses use the ASCII start of heading (SOH) character as the field delimiter.
Use the
Custom url
option to connect to a Hive target that requires Kerberos or LDAP authentication.
Ensure that the target tables are in CSV format. Fast Clone does not support loading data directly into tables that use other formats such as the Record Columnar File (RCFile) or Optimized Row Columnar (ORC) format.
Fast Clone does not support unloading source data to partitioned tables on Hive targets.
Fast Clone does not support unloading source data to Hive tables that use custom locations.
Fast Clone does not generate load scripts and control files for Hive targets.
Considerations for Netezza Targets
Netezza targets require the column separator to be a single ASCII character. You can also specify non-printable separator and enclosure character by using the \x
hh
escape sequence in hexadecimal notation.
To make unload and nzload processing faster, delete the default double quotation mark (") character in the
Enclosed by
field on the
Runtime Settings
tab >
Format Settings
view.
If you use the nzload utility to load data to a Netezza target, TIMESTAMP columns must use the format YYYY-MM-DD HH24:MI:SS.FF6, and DATE columns must use the format YYYY-MM-DD. The nzload utility supports these ISO date formats. Ensure that you specify the correct formats in the
Timestamp format
and
Date format
fields on the
Runtime Settings
tab >
Format Settings
view.
If you unload data from CLOB or LONG columns and load these values with the nzload utility to a Netezza target, clear the
Unload binary into separate files
option. The nzload utility can load up to 32,768 bytes of data into VARCHAR columns.
If you use DataStreamer to load data to a Netezza target, verify that the
Unload binary into separate files
,
Trailing column separator
, and
Suppress trailing null columns
options are not selected on the
Runtime Settings
tab >
Format Settings
view.
Alternatively, verify that the export_binary_to_separate_files, trailing_column_separator, and suppress_trailing_nullcols parameters are set to false in the configuration file.
Considerations for Teradata Targets
For Teradata targets, delete the default double quotation mark (") character in the
Enclosed by
field on the
Runtime Settings
tab >
Format Settings
view so that the field is empty.
If you want the output data file to be in Teradata binary format instead of text-delimited format, select the
Binary length prefix
option on the
Runtime Settings
tab >
Teradata Load Settings
view.
For Teradata targets, Informatica recommends that you create separate unload_td.cmd and unload_td.sh scripts that call the FastReaderTD executable for all unload jobs for this target type. You can then specify the unload_td file in the
Informatica Fast Clone Console start script
field on the
Runtime Settings
tab >
File Locations
view.
Data cannot be loaded to Teradata targets by using Fast Clone-generated load scripts and data files if the total row size exceeds the Teradata maximum buffer size. This error is most likely to occur with source tables that contain LOB columns because Fast Clone uses a default width of 32,768 bytes for LOBs.