Table of Contents

Search

  1. Preface
  2. Fast Clone Overview
  3. Configuring and Using the Fast Clone Server
  4. Creating Cloning Configuration Files in the Fast Clone Console
  5. Unloading Data from the Source Database
  6. Loading Data to a Target
  7. Remote Configuration Management
  8. Fast Clone Command Line Interface
  9. Troubleshooting
  10. Fast Clone Configuration File Parameters
  11. Glossary

User Guide

User Guide

Configuration Considerations for Targets

Configuration Considerations for Targets

To accurately load data to the target database, review the configuration considerations for your target type and the requirements of the target load utility.

Considerations That Apply to Multiple Target Types

  • Greenplum, Netezza, Teradata, and Vertica targets require data files to use a new-line character (\n) as a row separator. If the new-line character occurs in the Oracle data as a data value, the load fails. To prevent this problem, configure Fast Clone to replace each new-line character with the space character in one of the following ways:
    • In the Fast Clone Console, select the
      Replace new line with space
      option on the
      Runtime Settings
      >
      Format Settings
      view.
    • Manually enter the replace_new_line_with_space=true parameter in the configuration file.
    • At the command line, enter the replace_newline_with_space=true parameter.
    Replacement of the new-line character with the space character can increase the overhead of unload processing.
    For Teradata targets, if you unload data in binary format, Fast Clone does not use a row separator in the output data files.
  • Greenplum and Vertica targets require the column separator and the enclosure character to be a single ASCII character. You can also specify a non-printable separator character by using the \x
    hh
    escape sequence in hexadecimal notation.
  • Greenplum, Teradata, and Vertica targets require data files to not use a column separator at the end of each row. For Teradata and Vertica targets, you must manually configure Fast Clone to produce data files without a column separator at the end of each row in one of the following ways:
    • In the Fast Clone Console, clear the
      Trailing column separator
      option on the
      Runtime Settings
      tab >
      Format Settings
      view.
    • Manually enter the trailing_column_separator parameter=false parameter in the configuration file.
    • At the command line, enter the trailing_column_separator=false command-line parameter.
    For Greenplum targets, Fast Clone never writes a trailing column separator at the end of each row, regardless of the trailing_column_separator setting.

Considerations for Amazon Redshift Targets

  • Fast Clone connects to and writes source data to the Amazon Simple Storage Service (Amazon S3). Amazon S3 is a storage service that can copy data from a source and simultaneously move the data to Amazon Redshift clusters. After the source data is in Amazon S3 storage, Fast Clone issues a copy command that copies the data to an Amazon Redshift target table.
  • Fast Clone does not delete the temporary files that contain the source data from the Amazon S3 bucket if the Fast Clone executable ends with an error. You can delete these files manually, or use Amazon S3 Object Lifecycle Management to automatically delete the temporary files from the bucket when the lifetime of the temporary files expires.
  • Fast Clone cannot load data from source columns that have binary datatypes to Amazon Redshift targets.

Considerations for Flat File Targets

  • Fast Clone does not generate load scripts and control files for flat file targets.

Considerations for Cloudera and Hortonworks Targets

  • To unload data to Cloudera and Hortonworks targets, your Fast Clone license must include Hadoop target support.
  • Fast Clone does not generate load scripts and control files for Hadoop targets.

Considerations for Hive Targets

  • To unload data to Hive targets, your Fast Clone license must include Hadoop target support.
  • To unload data to remote Hive targets, the underlying Hadoop software must allow remote connections.
  • In the
    Column separator
    field on the
    Runtime Settings
    tab >
    Format Settings
    view of the Fast Clone Console, enter the field delimiter that the Hive target uses in data files. By default, Hive data warehouses use the ASCII start of heading (SOH) character as the field delimiter.
  • Use the
    Custom url
    option to connect to a Hive target that requires Kerberos or LDAP authentication.
  • Ensure that the target tables are in CSV format. Fast Clone does not support loading data directly into tables that use other formats such as the Record Columnar File (RCFile) or Optimized Row Columnar (ORC) format.
  • Fast Clone does not support unloading source data to partitioned tables on Hive targets.
  • Fast Clone does not support unloading source data to Hive tables that use custom locations.
  • Fast Clone does not generate load scripts and control files for Hive targets.

Considerations for Netezza Targets

  • Netezza targets require the column separator to be a single ASCII character. You can also specify non-printable separator and enclosure character by using the \x
    hh
    escape sequence in hexadecimal notation.
  • To make unload and nzload processing faster, delete the default double quotation mark (") character in the
    Enclosed by
    field on the
    Runtime Settings
    tab >
    Format Settings
    view.
  • If you use the nzload utility to load data to a Netezza target, TIMESTAMP columns must use the format YYYY-MM-DD HH24:MI:SS.FF6, and DATE columns must use the format YYYY-MM-DD. The nzload utility supports these ISO date formats. Ensure that you specify the correct formats in the
    Timestamp format
    and
    Date format
    fields on the
    Runtime Settings
    tab >
    Format Settings
    view.
  • If you unload data from CLOB or LONG columns and load these values with the nzload utility to a Netezza target, clear the
    Unload binary into separate files
    option. The nzload utility can load up to 32,768 bytes of data into VARCHAR columns.
  • If you use DataStreamer to load data to a Netezza target, verify that the
    Unload binary into separate files
    ,
    Trailing column separator
    , and
    Suppress trailing null columns
    options are not selected on the
    Runtime Settings
    tab >
    Format Settings
    view.
    Alternatively, verify that the export_binary_to_separate_files, trailing_column_separator, and suppress_trailing_nullcols parameters are set to false in the configuration file.

Considerations for Teradata Targets

  • For Teradata targets, delete the default double quotation mark (") character in the
    Enclosed by
    field on the
    Runtime Settings
    tab >
    Format Settings
    view so that the field is empty.
  • If you want the output data file to be in Teradata binary format instead of text-delimited format, select the
    Binary length prefix
    option on the
    Runtime Settings
    tab >
    Teradata Load Settings
    view.
  • For Teradata targets, Informatica recommends that you create separate unload_td.cmd and unload_td.sh scripts that call the FastReaderTD executable for all unload jobs for this target type. You can then specify the unload_td file in the
    Informatica Fast Clone Console start script
    field on the
    Runtime Settings
    tab >
    File Locations
    view.
  • Data cannot be loaded to Teradata targets by using Fast Clone-generated load scripts and data files if the total row size exceeds the Teradata maximum buffer size. This error is most likely to occur with source tables that contain LOB columns because Fast Clone uses a default width of 32,768 bytes for LOBs.

0 COMMENTS

We’d like to hear from you!