Table of Contents

Search

  1. Preface
  2. Data Replication Overview
  3. Understanding Data Replication
  4. Sources - Preparation and Replication Considerations
  5. Targets - Preparation and Replication Considerations
  6. Starting the Server Manager
  7. Getting Started with the Data Replication Console
  8. Defining and Managing Server Manager Main Servers and Subservers
  9. Creating and Managing User Accounts
  10. Creating and Managing Connections
  11. Creating Replication Configurations
  12. Materializing Targets with InitialSync
  13. Scheduling and Running Replication Tasks
  14. Implementing Advanced Replication Topologies
  15. Monitoring Data Replication
  16. Managing Replication Configurations
  17. Handling Replication Environment Changes and Failures
  18. Troubleshooting
  19. Data Replication Files and Subdirectories
  20. Data Replication Runtime Parameters
  21. Command Line Parameters for Data Replication Components
  22. Updating Configurations in the Replication Configuration CLI
  23. DDL Statements for Manually Creating Recovery Tables
  24. Sample Scripts for Enabling or Disabling SQL Server Change Data Capture
  25. Glossary

Data Replication Runtime Parameters

Data Replication Runtime Parameters

Depending on the source and target types that you use for a replication job, the Data Replication Console lists the runtime parameters that are available for the Extractor, Applier, and InitialSync components.
In the Data Replication Console, you can optionally customize runtime parameters on the
Runtime Settings
tab >
Advanced Settings
view. The following list describes the runtime parameters that you can edit in the Console.

Amazon Redshift Parameters

redshift.s3.bucket_name
Specifies the name of the Amazon S3 bucket that stores the temporary files that contain the data to be applied to the Amazon Redshift target.
redshift.s3.file_size
Specifies the maximum size of the temporary files, in kilobytes, that Data Replication creates in the Amazon S3 bucket before loading the source data to Amazon Redshift.
Data Replication treats this parameter value as a soft limit on the temporary file size.
Valid values are 1700 to 100000.
Default value: 5120 KB
redshift.s3.key_id
Specifies an AWS access key ID that Data Replication must use to access the Amazon S3 account resources including the bucket where Data Replication creates temporary files.
redshift.s3.path
Specifies the name of the directory in the Amazon S3 bucket where Data Replication creates the temporary files that store source data.
redshift.s3.secret_key
Specifies the secret access key for the access key ID that is specified in the redshift.s3.key_id parameter. The access key ID must have the authority to access the Amazon S3 account resources.

Apache Kafka Parameters

apply.avro.avro_before_image_prefix
Specifies a prefix for the names of Avro message fields that contain the before image data for an Update operation. Valid values are the digits 0-9, Latin letters A-Z and a-z, and the underscore (_) character.
You can also specify a suffix in the apply.avro.avro_before_image_suffix parameter. If you specify both a prefix and suffix, the Avro before-image field name uses the following format:
<prefix>
<source_column_name>
<suffix>
.
If you do not specify a value for this parameter or the apply.avro.avro_before_image_suffix parameter, the Avro before image field name duplicates the source column name. Informatica recommends that you enter a value in at least one of these parameters to create unique before-image field names.
For this parameter to be used, the apply.avro.avro_include_before_image parameter must be set to 1.
Default value: Null
apply.avro.avro_before_image_suffix
Specifies a suffix for the names of the Avro message fields that contain the before image data for an Update operation. Valid values are the digits 0-9, Latin letters A-Z and a-z, and the underscore (_) character.
You can also specify a prefix in the apply.avro.avro_before_image_prefix parameter. If you specify both a suffix and prefix, the Avro before-image field name uses the following format:
<prefix>
<source_column_name>
<suffix>
.
If you do not specify a value for this parameter or the apply.avro.avro_before_image_prefix parameter, the Avro before-image field name duplicates the source column name. Informatica recommends that you enter a value in at least one of these parameters to create unique before-image field names.
For this parameter to be used, the apply.avro.avro_include_before_image parameter must be set to 1.
Default value: _OLD
apply.avro.avro_include_before_image
Indicates whether Data Replication creates Avro message fields that contain the before image of updated columns when replicating an Update operation to a Kafka target. Valid values are:
  • 0
    . Do not create fields that contain the before image of columns.
  • 1
    . Create fields that contain the before image of columns.
Default value: 1
To add a prefix or suffix to the names of before-image fields, use the apply.avro.avro_before_image_prefix and apply.avro.avro_before_image_suffix parameters.
apply.avro.avro_include_is_present
Indicates whether Data Replication creates Avro message fields that contain a Boolean value indicating one of the following reasons for the occurrence of a null value in an Avro message field:
  • The corresponding source column value is also null (null "is present").
  • The corresponding source column has a value that is not null (null is not present). However, Data Replication does not capture the value because it did not change during the current replication cycle.
Valid values are:
  • 0
    . Do not create fields that contain a Boolean value indicating whether the null value is present or is not present in the source data.
  • 1
    . Create fields that contain a Boolean value indicating whether the null value is present or not present in the source data.
Default value: 1
To add a prefix or suffix to the names of these fields, use the apply.avro.avro_is_present_prefix and apply.avro.avro_is_present_suffix parameters.
apply.avro.avro_is_present_prefix
Specifies a prefix for the named of the Avro message fields that indicate whether a null value in an Avro message is also present in the corresponding source data. Valid values are the digits 0-9, Latin letters A-Z and a-z, and the underscore (_) character.
You can also specify a suffix in the apply.avro.avro_is_present_suffix parameter. If you specify both a prefix and suffix, the Avro field name uses the following format:
<prefix>
<source_column_name>
<suffix>
.
If you do not specify a value for this parameter or the apply.avro.avro_is_present_suffix parameter, the names of the Avro fields that indicate whether a null is present in the source data duplicate the source column names. Informatica recommends that you enter a value in at least one of these parameters to create unique field names.
For this parameter to be used, the apply.avro.avro_include_is_present parameter must be set to 1.
Default value: Null
apply.avro.avro_is_present_suffix
Specifies a suffix for the names of the Avro message fields that indicate whether a null value in an Avro message is also present in the corresponding source data. Valid values are the digits 0-9, Latin letters A-Z and a-z, and the underscore (_) character.
You can also specify a prefix in the apply.avro.avro_is_present_prefix parameter. If you specify both a suffix and prefix, the field name uses the following format:
<prefix>
<source_column_name>
<suffix>
.
If you do not specify a value for this parameter or the apply.avro.avro_is_present_prefix parameter, the names of the Avro fields that indicate whether a null is present in the source data duplicates the source column names. Informatica recommends that you enter a value in at least one of these parameters to create unique field names.
For this parameter to be used, the apply.avro.avro_include_is_present parameter must be set to 1.
Default value: _PRESENT
apply.avro.avro_metadata_prefix
Specifies a prefix for the names of any metadata fields that are added to an Avro schema to distinguish these metadata fields from the source columns. You can also specify a suffix by using the apply.avro.avro_metadata_suffix parameter. If you specify both a prefix and suffix, the names of the metadata fields use the following format:
<prefix>
<metadata_field_name>
<suffix>
.
Default value: Null
apply.avro.avro_metadata_suffix
Specifies a suffix for the names of any metadata fields that are added to an Avro schema to distinguish these metadata fields from the source columns. You can also specify a prefix by using the apply.avro.avro_metadata_prefix parameter. If you specify both a suffix and prefix, the names of the metadata fields use the following format:
<prefix>
<metadata_field_name>
<suffix>
.
Default value: Null
apply.avro.avro_name_replacement_value
Specifies a character that will replace any unsupported character in source column or table names when Data Replication generates Avro field names for Kafka messages. Avro field names do not allow spaces or special characters. If this parameter is not specified, Data Replication will not replace unsupported Avro characters from source column or table names. For example, if this parameter specifies the underscore (_) character and a source column is named
column#name with space
, the corresponding Avro field name is
column_name_with_space
.
Default value: Underscore (_)
apply.avro.avro_schema_cache_directory
Specifies the subdirectory in the
DataReplication_installation
/output/
replication_configuration
directory where the Avro formatting code stores Avro schema objects that are used to produce Avro messages for each mapped source table.
Default value: overflow
apply.avro.avro_schema_cache_memory_size
Specifies the maximum number of Avro schema files in a single Least Recently Used (LRU) cache instance to keep in memory. Data Replication writes an Avro schema file for each mapped source table in a configuration to the directory specified in the apply.avro.avro_schema_cache_directory parameter. When the number of Avro schema files in the cache instance exceeds the limit specified in this parameter, the excess files are removed from memory.
Valid values are 10 through 1000000000
Default value: 1000
apply.kafka.kafka_checkpoint_file_directory
Specifies the subdirectory in the
DataReplication_installation
/output/
replication_configuration
directory to which Data Replication writes checkpoint files when applying data to a Kafka target. The Applier creates this subdirectory when it first runs.
Default value: checkpoints
apply.kafka.kafka_file_checkpoint_file_name
Specifies the name of the checkpoint file that Data Replication creates when applying data to a Kafka target. When the value of this parameter is DEFAULT, each checkpoint file name matches the corresponding configuration name.
Default value: DEFAULT
apply.kafka.kafka_client_libraries_directory
Specifies the directory that Data Replication uses to access Kafka client libraries. Copy the files from the
kafka_version
/libs directory to the default Data Replication directory or enter the path to another directory. The library files must match the version of Kafka to which Data Replication connects, and the files must be located in a directory that runs on the same machine as the Data Replication Applier.
Default value:
DataReplication_installation
/lib/queueAdapterLib.
apply.kafka.kafka_message_key
Specifies an optional message key that Data Replication uses to assign a topic partition to the Avro messages sent to a Kafka target. All messages that use the same topic and message key are assigned to the same partition in the topic. If you specify USE_TABLE_NAME, Data Replication uses the name of each mapped source table as the key name in the messages.
Default value: Null
apply.kafka.kafka_producer_guarantee_delivery
Specifies whether or not to enable guaranteed delivery of messages to a Kafka target. When guaranteed delivery is enabled, the CDC Publisher attempts to guarantee that no messages are lost or duplicated on the target if a network outage occurs. Guaranteed delivery provides for the highest level of data integrity on the target but might degrade performance. Valid values are:
  • 0
    . Disable guaranteed delivery. The Applier writes a checkpoint to the checkpoint file after each Commit operation, or after the timeout period that is specified in the apply.qadapter.checkpoint_request_timeout runtime parameter elapses. If you restart apply processing after an outage, duplicate or missing messages might occur on the target.
  • 1
    . Enable guaranteed delivery. The Applier writes a checkpoint to the checkpoint file after each message is successfully sent to the target topic. All messages are sent synchronously if the apply.kafka.kafka_producer_send_synchronously runtime parameter is set to 1. If you restart apply processing after an outage, no duplicate or missing messages should occur on the target.
Default value: 1
apply.kafka.kafka_producer_max_retry_attempts
Specifies the maximum number of times that the CDC Publisher Connector will retry sending a Kafka message that produced an error.
Default value: 5
apply.kafka.kafka_producer_partition_ID
Specifies the topic partition ID that Data Replication will include in each message sent to Kafka. A value of -1 causes Data Replication to send a null partition ID to Kafka.
Default value: -1
apply.kafka.kafka_producer_send_synchronously
Controls whether to use synchronous delivery of messages to Kafka. When synchronous delivery is enabled, Kafka must acknowledge each message as received before Data Replication will send the next message. In this mode, Kafka is unlikely to receive duplicate messages, but apply performance might be slower. Valid values are:
  • 0
    . Disable synchronous delivery. The Applier sends messages asynchronously to the target.
  • 1
    . Enable synchronous delivery.
Default value: 0
With either option, the CDC Publisher maintains changes in commit order as they are streamed to the target.
apply.kafka.kafka_table_name_include_schema
When USE_TABLE_NAME is specified in the Data Replication Console or in the apply.kafka.kafka_message_key parameter to use the mapped source table name as the Kafka topic name or message key name, you can use this parameter to also include the source table schema name in the topic and key names. Valid values are:
  • 0
    . Do not include the source table schema name in the topic and key names.
  • 1
    . Include the source table schema name in the topic and key names. With this setting, a Kafka topic name will have the following format:
    schema_name
    .
    source_table_name
    .
Default value: 0
apply.kafka.kafka_table_name_prefix
When USE_TABLE_NAME is specified in the Data Replication Console or in the apply.kafka.kafka_message_key parameter to use the mapped source table name as the Kafka topic name or message key name, you can use this parameter to also include a prefix in the topic and key names. For example, if you specify a prefix in this parameter and set the apply.kafka.kafka_table_name_include_schema parameter to 1, a Kafka topic name will have the following format:
prefix
_
schema_name
.
source_table_name
.
Default value: Null
apply.qadapter.cache_overflow_directory
Specifies the subdirectory in the
DataReplication_installation
/output/
replication_configuration
directory that Data Replication uses to write Least Recently Used (LRU) cache overflow files from the CDC Publisher. The Applier creates this subdirectory when it first runs.
Default value: overflow
apply.qadapter.cache_memory_size
For the CDC Publisher, specifies the maximum number of items in a single Least Recently Used (LRU) cache instance to keep in memory. When the number of cached items exceeds this limit, the overflow items are stored in the overflow directory that is set in the apply.qadapter.cache_overflow_directory parameter.
Valid values are 10 through 1000000000
Default value: 1000
apply.qadapter.checkpoint_request_timeout
Specifies the time interval (in milliseconds) that the Applier waits for the CDC Publisher to validate that all data was successfully sent to a Kafka target before timing out. The Applier typically requests this checkpoint information at the end of each Apply cycle before the intermediate files are deleted. If the CDC Publisher does not validate that all data was sent to a Kafka target before this time elapses, the CDC Publisher reports a timeout error. The Applier then checks for error conditions, retries the request, or fails. Use this parameter and the apply.qadapter.checkpoint_retries parameter to provide enough time for the Applier to write all data to a Kafka target so that Data Replication can release the intermediate files for the Apply cycle.
If Data Replication persistently issues an IDR-070218 error message and reports a syncKafkaEnvironment failure, try adjusting the apply.qadapter.checkpoint_request_timeout and apply.qadapter.checkpoint_retries parameter values.
Default value: 60000 (60 seconds)
apply.qadapter.checkpoint_retries
After a timeout error, specifies the maximum number of consecutive times that the Applier retries a request for the CDC Publisher to validate that all data was successfully sent to a Kafka target. If the CDC Publisher does not validate that all data was sent to a Kafka target after the maximum number of retries, the Applier ends with an error. For example, if the value of the apply.qadapter.checkpoint_request_timeout parameter is 60000 (60 seconds) and the value of this parameter is 6, then 6 minutes must expire before the Applier stops waiting for all data to be written to a Kafka target and ends with an error.
If Data Replication persistently issues an IDR-070218 error message and reports a syncKafkaEnvironment failure, try adjusting the apply.qadapter.checkpoint_request_timeout and apply.qadapter.checkpoint_retries parameter values.
Default value: 5
apply.qadapter.create_missing_directories
When the directories that are specified in the apply.kafka.kafka_checkpoint_file_directory and apply.qadapter.cache_overflow_directory parameters do not already exist, determines whether to create these directories or report an access error. Valid values are:
  • 0
    . Report an access error, and do not create the directories.
  • 1
    . Create the directories.
Default value: 1
apply.qadapter.properties_file
Specifies a file that contains additional CDC Publisher properties for use by Informatica Global Customer Support or other Informatica personnel. The properties in this file are used to change or augment default CDC Publisher behavior with Data Replication. Edit this parameter only at the direction of Informatica Global Customer Support.

DB2 for Linux, UNIX, and Windows Parameters

apply.db2.create_lob_nologging
For configurations that have DB2 targets and any source other than DB2, determines the length of DB2 LOB columns that Data Replication creates when replicating CREATE TABLE and ADD COLUMN operations or when generating target tables based on source tables. The length determines whether logging can be enabled for these columns. Valid values are:
  • 0
    . Create LOB columns of 1 GB in length with the default LOGGED option. The LOGGED option enables logging.
  • 1
    . Create LOB columns of 2 GB in length with the NOT LOGGED option. The NOT LOGGED option disables logging.
    DB2 for Linux, UNIX, and Windows does not support logging for LOB columns that are larger than 1 GB.
Default value: 0
db2.connection_additional_params
Specifies additional parameters for the ODBC connection string.
No default value
db2.direct_buffer_size
Specifies the buffer size, in bytes, that InitialSync uses for loading and unloading data when using the DB2 APIs.
Default value: 1048576 bytes
db2.initial_lsn
Specifies the LSN from which the Extractor starts extracting changes when you run the Extractor for the first time.
Default value: Current LSN
db2.max_failed_tasks
When you run InitialSync with DB2 sources
and targets
, Data Replication creates a separate process for each table. This parameter specifies the maximum number of processes that can fail before the InitialSync task fails.
Default value: 5
db2.node_name
Specifies the DB2 node name that the InitialSync component or Extractor uses to connect to a DB2 source. This parameter is required for DB2 sources.
No default value
db2.odbc.enable_decimal_period
If you use the ODBC driver to connect DB2 targets when running an Applier or InitialSync task, indicates whether the ODBC driver uses a period (.) as the decimal separator in float values that are loaded to the target. Options are:
  • 0
    . Preserve the decimal separator that is in the source float values when loading the float values to the target.
  • 1
    . Use a period (.) as a decimal separator in the target float values.
Default value: 1
db2.task_timeout
When you run InitialSync with DB2 sources
and targets
, Data Replication creates a separate process for each table. If a single process cannot open a pipe before the time limit that is specified in this parameter expires, the process fails. If you accept the default value of 0, Data Replication uses one of the following values:
  • 5 seconds on Windows
  • 20 seconds on Linux and UNIX
Default value: 0
extract.db2.checkpoint_size
Specifies the amount of transaction log, in MB, at which an extraction checkpoint is taken. If the extraction process fails, the next time you run it, it resumes from the last checkpoint.
Default value: 100 MB
extract.db2.db_alias_name
Specifies a database alias that the Extractor uses instead of the connection details on the
Source Database
tab to connect to a remote DB2 source database. This alias must be defined on the system where the Extractor runs.
If the parameter value is an empty string, the Extractor uses the connection details on the
Source Database
tab to connect to the DB2 source database.
Default value: empty string
extract.db2.enable_row_compression_support
Determines whether the DB2 Extractor requests the filtered log buffer for compressed rows from the db2ReadLog API. The Extractor needs the filtered log buffer to extract change data from tables that use row compression. Options are:
  • 0
    . Do not request the filtered log buffer for compressed rows. Use this option if none of the mapped source tables use row compression.
  • 1
    . Request the filtered log buffer for compressed rows. This option might degrade replication performance. Use this option only if one or more mapped source tables use row compression.
Default value: 0
  • When you create a replication configuration that maps a source table that uses row compression, the Data Replication Console sets this parameter to 1.
  • To prevent data loss, set this parameter to 1 before enabling row compression for a mapped source table.
extract.db2.filter_option_forced
Indicates whether the DB2 Extractor always requests the DB2 filtered log buffer from the db2ReadLog API. Options are:
  • 0
    . Request the unfiltered log buffer. However, if the extract.db2.enable_row_compression_support runtime parameter is set to 1, the Extractor requests the filtered log buffer to get compressed records.
  • 1
    . Always request the filtered log buffer. The Extractor processes propagatable records from the filtered log buffer. These records do not contain DDL operations and LOB data that is not stored inline. This setting improves the Extractor performance.
    If you enable DDL replication or map source LOB columns, the Data Replication Console sets the parameter to the default value of 0.
Default value: 0
extract.db2.lob_lookup_threads
Specifies the number of threads that the Extractor uses to extract LOB data from the source database. If you use the default value of 0, the Extractor ignores this parameter and extracts LOB data from the transaction logs.
Default value: 0
extract.db2.lob_truncate_size
Specifies the maximum amount of LOB data, in KB, that the Extractor extracts. If LOB data in a column is larger than the specified value, the data is truncated.
Default value: 64 KB
extract.db2.log_buffer_size
Specifies the buffer size, in kilobytes, that is used for parsing the transaction log. To improve performance, set this parameter to the estimated average log record size. A value that is too low or too high might significantly increase CPU consumption on the DB2 server. Lower values cause the Extractor to read the transaction log more frequently and might degrade Extractor performance.
Default value: 1024 KB
extract.db2.max_read_retries
Specifies the maximum number of times that the Extractor can retry a request to the DB2 API for the DB2 archive logs after the first request returns the log buffer in an invalid format. If all of the retries fail, the Extractor ends with an error.
Default value: 5
global.lsn_table_connection_attempts
Indicates the number of times the Extractor or InitialSync retries a query to the DBSYNC_SYNC_LSN service table after the initial query timeout interval elapses. The query timeout interval is specified in the global.lsn_table_connection_timeout parameter.
Default value: 3
global.lsn_table_connection_timeout
Indicates the number of seconds after which an Extractor or InitialSync query to the DBSYNC_SYNC_LSN service table will time out. After this interval elapses, the Extractor or InitialSync retries the query up to the number of attempts specified in the
global.lsn_table_connection_attempts
parameter.
Default value: 15
initial.db2.sync_lsn_table
Specifies the name of the service table. The Extractor uses the service table to determine the point in the transaction log from which the Applier should begin applying changes. The Extractor passes this information to the Applier.
You can change the default service table name when you save the configuration the first time. The Data Replication Console asks if you want to select a schema or to edit the service table name before it creates the table in the DB2 database.
After the configuration is saved, this parameter shows the full name of the service table in the following format:
schema_name
.
table_name
.
Default value: DBSYNC_SYNC_LSN
initial.db2.table_lock_enabled
Indicates whether to lock the source tables that the InitialSync component is processing. Options are:
  • 1
    . Lock.
  • 0
    . Do not lock.
InitialSync acquires a share lock on the source tables that are being synchronized to guarantee a consistent snapshot. However, you might want to prevent locking in some situations, for example, for test purposes.
Default value: 1
initial.db2.til
Specifies which transaction isolation level to use during the initial synchronization process. Options are:
  • 0
    . Default database transaction isolation level.
  • 1
    . Uncommitted read.
  • 2
    . Cursor stability.
  • 3
    . Read stability.
  • 4
    . Repeatable read.
Default value: 0

Greenplum Parameters

apply.enable_greenplum_table_redistribution
Indicates whether the Applier runs an ALTER TABLE
table
SET DISTRIBUTED
column
,... statement before replicating a DDL operation that adds a primary key constraint or unique index to a Greenplum target. Options are:
  • 0
    . Do not run the ALTER TABLE ... SET DISTRIBUTED statement.
    If the target table uses the random distribution policy, the Applier cannot replicate DDL operations that add a primary key constraint or unique index.
  • 1
    . Run the ALTER TABLE ... SET DISTRIBUTED statement on a Greenplum target before applying a DDL operation that adds a primary key constraint or unique index. This option sets the distribution key columns to the unique constraint columns.
    If the target table contains a large number of rows, this option can be resource intensive.
Default value: 1
Change the default value only at the request of Informatica Global Customer Support.

Microsoft SQL Server Parameters

extract.mssql.checkpoint_size
Specifies the amount of a transaction log, in megabytes, at which an extraction checkpoint is taken. If the extraction process fails, the next time you run it, it resumes from the last checkpoint.
Default value: 100 MB
extract.mssql.ignore_inactive_backup_locations
If you add shared backup log locations on the
Extract Range
tab, indicates whether the Extractor ends with an error when a location becomes temporarily unavailable. Options are:
  • 0
    . The Extractor ends with an error if it cannot access a backup log location.
  • 1
    . The Extractor continues processing even if it cannot access a backup log location.
    Use this option for Microsoft SQL Server sources that use Always On Availability Groups.
Default value: 0
extract.mssql.log_buffer_size
Specifies the buffer size, in megabytes, that the Extractor uses to parse the backup and online logs. Higher values might improve Extractor performance but require more memory to allocate log data. Lower values cause the Extractor to read the transaction log more frequently and might degrade Extractor performance.
Default value: 100 MB
extract.mssql.process_updates_as_updates
Indicates whether the Extractor handles each SQL Server Update as a pair of Delete and Insert operations or as a single Update. Options are:
  • 0
    . Handle each Update as a pair of Delete and Insert operations.
  • 1
    . Handle each Update as an Update. The Extractor writes a single Update operation to the intermediate files. As a result, the performance of Applier processing of Updates in SQL Apply mode is improved.
Default value: 1
  • Microsoft SQL Server 2008 R2 generates a pair of Delete and Insert operations for each Update. If you set this parameter to 1, the Extractor merges these Delete and Insert operations into a single Update operation. Other supported SQL Server versions process Updates as Updates. If you set this parameter to 0, the Extractor splits each Update operation into a pair of Delete and Insert operations.
  • For configurations that include Merge Apply mappings, the Data Replication Console sets this parameter to 0.
extract.mssql.together_with_native_replication
Indicates whether the Extractor manages the secondary truncation checkpoint. Valid values are:
  • 0
    . Enable Extractor management of the secondary truncation checkpoint.
  • 1
    . Disable Extractor management of the secondary truncation checkpoint. Set this parameter to
    1
    in the following cases:
    • You run the Extractor that reads transaction logs and SQL Server Change Data Capture (CDC) or native transactional replication against the same database. In this case, SQL Server must manage the secondary truncation checkpoint and produce backup logs.
    • You run two Extractors that read transaction logs from the same database. In this case, only one Extractor must manage the secondary truncation checkpoint.
    • You run the Extractor that reads transaction logs and run the Microsoft SQL Server Backup task with the
      Run the sp_repldone procedure
      check box selected.
    After you disable the secondary truncation checkpoint management by the Extractor, ensure that the Extractor reads both the transaction and backup logs to prevent incomplete data capture.
This parameter is a synonym for the
Disable secondary truncation checkpoint
parameter that you can set on the
Runtime Settings tab
General view
.
Default value: 0
extract.use_db_connection
Indicates whether the Microsoft SQL Server Extractor is connected to the SQL Server source database. The Extractor must connect to the source database when reading online logs or working in continuous mode. Options are:
  • 0
    . Do not connect to the SQL Server source database.
  • 1
    . Connect to the SQL Server source database.
Default value: 1
extractor.mssql.directory_for_decompressed_backup_logs
Specifies the name of the directory to which the Microsoft SQL Server Extractor decompresses compressed backup transaction logs before processing them.
Default value:
DataReplication_installation
/tmp
mssql.enable_quoted_identifier
For Microsoft SQL Server
sources and
targets, indicates whether InitialSync adds the ODBC EnableQuotedIdentifiers=1 parameter to the connection string. When EnableQuotedIdentifiers is set to 1, SQL Server allows identifiers that are delimited by double quotation marks in Transact-SQL statements. Quoted identifiers can include reserved keywords and special characters that are not otherwise allowed in Transact-SQL statements. Options are:
  • 0
    . Do not add the EnableQuotedIdentifiers parameter to the connection string.
  • 1
    . Set the EnableQuotedIdentifiers parameter to 1 in the connection string to allow quoted identifiers.
Default value: 0

MySQL Parameters

mysql.connection_encoding
Specifies the connection-related character set that the InitialSync and Applier tasks use to load data to MySQL targets. The MySQL server uses this setting to convert the statements that are executed by the InitialSync and Applier tasks to the character set that is used by the database server.
The default connection-related character set for the InitialSync and Applier tasks is determined by the MySQL server character settings. Use this runtime parameter if the data that InitialSync unloads or that the Extractor extracts has a character set other than the default connection character set for the MySQL target. For a list of supported connection character sets for MySQL databases, see the MySQL documentation.

Netezza Parameters

apply.asyncloader.exit_on_odbc_error
Indicates whether the Applier is forced to end with an error after an ODBC error occurs on the Netezza target. Options are:
  • 0
    . The Applier completes normal processing of the ODBC error.
  • 1
    . The Applier is forced to end with an error.
Default value: 0
apply.asyncloader.reopen_pipe_timeout
If a Netezza target fails to open the pipe after an ODBC error, the Applier might continue to write change data to the pipe and then hang. In this case, Data Replication checks the pipe two times to determine if the Applier is still writing data to this pipe. If the
apply.asyncloader.exit_on_odbc_error
runtime parameter is set to 0, this parameter specifies the number of seconds between the first and second checks.
If the Applier is writing data to the pipe after the second check, Data Replication tries to open and close the pipe to complete normal processing of the error.
Default value: 1
Increase this parameter value if high loads occur on the Netezza target.
apply.netezza.connection_timeout
Specifies the number of seconds that the Applier waits for a response from the Netezza target. If this timeout period elapses without the response, the Applier ends with an error message.
Default value: 600 seconds
Increase this parameter value if high loads occur on the Netezza target.
nzload.ignorezero
Indicates whether the Applier uses the Netezza IGNOREZERO external table option to strip null characters (0x00) from extracted source data when applying the data to a character column on the Netezza target. The null characters can be null values within the source data or null terminator characters. Options are:
  • 0
    . Do not strip null characters from the data. The Netezza target returns an error if the Applier tries to apply data that includes null characters to a character column.
  • 1
    . Strip null characters from the data before applying the data to a character column. If a column contains null characters with other characters, the Applier forces Netezza to strip the null characters and then applies the other characters to the Netezza target column. For fixed-length target columns, the Applier pads the character string with blanks up to the fixed column length. If a source column contains only null characters, the Applier applies a null value to the Netezza target column.
    If the Applier applies a null value to a target column that is defined with the NOT NULL constraint, a constraint violation occurs.
The InitialSync task truncates source data at the first null character that it encounters, regardless of this parameter setting.
Default value: 0

Oracle Parameters

apply.direct_load_for_audit_tables
For data replications in Audit Apply mode, indicates whether to use the native database load utility or the Oracle ODBC driver, which is delivered with the Oracle Client, to load data to audit log tables on the Oracle target. Options are:
  • 0
    . Use the ODBC driver for Oracle to load data to the audit log tables on the Oracle target.
  • 1
    . Use the native database load utility to load data to the audit log tables on the Oracle target.
Default value: 1
For configurations that include Audit Apply mappings of tables with LOB columns, the Data Replication Console sets this parameter to 0.
apply.disable_target_triggers
For Oracle targets of 10.2.0.5 or a later 10.2 patch and for Oracle targets of 11.2.0.2 or a later 11.2 patch, indicates whether to disable suppressing triggers. Options are:
  • 0
    . Suppress triggers.
  • 1
    . Do not suppress triggers.
Default value: 0
apply.local_time_zone_hours
Specifies the hour value in the local time zone GMT offset for the source database. The Applier task uses this information to process TIMESTAMP WITH LOCAL TIMEZONE columns.
Default value: 0
apply.local_time_zone_minutes
Specifies the minute value in the local time zone GMT offset for the source database. The Applier task uses this information to process TIMESTAMP WITH LOCAL TIMEZONE columns.
Default value: 0
apply.oracle.use_returning_into
For Oracle targets, determines whether the Applier uses the RETURNING INTO clause to insert change data into LOB and LONG columns. Options are:
  • 0
    . Do not use the RETURNING INTO clause. The Applier cannot insert change data into LOB and LONG columns if the total row data size exceeds 4000 bytes. This setting can result in faster apply processing.
  • 1
    . Use the RETURNING INTO clause. The Applier can insert change data into LOB and LONG columns even if the total row data size exceeds 4000 bytes. This setting can result in slower apply processing.
Default value: 0
extract.max_flob_size
Determines the maximum buffer size that the Extractor allocates for Oracle LOB data when you perform row filtering based on a LOB value. If you enter a large value for this parameter, Data Replication allocates large object data in memory, which can degrade replication performance. If the LOB data is larger than the specified buffer size, the Extractor filters the data so that some of it fits in the buffer and skips the rest.
Default value: 12 KB
extract.oracle.bsc.enable
Indicates whether the Oracle Extractor reads archived redo logs in backup sets that were created by the RMAN utility. Valid values are:
  • 0
    . Do not process redo logs from backup sets.
  • 1
    . Process redo logs from backup sets.
Default value: 0
extract.oracle.bsc.number_of_threads_to_build_indexes
Specifies the number of threads that the Oracle Extractor can use for indexing archived redo logs in backup set containers.
For this parameter to be used, the extract.oracle.bsc.enable parameter must be set to 1.
Default value: 4
Increase the number of indexing threads if the hard disk system is not fully loaded when indexing redo logs in backup set containers. Decrease the number if the hard disk system usage reaches 100 percent.
extract.oracle.check_size_of_parsed_logs
Indicates whether the Oracle Extractor ends with an error after processing a redo log file if the file size that is specified in the log header does not match the number of bytes that the Extractor processed. Options are:
  • 0
    . The Extractor logs a warning message and continues processing.
  • 1
    . The Extractor ends with an error.
Default value: 1
extract.oracle.double_verification
Indicates whether the Oracle Extractor uses a checksum to verify the data integrity of online and archived redo logs. Valid values are:
  • 0
    . Do not verify the data integrity of redo logs.
  • 1
    . Verify the data integrity of the current online redo log.
  • 2
    . Verify the data integrity of all online redo logs.
  • 3
    . Verify the data integrity of all online and archived redo logs.
Default value: 1
extract.oracle.print_logs_for_processing
Indicates whether the Extractor logs a message that lists the Oracle redo log files for processing to the Extractor execution log. In continuous mode, the Extractor logs this list at the beginning of each Extractor microcycle. If you run the replication schedule on demand or periodically, the Extractor logs this list when the Extractor task starts. Options are:
  • 0
    . Do not log the message that lists the redo log files for processing.
  • 1
    . Log the message that lists the redo log files for processing.
Default value: 0
extract.oracle.print_truncate_operations
Indicates whether the Extractor logs a message each time it encounters a TRUNCATE TABLE operation in the Oracle redo logs. The message includes the TRUNCATE TABLE statement. Options are:
  • 0
    . Do not log the message that reports the TRUNCATE TABLE statement.
  • 1
    . Log the message that reports the TRUNCATE TABLE statement.
Default value: 0
extract.oracle.rds.fetch_archive_log_list_attempts
Specifies the maximum number of times that the Oracle Extractor retries the request for the list of archived redo log files from the Amazon RDS for Oracle instance.
The Extractor retries the request at the interval that is specified in the extract.oracle.rds.fetch_archive_log_list_timeout runtime parameter.
Default value: 5
extract.oracle.rds.fetch_archive_log_list_timeout
Specifies the interval, in seconds, that the Oracle Extractor waits after a request to get the list of archived redo log files from the Amazon RDS for Oracle instance fails before requesting the list again.
Default value: 1
extract.oracle.rds_read_enable
Controls whether the Extractor can extract change data from online and archived redo logs for Amazon Relational Database Service (RDS) for Oracle instances in the cloud. Options are:
  • 0
    . The Extractor cannot process redo logs from Amazon RDS for Oracle instances.
  • 1
    . The Extractor can process redo logs from Amazon RDS for Oracle instances.
Default value: 0
extract.oracle.read_log_list_attempts
Specifies the maximum number of times the Extractor tries to get the list of Oracle redo log files for processing. If the redo logs do not pass integrity checking after the Extractor reaches the specified maximum number of attempts, the Extractor stops processing the redo logs.
Default value: 5
extract.oracle.receive_timeout
Specifies the value that the Oracle Extractor uses for the Oracle SQLNET.RECV_TIMEOUT profile parameter if this parameter is not specified in the sqlnet.ora configuration file.
The SQLNET.RECV_TIMEOUT parameter specifies the time, in seconds, for a client to wait for response data from the database server after establishing a connection.
Default value: 600
extract.oracle.send_timeout
Specifies the value that the Oracle Extractor uses for the Oracle SQLNET.SEND_TIMEOUT profile parameter if this parameter is not specified in the sqlnet.ora configuration file.
The SQLNET.SEND_TIMEOUT parameter specifies the time, in seconds, for a client to complete send operations to the database server after establishing a connection.
Default value: 600
extract.oracle.skip_encrypted_tablespace_records
For Oracle sources that use Transparent Data Encryption (TDE) for tablespaces, determines the Extractor behavior when it cannot decrypt a record from an encrypted tablespace. Options are:
  • 0
    . Do not skip records from an encrypted tablespace that the Extractor cannot decrypt. In this case, the Extractor ends with an error.
  • 1
    . Skip records from an encrypted tablespace only if you did not open an Oracle PKCS #12-format wallet in the Data Replication Console.
  • 2
    . Skip records from an encrypted tablespace if you did not open an Oracle PKCS #12-format wallet or if the wallet that you opened in the Data Replication Console does not include a master key to decrypt these records.
Default value: 1
extract.oracle.skip_lob_data_for_table
Specifies the name of an Oracle source table for which you want the Extractor to replace LOB data with null values in the intermediate files. Use this parameter for a table that includes SecureFile LOB columns that use unsupported options, such as NOLOGGING or FILESYSTEM_LIKE_LOGGING. Otherwise, the Extractor ends with an error.
Default value: Empty
extract.oracle.threads_offline_at_start
Specifies a comma-separated list of the Oracle redo threads that are offline at the SCN that corresponds to the Start Point value for the Extractor. The Extractor starts processing change data records from an Oracle redo thread that is in this list only after a Thread Enable Marker (TEM) appears in the redo log. The Data Replication Console populates this list of offline redo threads. Informatica recommends that you do not change the populated values for this parameter.
extract.oracle.threads_online_at_start
Specifies a comma-separated list of the Oracle redo threads that are online at the SCN that corresponds to the Start Point value for the Extractor. The Extractor starts processing change data records from Oracle redo threads that are in this list. The Data Replication Console populates this list of online redo threads. Informatica recommends that you do not change the populated values for this parameter.
extract.oracle.use_log_file_header_cache
Indicates whether the Oracle Extractor caches the headers of processed redo log files. Valid values are:
  • 0
    . Disable caching of redo log headers.
  • 1
    . Enable caching of redo log headers.
Default value: 1
extract.process_missing_columns
Determines the Oracle Extractor behavior after it encounters a source column that is missing from the configuration. Options are:
  • 0
    . The Extractor issues a parser exception. Subsequent Extractor behavior depends on the
    extract.stop_on_parsing_error
    runtime parameter setting.
  • 1
    . The Extractor writes change data for the missing column to an intermediate file.
    Use this option to correctly replicate ADD COLUMN ... DEFAULT operations that occur on the source. The Applier updates all of the existing rows in the mapped target table with the default value for the added column. If the target table contains many rows, this operation might take a long time.
  • 2
    . The Extractor does not write change data for the missing column to an intermediate file.
    Use this option to reduce overhead on Merge Apply targets after an ADD COLUMN ... DEFAULT operation on the source. The Applier does not update the existing rows in the mapped target table with the default value of the added source column.
Default value: 0
extract.stop_on_table_rename
Indicates whether the Extractor ends with an error after a mapped table is renamed in the Oracle source database. Options are:
  • 0
    . Log a warning message and continue Extractor processing. In this case, the Extractor continues extracting change data from the renamed table. The Applier continues to apply change data to the mapped target table. However, the target table is not renamed.
  • 1
    . End with an error.
Default value: 0
extract.use_strict_intermediate_file_size
For Oracle sources, determines whether to treat the value that is specified in the
Maximum size of each intermediate file
field on the
Runtime Settings
tab >
General
view as a soft limit or hard limit on the intermediate file size. Options are:
  • 0
    . The Extractor treats the value in the
    Maximum size of each intermediate file
    field as a soft limit on the intermediate file size. The Extractor checks the size of the intermediate file at each checkpoint. If the file size exceeds the specified limit, the Extractor writes the data and closes the intermediate file.
  • 1
    . The Extractor treats the value in the
    Maximum size of each intermediate file
    field as a hard limit on the intermediate file size. The Extractor estimates the size of the intermediate file when processing each change record. If the estimated file size exceeds the specified limit, the Extractor forces a checkpoint, writes the data, and closes the intermediate file.
Default value: 0
extract.verbose_ddl
Indicates whether to include debug messages that are related to Extractor processing of source DDL in the Extractor output. The messages provide information related to DDL changes that the Extractor parses from Oracle system tables. Use these messages to identify potential DDL processing errors. Options are:
  • 0
    . Do not include debug messages that are related to source DDL processing in the Extractor output.
  • 1
    . Include debug messages that are related to source DDL processing in the Extractor output.
Default value: 0
general.number_overflow_abort
For Amazon Redshift, DB2 for Linux, UNIX, and Windows, Greenplum, Microsoft SQL Server, Netezza, Teradata, and Vertica targets, if you use the general.number_overflow_verify parameter setting of 1, indicates whether the Applier and InitialSync end with an error if the whole number portion of an Oracle source NUMBER value exceeds the size of the target column. This situation is called a
numeric data overflow
. Options are:
  • 0
    . The Applier and InitialSync do not end with an error if a numeric data overflow occurs. Replace the number with the value that is specified in the general.number_overflow_replace parameter.
  • 1
    . The Applier and InitialSync end with an error if a numeric data overflow occurs.
Default value: 1
general.number_overflow_replace
For Amazon Redshift, DB2 for Linux, UNIX, and Windows, Greenplum, Microsoft SQL Server, Netezza, Teradata, and Vertica targets, if you use the general.number_overflow_verify setting of 1 and the general.number_overflow_abort setting of 0, specifies the replacement value for the Oracle NUMBER values that cause a numeric data overflow on the target. Valid values are any floating-point numbers that do not cause a numeric data overflow on the target.
When a numeric data overflow occurs, the Applier and InitialSync do not end with an error. Instead, the task replaces the NUMBER values that caused the overflow on the targets with the specified replacement value.
Default value: Null value
general.number_overflow_verify
For Amazon Redshift, DB2 for Linux, UNIX, and Windows, Greenplum, Microsoft SQL Server, Netezza, Teradata, and Vertica targets, indicates whether the Applier and InitialSync handle numeric data overflows for Oracle NUMBER columns that are replicated to target columns. Options are:
  • 0
    . Disable Applier and InitialSync handling of numeric data overflows. Allow the target database to handle numeric data overflow errors.
  • 1
    . Enable Applier and InitialSync handling of numeric data overflows. Use the general.number_overflow_abort and general.number_overflow_replace parameters to change the default handling of numeric data overflows on the target.
Default value: 1
initial.dblink.disable_global_names
If InitialSync uses database links (dblinks) for initial synchronization of Oracle sources and Oracle targets, indicates whether to override the Oracle GLOBAL_NAMES parameter value of true with false for the InitialSync targets. An override is required when the GLOBAL_NAMES parameter is set to true because InitialSync creates database link names that differ from the target database names. Options are:
  • 0
    . Do not change the GLOBAL_NAMES parameter setting for InitialSync sessions that have Oracle targets. Use this option only if the GLOBAL_NAMES parameter is set to false for the Oracle targets.
  • 1
    . Override the GLOBAL_NAMES parameter setting of true with false for InitialSync sessions that have Oracle targets. Use this option if the GLOBAL_NAMES parameter is set to true for the Oracle targets.
Default value: 0
If the GLOBAL_NAMES parameter setting for an Oracle target is true, the Data Replication Console sets this initial.dblink.disable_global_names parameter to 1. If you set this parameter back to 0, InitialSync cannot create database links.
initial.oracle_clob_text_buffer_size
For Oracle sources, indicates whether InitialSync unloads CLOB data as character data of limited size. Enter a value greater than 0 to specify the buffer size, in bytes, that InitialScync can use to unload CLOB data as character data. Enter 0 to have InitialSync unload CLOB data as LOB data.
For Teradata and Netezza targets, this parameter is honored if the initial.enforce_odbc_load runtime parameter is set to 1. For Oracle targets, this parameter is honored if InitialSync does not use dblinks.
Default value: For configurations with Teradata targets, the default is the value of the global.lob_truncation_size runtime parameter. For configurations with any other target type, the default is 0.
initial.oracle.parallel_sample_percentage
For Oracle sources, specifies the percentage of the total number of rows in a source table that InitialSync randomly samples to get row IDs for distributing the table rows across multiple InitialSync subtask threads. InitialSync executes a SELECT statement with a WHERE clause that uses the sampled row IDs to define a range of rows to be processed on a thread. In this manner, Data Replication uses the sampled row IDs to distribute all of the table rows across InitialSync threads for multithreaded load processing. For example, if you enter 2 percent for a table that contains 100 rows and the sampled row IDs are 30 and 60, Data Replication could distribute the table rows as follows:
  • Thread1 processes rows 0-30.
  • Thread2 processes rows 31-60.
  • Thread3 processes rows 61-100.
The number of InitialSync subtask threads is limited by the initial.oracle.parallel_subtask_limit runtime parameter.
For more information about using multiple InitialSync threads, see Considerations for Running InitialSync.
Default value: 0
If you use the default value of 0, InitialSync processes all table rows with a single thread. Use the default value of 0 for source tables for which InitialSync does not support multithreaded processing, including Oracle source tables that have subpartitions and tables that have virtual columns associated with Tcl scripts or SQL expressions.
Use the initial.oracle.parallel_sample_seed runtime parameter to select a reproducible sample of Oracle source table rows and row IDs for distributing table rows across multiple InitialSync threads, each time you run InitialSync. This way, you can better compare InitialSync performance across multiple executions that use different initial.oracle.parallel_sample_percentage parameter settings or different Data Replication versions.
initial.oracle.parallel_sample_seed
If you set the initial.oracle.parallel_sample_percentage parameter to a positive value, specifies the SEED value in the SELECT…SAMPLE statement that InitialSync uses to select a reproducible sample of Oracle source table rows and row IDS for distributing table rows across multiple InitialSync threads, each time you run InitialSync. By using a reproducible sample of rows, you can better compare InitialSync performance across multiple executions that use different initial.oracle.parallel_sample_percentage parameter settings or different Data Replication versions.
Valid values are 0 through 2147483647
Default value: 0, which provides a different distribution of rows across InitialSync threads each time InitialSync runs.
initial.oracle.parallel_subtask_limit
If the initial.oracle.parallel_sample_percentage runtime parameter is set to a non-zero value, specifies the maximum number of subtask threads that InitialSync can use to load data in parallel to the targets that you are materializing from Oracle sources.
Default value: 4
initialsync.oracle.check_corruption_event
Indicates whether to detect a corrupted Oracle source when running an InitialSync task. This option uses event 10200 to indicate the error.
Default value: false
initialsync.oracle.no_parallel_select
For Oracle sources, indicates whether to disable parallel query processing when running an InitialSync task. Options are:
  • false
    . Use parallel query processing for Oracle sources.
  • true
    . Disable parallel query processing for Oracle sources for InitialSync.
Default value: false
initialsync.oracle.skip_corruption_event
Indicates whether to skip corrupted blocks in an Oracle source database when running the InitialSync task. This option uses events 10231 and 10233 to indicate the error.
Default value: false
initialsync.oracle.use_dblink
For Oracle-to-Oracle initial synchronization only, indicates whether Data Replication can use database links (dblinks). The use of dblinks can improve performance. Options are:
  • 1
    . Enables Data Replication to use dblinks.
  • 0
    . Prevents Data Replication from using dblinks.
Default value: 1
oracle.replace_empty_string_characters
If you replicate source columns that contain empty strings to Oracle targets, specifies the character that the Applier and InitialSync use to replace the empty strings when writing data to the corresponding target columns.
You can enter any string that is valid for the target column datatype. The default value is a space, which causes the Applier and InitialSync to replace the empty strings with the space character on the target.
For this parameter to be honored, you must set the oracle.replace_empty_string_enabled parameter to 1.
Default value: Space character
oracle.replace_empty_string_enabled
For Oracle targets, indicates whether the Applier and InitialSync replace empty strings from the source with a character value when writing the strings to the target. Options are:
  • 0
    . Do not replace empty strings from the source with a character value on the target. The Applier and InitialSync replicate the empty strings as null values to the target.
  • 1
    . Replace empty strings from the source with a character value on the target. The oracle.replace_empty_string_characters runtime parameter specifies the replacement character that the Applier and InitialSync use. Use this option if you map source columns that include empty strings to NOT NULL target columns to avoid NOT NULL constraint violations.
Default value: 0
oracle.replace_empty_string_not_null_only
For Oracle targets, if you set the oracle.replace_empty_string_enabled parameter to 1, indicates whether the Applier and InitialSync replace empty strings from the source with a replacement character when writing data to all of the target columns or only to NOT NULL target columns. Options are:
  • 0
    . Replace the empty strings with a replacement character in both NOT NULL and NULL target columns.
  • 1
    . Replace the empty strings with a replacement character only in NOT NULL target columns.
You can specify the replacement character in the oracle.replace_empty_string_characters parameter.
Default value: 1
oracle.truncate_char_to_4000
For Oracle targets, indicates whether the Applier truncates character data to 4,000 bytes if the corresponding source tables have character data that is larger than 4,000 bytes. Options are:
  • 0
    . Do not truncate character data.
  • 1
    . Truncate character data to 4,000 bytes.
Default value: 0

Teradata Parameters

apply.teradata.td_max_sessions
Specifies the maximum number of sessions for TPT libraries that an Applier thread can use for loading data to a target. The value for this parameter must be less than or equal to the number of AMPs (Access Module Processors) on the Teradata target database.
Default value: 8
apply.teradata.td_tenacity_hours
Adjusts the TENACITY runtime option for the TPT Load or Update operator when you run the Applier task. This parameter specifies the number of hours that the TPT Load or Update operator tries to log on when the maximum number of load jobs is already running on the Teradata database.
For more information about the TENACITY option, see the Teradata documentation.
Default value: 0
initialsync.teradata.td_max_sessions
Specifies the maximum number of sessions for TPT libraries that an InitialSync thread can use for loading data to a target. The value for this parameter must be less than or equal to the number of AMPs on the Teradata target database.
Default value: 8
initialsync.teradata.td_tenacity_hours
Adjusts the TENACITY runtime option for the TPT Update operator when you run the InitialSync task. This parameter specifies the number of hours that the TPT Update operator tries to log on when the maximum number of load jobs is already running on the Teradata database.
For more information about the TENACITY option, see the Teradata documentation.
Default value: 0
teradata.disable_utf8_load
Indicates whether InitialSync and the Applier load source character data to a Teradata target by using UTF-8 encoding. Options are:
  • 0
    . Use UTF-8 encoding when loading source character data to the target.
  • 1
    . Do not use UTF-8 encoding when loading source character data to the target. This setting might improve replication performance.
    Specify 1 if source character data includes only Latin characters.
Default value: 0
teradata.stream_load_hook
If the Applier uses the TPT Stream operator, indicates whether the Applier locks Teradata target tables before loading change data. Options are:
  • 0
    . Lock the target tables before loading change data.
  • 1
    . Do not lock the target tables before loading change data.
Default value: 1
teradata.stream_operator_enabled
Indicates the type of TPT loader that the Applier uses to load change data to Teradata targets. Options are:
  • 0
    . Always use the TPT Load operator to apply change data to Teradata targets in Merge Apply mode. Always use the TPT Update operator to apply change data to Teradata targets in Audit Apply mode.
  • 1
    . Use the TPT Stream operator or the TPT Load or Update operator to load change data to Teradata targets, depending on the amount of change data. If the amount of change data from a source table is greater than the value that is specified by the teradata.threshold_table_size parameter, the Applier uses the TPT Load operator in Merge Apply mode and uses the TPT Update operator in Audit Apply mode. If this amount is less than the teradata.threshold_table_size value, the Applier uses the TPT Stream operator in both Audit Apply and Merge Apply modes.
Default value: 1
teradata.stream_operator_max_connections
Specifies the maximum number of TPT Stream operator connections to a Teradata target that the Applier can use.
Default value: 4
teradata.td_trace_level
Specifies the tracing level for the Teradata driver. Options are:
  • TD_OFF
    . Disable Teradata driver trace messages.
  • TD_OPER
    . Enable trace messages for driver-specific activities.
  • TD_OPER_CLI
    . Enable trace messages for driver activities that involve CLIv2.
  • TD_OPER_NOTIFY
    . Enable trace messages for driver activities that involve the Notify feature.
  • TD_OPER_OPCOMMON
    . Enable trace messages for driver activities that involve the operator common library.
  • TD_OPER_ALL
    . Enable all driver-level trace messages.
Default value: TD_OFF
The Teradata driver writes trace messages to the log file that is specified by the teradata.td_trace_output_name runtime parameter.
teradata.td_trace_level_infr
Specifies the tracing level for the Teradata infrastructure. Options are:
  • TD_OFF
    . Disable trace messages for Teradata infrastructure.
  • TD_OPER
    . Enable trace messages for infrastructure-specific activities.
  • TD_OPER_CLI
    . Enable trace messages for infrastructure activities that involve CLIv2.
  • TD_OPER_NOTIFY
    . Enable trace messages for infrastructure activities that involve the Notify feature.
  • TD_OPER_OPCOMMON
    . Enable trace messages for infrastructure activities that involve the operator common library.
  • TD_OPER_ALL
    . Enable all infrastructure-level trace messages.
Default value: TD_OFF
The Teradata infrastructure writes trace messages to the log file that is specified by the teradata.td_trace_output_name runtime parameter.
teradata.td_trace_output_name
Specifies the name of the output file that includes the TPT trace messages. Data Replication uses the following naming pattern to generate this file in the
DataReplication_installation
/output directory:
file_name
.
schema_name
.
table_name
.
PID
.
timestamp
You can configure the tracing level of the Teradata driver by using the teradata.td_trace_level and teradata.td_trace_level_infr runtime parameters.
Default value: tpttrace. The full default name for the TPT trace log is tpttrace.
schema_name
.
table_name
.
PID
.
timestamp
.
teradata.threshold_table_size
Specifies the maximum amount of change data, in bytes, that the Applier can apply to a Teradata target table by using the TPT Stream operator in Merge Apply or Audit Apply mode. If the amount of change data exceeds this limit, the Applier applies change data to the target by using either the TPT Load operator in Merge Apply mode or the TPT Update operator in Audit Apply mode.
To determine the size of change data, select the configuration from the
Server Manager
Configs
view in the Data Replication Console and click the
Intermediate Files
icon button. The
Intermediate Files for the Configuration <configuration_name>
dialog box appears. For each file listed in the
Intermediate Files
box on the left, the amount of extracted data, in bytes, is displayed in the
Transactions
box on the right by transaction.
Data Replication compresses the size of intermediate files by default. You can edit this setting and adjust the maximum size, in megabytes, of intermediate files on the
Runtime Settings
tab >
General
view.
Default value : 1048576
teradata.tpt_thread_safe
Indicates whether to use additional mutual exclusion mechanisms for thread-unsafe tasks when an Applier or InitialSync task loads data to Teradata. Options are:
  • 0
    . Do not use additional mutual exclusion mechanisms for thread-unsafe tasks.
  • 1
    . Use additional mutual exclusion mechanisms for thread-unsafe tasks.
Default value: 0

Common Parameters

apply.buffer_size_for_split_records
Specifies the maximum size of the buffer, in bytes, that the Applier uses to store data for a long-running transaction or for a table that is queued for apply processing in Merge Apply mode. When this buffer size limit is exceeded, the Applier flushes data from the buffer to a temporary spill file and then writes any additional change data to the spill file. The Applier creates a buffer for each long-running transaction or for each table that is queued for apply processing in Merge Apply mode.
Default value: 1048576 bytes
apply.conflict_resolution_json_offset
If you use the apply.conflict_resolution_pretty_json parameter, specifies the indentation level for the JSON log files. Data Replication uses the specified number of spaces to indent JSON elements.
Default value: 1
apply.conflict_resolution_pretty_json
Indicates whether to use the indented format for the JSON log file to which the Applier writes information about detected conflicts. You specify a path to this log file on the
Runtime Settings
tab >
General
view. Options are:
  • 0
    . Use a single-line format for the JSON log files. Use this option to reduce the size of the log files.
  • 1
    . Use an indented format for the JSON log files. The indentation level is determined by the apply.conflict_resolution_json_offset parameter.
Default value: 1
apply.force_log_table_suffix_for_audit
For configurations that include multiple targets, indicates whether the Applier alters the SQL apply statements for secondary targets that use Audit Apply mode to append the audit log suffix to the table names on those targets. The audit log suffix is specified in the
Log table suffix for merge apply
field on the
Runtime Settings
tab >
Calculated Columns
view and usually used for Merge Apply. Options are:
  • 0
    . Do not append the audit log suffix to the table names on the secondary targets that use Audit Apply mode. With this option, the names of the audit log tables on the secondary targets match the names of the regular target tables on the primary target. This option is typically used for configurations that replicate data from one primary target to multiple secondary targets in different databases.
  • 1
    . Append the audit log suffix to the table names on the secondary targets that use Audit Apply mode. With this option, the audit log table names on the secondary targets are composed of the primary target table names followed by the audit log suffix. This option is used in special situations where you need to use the audit log suffix to distinguish the audit log tables on secondary targets that use Audit Apply mode.
    You can use Audit Apply mode for a secondary target to add audit log tables in the primary target database that use the same schema as the regular target tables in the same database. In this case, if you use Merge Apply mode for the primary target, the audit log tables for Merge Apply mode must have a different schema. Specify a schema that includes audit log tables for Merge Apply mode in the
    Log table schema for merge apply
    field on the
    Runtime Settings
    tab >
    Calculated Columns
    view.
Default value: 0
apply.libpq_client_encoding
For Greenplum
and Vertica
targets, specifies the original source character encoding for data that Data Replication applies using the libpq library in Audit Apply and Merge Apply modes. InitialSync also uses this setting to specify original source character encoding for data that it loads using the libpq library. This parameter is equivalent to setting the PGCLIENTENCODING environment variable. For example, if the Oracle source system character set is WE8ISO8859P1, set this parameter to LATIN1. For more information about available character sets for a target, see your target database documentation.
No default value.
apply.libpq_custom_connection_string
For Greenplum
, Vertica
,
and PostgreSQL
targets, specifies a custom connection string for the libpq library. This parameter overrides the automatically generated libpq connection string, which has the following format:
host=
target_host_name
port=
target_port
user=
target_user
password=
target_password
dbname=
target_database_name
No default value.
apply.load_sessions_per_load_pass
Specifies the number of target tables that the Applier can load concurrently. By default, Data Replication uses this parameter only for Netezza
, Vertica,
and Teradata targets. To use this parameter for other target types, set the apply.loader_multiple_load_passes parameter to 1.
If you use the value of -1 for this parameter, the Applier determines the number of tables that it can load to the target concurrently based on the target database type. The Applier uses the following criteria:
  • For Netezza, the number of tables equals the number of Applier threads. Maximum is 20 tables.
  • For Vertica, the number of tables equals the number of Applier threads multiplied by 4. Maximum is 20 tables.
  • For other target types, the number of tables equals the number of Applier threads. No maximum limit applies.
Default value: -1
To change this value for Netezza targets, Informatica recommends that you clear the
Create output as pipes instead of files
option on the
Runtime Settings
tab >
General
view. If this option is selected, each Applier thread can load only a single audit log table at a time.
apply.loader_multiple_load_passes
Indicates whether the number of target tables that the Applier loads concurrently is limited. Options are:
  • 0
    . The Applier loads data to all of the target tables concurrently. No limit on the number of tables is in effect.
  • 1
    . The Applier loads data to a limited number of target tables concurrently. The number of tables is limited by the apply.load_sessions_per_load_pass runtime parameter.
Default value: 0. However, for Netezza
, Vertica,
and Teradata targets, Data Replication always uses the parameter value of 1 because the number of concurrent loader tasks on these targets is limited. For Apache Kafka targets, Data Replication always uses the parameter value of 0.
apply.loader_tx_map_size_rollback_key
Specifies the number of slots in a hash structure that is used to process rollback-to-savepoint records during each apply cycle. Increase this value if rollback-to-savepoint operations are common and process a lot of records.
Default value: 262144
apply.loader_tx_map_size_tx
Specifies the number of slots in a hash structure that is used to hold committed transactions during each apply cycle. For better performance, set this value to one-tenth of the number of opened or closed transactions in each apply cycle.
Default value: 500000
apply.merge.double_precision
The Applier rounds 8-byte floating-point numbers, such as DOUBLE and 8-byte FLOAT values, from the source before loading them to Merge Apply targets. This parameter specifies the number of digits after the decimal point that the Applier preserves in the 8-byte floating-point numbers that are applied to the target.
Default value: 15
apply.merge.float_precision
The Applier rounds 4-byte floating-point numbers, such as 4-byte FLOAT and REAL values, from the source before loading them to Merge Apply targets. This parameter specifies the number of digits after the decimal point that the Applier preserves in the 4-byte floating-point numbers that are applied to the target.
Default value: 8
apply.performance_degradation_warning_percentage
Specifies the percentage degradation of the average SQL statement processing time that must occur before Data Replication logs a warning. You can specify the intervals at which Data Replication calculates the average SQL statement processing times in the apply.print_statistics_interval and initial.print_statistics_interval parameters.
Data Replication does not log a performance degradation warning if the average SQL statement processing time is less than the value that is specified by the apply.performance_degradation_warning_threshold parameter.
Default value: 10%
apply.performance_degradation_warning_step
Specifies the number of performance readings that must indicate degradation of the average SQL statement processing time before Data Replication logs a warning. You can specify the intervals at which Data Replication calculates the average SQL statement processing times in the apply.print_statistics_interval and initial.print_statistics_interval parameters.
Data Replication does not log a performance degradation warning if the average SQL statement processing time is less than the value that is specified by the apply.performance_degradation_warning_threshold parameter.
Default value: 3
apply.performance_degradation_warning_threshold
Specifies the minimum average SQL statement processing time, in milliseconds, over an apply cycle for which Data Replication logs a performance degradation warning.
Default value: 5 milliseconds
apply.post_apply_script_enabled
Indicates whether the Applier runs post-apply SQL statements or stored procedure calls after each apply cycle on a target. Post-apply processing is supported for configurations that have Greenplum, Netezza, Oracle,
Vertica,
or Teradata targets and that use Audit Apply or Merge Apply mode. You must specify the post-apply SQL statements or stored procedure calls in the
%DBSYNC_HOME%/uiconf/afterapply.xsl
file. Options are:
  • 0
    . Disable post-apply processing.
  • 1
    . Enable post-apply processing.
Default value: 1
apply.print_statistics_interval
For configurations with targets that use ODBC drivers, specifies the number of seconds that must elapse before Data Replication updates the number of processed rows in the Applier output log. If you specify 0, Data Replication does not include the number of processed rows in the Applier output log.
Default value: 300 seconds
apply.process_intermediate_size_per_job
Specifies the maximum total size of all intermediate files, in megabytes, that Data Replication processes during a single apply cycle. Data Replication always processes entire intermediate files. Data Replication never splits an intermediate file to avoid exceeding the maximum total size that is specified in this parameter. You control the maximum size of a single intermediate file by setting the
Maximum size of each intermediate file
option on the
Runtime Settings
tab >
General
view.
Each apply cycle ends with a single commit by default except in certain situations when you use Applier multi-threaded processing. For more information, see Commit Processing and Target Constraints with Applier Multi-Threaded Processing. If you specify smaller values for this parameter, Data Replication apply processing reduces the transaction size and increases the commit frequency on the target. If you specify 0, Data Replication processes all available intermediate files within a single apply cycle.
Data Replication always processes at least one intermediate file during a single apply cycle, even if the intermediate file size exceeds the maximum total size that is specified in this parameter.
Default value: 256 MB
apply.realtime_clean_metadata_every_n_cycles
For continuous replication, specifies the number of apply microcycles that the Applier runs before it removes information about the processed intermediate files from the Applier SQLite database. This parameter is for use only by Informatica Global Customer Support. Do not modify it.
Default value: 1000 microcycles
apply.recovery_enabled
Indicates whether to enable recovery based on information in the recovery table for apply processing. Options are:
  • 0
    . Disable recovery based on information in the recovery table. When you save a new configuration in the Data Replication Console, this parameter is transparently set to 0 if you clicked
    No
    in the
    Enter Recovery Table Name and Schema
    dialog box to not create a recovery table with the specified name.
  • 1
    . Enable recovery based on information in the recovery table. When you save a new configuration in the Data Replication Console, this parameter is transparently set to 1 if you clicked
    Yes
    in the
    Enter Recovery Table Name and Schema
    dialog box to create a recovery table. Ensure that the apply.recovery_table parameter specifies the schema and table name of the recovery table.
For database target types other than
Cloudera,
Flat File,
Hortonworks,
and Kafka, Informatica recommends that you enter 1 to enable recovery for apply processing. For
Cloudera,
Flat File,
Hortonworks,
and Kafka targets that do not use a recovery table, enter 0.
Default value: 0 for the
Cloudera,
Flat File,
Hortonworks,
and Kafka targets, 1 for other targets.
apply.recovery_table
Specifies the schema and table name of the recovery table on the primary target. Only target types other than
Cloudera,
Flat File,
Hortonworks,
and Kafka use recovery tables. For the recovery table to be used, ensure that the apply.recovery_enabled parameter is set to 1.
In the Data Replication Console, when you save a new configuration, the Console prompts you to specify the schema and table name of the recovery table. You can accept the default values or override them. The default schema is the database user, and the default table name is IDR_RECOVERY. If you click
Yes
, this parameter is populated with the full recovery table name. The Applier uses this parameter value to create the recovery table. If you click
No
, the recovery table is not created and the apply.recovery_enabled parameter is set to 0. If you want to use a recovery table later, you can create it in one of the following ways:
  • Manually create the table by using the DDL in DDL Statements for Manually Creating Recovery Tables. Then enter the schema and table name of the table that you created in this parameter and set the apply.recovery_enabled parameter to 1, before saving the configuration again.
  • In the Data Replication Console, set the apply.recovery_enabled runtime parameter to 1 and save the configuration again. The Console prompts you for the schema and table name. After you specify these names and click
    Yes
    , the Console enters the full
    schema.table_name
    value in this parameter. When you start the Applier later, it will create the recovery table.
To specify recovery tables for secondary targets in a configuration that has multiple targets, use the
Recovery table name
field on the
Routing
tab >
Override Apply
subtab.
Default value: IDR_RECOVERY
apply.script_engine_enabled
Indicates whether the Tcl Script Engine executes the following Tcl scripts that are located in the
DataReplication_installation
/support/Tcl directory during apply processing:
  • Execute dbs_main_start.tcl when the Applier task starts.
  • Execute dbs_main_end.tcl when the Applier task ends.
  • Execute dbs_cycle_start.tcl when an apply cycle starts.
  • Execute dbs_cycle_end.tcl when an apply cycle ends.
Options are:
  • 0
    . Do not execute these Tcl scripts during apply processing.
  • 1
    . Execute these Tcl scripts during apply processing.
Default value: 0
apply.skip_if_pk_not_defined
Indicates whether the Applier is forced to end with an error if it encounters an Update or Delete record that does not contain undo values for all of the primary key columns. Options are:
  • 0
    . The Applier ends with an error.
  • 1
    . In SQL Apply and Merge Apply modes, the Applier skips any Update or Delete record that does not include undo values for all of the primary key columns and continues apply processing. This setting can result in data inconsistencies.
    In Audit Apply mode, the Applier applies these records to the target instead of skipping them.
Default value: 0
apply.skip_oracle_piecewise_operations
Indicates whether the Applier skips Oracle piecewise operations.
Data Replication supports Oracle piecewise operations only for Oracle targets in Audit Apply mode if apply.direct_load_for_audit_tables is set to 0 and in SQL Apply mode. This parameter determines the Applier behavior when processing change records that contain piecewise operations in the following cases:
  • For target databases other than Oracle.
  • For Oracle targets in Merge Apply mode.
  • For Oracle targets in Audit Apply mode if apply.direct_load_for_audit_tables is set to 1.
Options are:
  • 0
    . Do not skip Oracle piecewise operations. If piecewise operations occur, the Applier ends with an error.
  • 1
    . Skip Oracle piecewise operations.
Default value: 0
apply.sql_statement_text_id_cache_optimization_enabled
For source tables mapped in SQL Apply mode, controls whether the Applier caches the SQL statements that it generates for applying data to target tables in memory. Enable caching of these SQL statements to improve Applier performance. Options are:
  • 0
    . Do not cache the SQL statements for applying data to the target tables.
  • 1
    . Cache the SQL statements for applying data to the target tables.
Default value: 1
apply.threshold_time_for_open_transaction
Specifies the number of minutes that a transaction must be open before the Applier logs a warning message about the open transaction. The Applier compares this threshold time to the time the transaction has been open, which is calculated as the difference between the time of the last record from the open transaction and the time of the last processed record from any transaction.
Default value: 60 minutes
apply.trace_db_execution
For Oracle targets, indicates whether to use Oracle SQL tracing. This tracing is enabled using SQL event 10046 with trace levels 1, 4, 8, or 12. Options are:
  • 0
    . Do not use SQL tracing.
  • 1
    . Use standard SQL trace level 1. Provides statistics for parse, execute, fetch, commit and rollback database calls.
  • 4
    . Use SQL trace level 4. This option displays level 1 information and additional information about bind variables.
  • 8
    . Use SQL trace level 8. This option displays level 1 information and additional information about wait time.
  • 12
    . Use SQL trace level 12. This option enables trace levels 4 and 8 simultaneously.
Default value: 0
apply.verbose_ddl
Indicates whether to include the DDL statements that the Applier uses to apply source DDL changes to the target in the Applier output. The Applier translates the source DDL statements to DDL statements that the target database can process. You can include these translated DDL statements in the Applier output to determine the specific DDL statements that the Applier used to apply structural changes on the target. Options are:
  • 0
    . Do not include the SQL statements for the target that correspond to source DDL changes in the Applier output.
  • 1
    . Include the SQL statements for the target in the Applier output.
Default value: 0
apply.verbose_level
Specifies the level of verbose output for the Applier process. Larger values provide more detailed output. The specific output information depends on the Applier type. Valid values are 0 to 4.
Default value: 0
apply.verbose_process_long_running_txns
Indicates whether the Applier logs messages that are related to processing memory buffers and temporary spill files for change data that belongs to long-running open transactions. Options are:
  • 0
    . Do not include the messages that are related to processing memory buffers and temporary spill files for long-running transactions in the Applier output.
  • 1
    . Include the messages that are related to processing memory buffers and temporary spill files for long-running transactions in the Applier output.
Default value: 1
apply.verbose_print_every_n_records
Determines the frequency at which Data Replication updates progress when the Applier task is running. When the specified number of records are reached, Data Replication updates progress information.
Default value: 1000000 records
apply.verbose_process_queued_tables
Indicates whether the Applier logs messages that are related to processing memory buffers and temporary spill files for queued source tables. The Applier cannot apply change data from these tables during the current apply cycle because of target loader limitations. Options are:
  • 0
    . Do not include the messages that are related to processing memory buffers and temporary spill files for queued tables in the Applier output.
  • 1
    . Include the messages that are related to processing memory buffers and temporary spill files for queued tables in the Applier output.
Default value: 0
apply.verify_modified_row_count
Indicates whether Data Replication verifies that the number of rows that the Applier expected to process for Inserts, Updates, and Deletes during the apply cycle matches the actual count of rows for these SQL changes that were applied. Options are:
  • 0
    . Do not verify that the expected row count matches the actual row count.
  • 1
    . Verify that the expected row count matches the actual row count.
    A row count mismatch might occur in the following situations:
    • Data Replication replicates a row for an Insert operation, but an Insert row was already written to the target.
    • Data Replication replicates a row for a Delete or Update operation, but no rows match the WHERE condition that is used when applying changes to the target.
    With this setting, the apply cycle is rolled back and a new apply cycle begins with row-by-row processing. Data Replication prints the SQL statements and bind variable values with the mismatched row count to the log. The Applier then reports the mismatch and either stops or continues depending on the apply.verify_modified_row_count_abort_on_mismatch parameter setting.
This type of verification works only for SQL Apply table mappings. It does not work for Audit Apply or Merge Apply mappings.
Default value: 1
apply.verify_modified_row_count_abort_on_mismatch
If you set the apply.verify_modified_row_count parameter to 1, use this parameter to control whether the Applier task stops or continues when a mismatch occurs between the expected count of Insert, Update, and Delete rows and the actual row count. Options are:
  • 0
    . Report an error and continue apply processing.
  • 1
    . Stop apply processing.
Default value: 1
data_direct_optimize_prepare
Specifies the value for the OptimizePrepare (OP) attribute in the connection string that is generated for the DataDirect ODBC drivers. The OptimizePrepare attribute determines whether the driver creates stored procedures on the database server for calls to the ODBC SQLPrepare function. Options are:
  • 0
    . Create stored procedures for each call to the SQLPrepare function. This setting might degrade performance when processing SQL statements that do not contain parameters.
  • 1
    . Create stored procedures only if the SQL statement includes parameters. If the SQL statement does not include parameters, the driver caches the SQL statement and runs it directly when calling the SQLExecute function.
  • 2
    . Do not create stored procedures. The driver caches the SQL statement and runs it directly in the SQLExecute function. The driver reports syntax errors during the call to the function or when the SQLExecute function runs.
  • 3
    . Do not create stored procedures. The driver caches the SQL statement and runs it directly in the SQLExecute function. However, the driver reports syntax errors during the call to the function or when the SQLPrepare function runs.
Default value: 0
extract.log_wait_cycle_time
Specifies the number of seconds that the Extractor waits during a cycle for archived log files from each Oracle RAC to appear before starting a new cycle.
Default value: 3 seconds
extract.log_wait_cycles
Specifies the number of cycles that the Extractor runs while waiting for Oracle RAC archived logs before it exits with an error message.
Default value: 40
extract.log_wait_print_warning_cycles
Specifies the number of cycles that the Extractor runs while waiting for Oracle RAC archived logs before it prints a warning message that reports its wait status.
Default value: 5
extract.process_old_logs
For
Microsoft SQL Server
and Oracle sources, indicates whether the Extractor reads all available transaction logs. Options are:
  • 0
    . Start reading the logs from the Start Point value and skip any older logs. The default Start Point value is calculated by the Data Replication Console. You can override the default Start Point value to customize Extractor processing.
  • 1
    . Ignore the Start Point value and process all available log files.
Default value: 0
extract.rawdevice_offset
On Linux and AIX operating systems only, if you use raw devices to store data for an Oracle source database, controls the offset from the beginning of the raw device storage at which the Oracle Extractor begins reading data. The offset is expressed in bytes. The default offset depends on the operating system type. Valid values are the integers 0 through 2147483647. Data Replication interprets any negative integer value for this parameter as the default value.
Default value: -1. On Linux, the default value causes the offset to be 0 bytes. On AIX, the default value causes the offset to be 4096 bytes.
extract.show_throughput
For Oracle sources, the Extractor shows the amount of the redo logs in bytes that is processed per time unit. Options are:
  • 0
    . Show the amount of redo logs processed per time unit.
  • 1
    . Do not show the amount of redo logs processed per time unit.
Default value: 0
extract.stop_on_parsing_error
Determines the Extractor behavior when the Extractor cannot parse a transaction record in the log. Options are:
  • 0
    . Skip an inconsistent log record and continue processing. The Extractor logs a message for each skipped record and ends successfully if no other errors occur.
  • 1
    . Log an error and stop processing the transaction log.
When a parsing error occurs, the Oracle Extractor also creates a binary file with the transaction record and a .LOG file in the
DataReplication_installation
/dump subdirectory.
Default value: 1
extract.throughput_window
The number of measurements that Data Replication uses to calculate Extractor throughput. Throughput is reported in an Extractor message in the output log. Valid values are positive integers, beginning with 1.
Default value: 5
extract.verbose_level
Specifies the level of verbose output for the Extractor process. Larger values provide more detailed output. The specific output information depends on the Extractor type. Valid values are 0 to 4.
Default value: 0
global.encrypt_algorithm
Specifies the encryption algorithm that Data Replication uses to encrypt the intermediate files. This setting is also configurable in the Data Replication Console on the
Runtime Settings
tab >
General
view. In the Encryption Algorithm box, options in the
Type
list correspond to the options for this parameter. Options are:
  • 0
    . Do not encrypt the intermediate files.
  • 1
    . Use the Triple DES algorithm.
  • 2
    . Use the AES algorithm with 128-bit key.
  • 3
    . Use the AES algorithm with 192-bit key.
  • 4
    . Use the AES algorithm with 256-bit key.
Default value: 0
global.fallback_connection
When an attempt to connect to an Oracle source or target fails, indicates whether the Extractor and Applier rebuild the ODBC connection string and try to connect to the source or target again. Options are:
  • 0
    . If an attempt to connect to an Oracle source or target fails, the replication tasks do not try to connect to the source or target again. The tasks end with an error.
  • 1
    . If an attempt to connect to an Oracle source or target fails, the replication tasks rebuild the ODBC connection string and try to connect to the source or target again. The replication tasks rebuild the connection string as follows:
    • If the replication configuration does not specify a custom ODBC connection string for the source or target, the replication tasks rebuild the ODBC connection string for the second connection attempt by switching between the Oracle SID and SERVICE_NAME values.
      By default, the Extractor and Applier use the SERVICE_NAME value to build the ODBC connection string for the first connection attempt. If you select the
      Use SID instead of SERVICE_NAME
      option on the
      Source Database
      or
      Target Database
      tab, the Extractor or Applier task uses the Oracle SID to build the ODBC connection string for the first connection attempt. If the first connection attempt fails, the Extractor or Applier task switches to the other value and builds the ODBC connection string based on this value for the second connection attempt.
    • If the replication configuration specifies a custom ODBC connection string for the source or target, the replication tasks build the ODBC connection string for the second connection attempt based on the connection settings or database connection that you specify on the
      Source Database
      or
      Target Database
      tab.
Default value: 1
global.file_open_timeout
Specifies the number of seconds that a Data Replication component retries opening a file after the first open attempt fails. If this timeout period elapses without the file being opened, the component ends with an error message. This parameter affects only Data Replication components that run on Linux.
Valid values are 1 to 2147483647.
Default value: 600
global.fix_invalid_db_characters
For Amazon Redshift, Greenplum, and Teradata targets, indicates whether to fix UTF-8 encoded characters in character data that the targets cannot process correctly during InitialSync and Applier processing.
The UTF-8 encoded characters are valid but the targets do not support them.
Options are:
  • 0
    . Do not fix UTF-8 encoded symbols in the character data that the target cannot process correctly. The InitialSync and Applier tasks might fail.
  • 1
    . Replace each byte of UTF-8 encoded symbols in character data that the target cannot process correctly with the character that is specified in global.invalid_character_replacement parameter.
Default value: 0
global.fix_invalid_encoding_characters
For
Amazon Redshift, Greenplum, Netezza, Teradata, and Vertica
targets, indicates whether to fix invalid UTF-8 byte sequences in character data during InitialSync and Applier processing. Options are:
  • 0
    . Do not fix invalid UTF-8 byte sequences. The InitialSync and Applier tasks might fail.
  • 1
    . If the global.icu_enabled parameter is set to 1 or the source data contains UTF-16 characters, replace invalid UTF-8 byte sequences with the character that is specified in the global.invalid_character_replacement parameter. If the global.icu_enabled parameter is set to 0 or the source data contains only UTF-8 characters, replace each byte of the invalid UTF-8 byte sequences with the character that is specified in the global.invalid_character_replacement parameter.
Default value: 0
global.handle_os_signals
For Applier, Extractor, and InitialSync tasks that run on Linux or UNIX, indicates which error log files these tasks generate in the
DataReplication_installation
directory when they end abnormally. Options are:
  • 0
    . Generate a core file.
  • 1
    . Generate a dbsync_crash.log file.
  • 2
    . Generate a core file and dbsync_crash.log file.
Default value: 2
Change this value only at the request of Informatica Global Customer Support.
global.icu_enabled
For configurations that have a DB2 for Linux, UNIX, and Windows, Microsoft SQL Server, or Oracle source and an Amazon Redshift, Greenplum, Netezza,
Vertica,
or Teradata target, indicates whether the Applier and InitialSync convert character data from the source database encoding to the target database encoding. Options are:
  • 0
    . Do not convert the source character data to the target database encoding. Use this setting if the source and target databases use incompatible character sets and the source character data includes only Latin characters.
    If the source character data includes non-Latin characters and the source and target databases use incompatible character sets, Data Replication ends with an error regardless of the global.icu_enabled setting.
  • 1
    . Convert the source character data to the target database encoding.
Default value: 0
The Data Replication Console sets this parameter to 1 for new configurations.
global.insert_buffer_size
Specifies the maximum number of DML operations that the Applier and InitialSync accumulate in the internal buffer prior to flushing the buffer contents to the target. Each buffer stores operations of a single type.
If you use the default value of 0, the Applier and InitialSync determine the appropriate number of operations that can be buffered for the target type. The Applier and InitialSync use one of the following maximums:
  • 1 operation for
    MySQL
    and
    PostgreSQL
    targets.
  • 100 operations for Teradata targets.
  • 200 operations for
    DB2 for Linux, UNIX, and Windows,
    Greenplum, Microsoft SQL Server, Netezza,
    Vertica,
    and Oracle targets.
If a replication configuration includes mapped LOB columns, Data Replication always uses the parameter value of 1.
For
MySQL
and
PostgreSQL
targets, enter a value other than the default value only at the request of Informatica Global Customer Support.
Set this parameter to 1 to ensure correct Applier processing in the following cases:
  • You configure Data Replication to continue Applier processing when specific database errors occur.
In other cases, use the default value of 0 for optimal Applier performance. If you set this parameter to 1 or any other low non-zero value, Applier performance might be degraded.
Default value: 0
global.invalid_character_replacement
Specifies the ASCII character that replaces UTF-8 encoded characters during InitialSync and Applier processing in the following situations:
  • If you set the global.fix_invalid _db_characters parameter to 1 for Greenplum and Teradata targets, replaces UTF-8 encoded characters that the targets cannot process correctly.
  • If you set the global.fix_invalid_encoding_characters parameter to 1 for Amazon Redshift, Greenplum, Netezza, Teradata, and Vertica targets, replaces invalid UTF-8 byte sequences.
Default value: ? (question mark)
global.lob_truncation_size
Determines the maximum size of a source LOB value, in bytes, that the Applier and InitialSync replicate to the target. If the size of a LOB value exceeds this limit, the data is truncated to the specified maximum size.
Default value: 0
If you use the default value of 0, the Applier and InitialSync truncate the source LOB data to 16 KB for Merge Apply targets and to 16 MB for other targets.
  • Netezza supports row sizes up to 64 KB when the data is loaded with Netezza high-performance loaders.
  • The Extractor does not extract more than 50 MB from source LOB columns. Consequently, the maximum supported LOB size for change data capture is 50 MB, regardless of this parameter setting.
  • If a target database uses a multibyte encoding, the Applier might truncate the last character from a character LOB column, causing an incomplete byte sequence. The Applier cannot apply the incomplete byte sequence.
    To avoid this problem on Amazon Redshift, Greenplum, Netezza, Teradata, and Vertica targets, set the global.fix_invalid_db_characters and global.fix_invalid_encoding_characters runtime parameters to 1. The Applier then replaces each byte of the invalid UTF-8 byte sequence or the entire sequence with the character that is specified in the global.invalid_character_replacement parameter.
global.max_output_log_size
Specifies the maximum size of a log file, in megabytes, to which an Extractor, Applier, or InitialSync task logs messages. When this log size limit is exceeded, the Data Replication task creates a new log file.
If the time period in the global.time_for_output_log_rotation parameter elapses before this log size limit is exceeded, the Data Replication task creates a new log file.
Default value: 10
global.number_of_historical_metadata_records_kept
The maximum number of records that represent processed intermediate files in the Applier SQLite database. When the number of intermediate file records in the Applier SQLite database exceeds this value, the Applier deletes records for the earliest intermediate files.
Default value: 40000 records.
global.nz_pg_empty_strings_as_null
For Greenplum
, Vertica,
PostgreSQL,
and Netezza targets, indicates whether the Applier and InitialSync load empty strings and LOBs from the source as null values on the target. Options are:
  • 0
    . Load empty strings and LOBs from the source as empty strings on the target.
  • 1
    . Load empty strings and LOBs from the source as null values on the target.
Default value: 1
global.print_runtime_settings
Indicates whether the InitialSync, Extractor, and Applier tasks log a list of the runtime parameters for which you specified a non-default value to the execution logs each time the tasks start. Options are:
  • 0
    . Do not log the list of parameters with non-default values.
  • 1
    . Log the list of parameters with non-default values.
Default value: 1
global.print_timestamps
Indicates whether to append date and time to the logged messages that provide debug information. Options are:
  • 0
    . Do not append the date and time.
  • 1
    . Append the date and time.
Default value: 1
global.show_transaction_locking
Indicates whether the Applier and InitialSync tasks retrieve information about database locking sessions that other applications establish. Options are:
  • 0
    . The Applier and InitialSync tasks do not provide information about database locking sessions and silently wait until the sessions release the row lock to continue processing.
  • 1
    . If the Applier or InitialSync task cannot load data into a locked target row, the task retrieves information about the database locking sessions and includes this information in the processing statistics. The Applier updates processing statistics at 5-minute intervals by default. InitialSync updates processing statistics at 10-second intervals by default. To change the default interval for the Applier or InitialSync task, use the apply.print_statistics_interval or initial.print_statistics_interval runtime parameter.
Default value: 1
global.time_for_output_log_rotation
Specifies the maximum number of hours that an Extractor, Applier, or InitialSync task can write messages to its task execution log. When this time period elapses, the task creates a new log file.
If the size of a log file exceeds the global.max_output_log_size parameter value before this time period elapses, the Data Replication task creates a new log file.
Default value: 24
global.unicode_normalization
Indicates whether Data Replication converts source Unicode character data to Normalization Form C (NFC) format by using the International Components for Unicode (ICU) library. Valid values are:
  • 0
    . Do not normalize Unicode character data.
  • 1
    . Normalize Unicode character data. Use this option for Netezza targets if the source character data includes decomposed characters.
Default value: 0
If you set this parameter to 1 and the source character set matches the target character set, Data Replication converts the source data to UTF-16 for Unicode normalization. Then, the normalized Unicode character data is converted back to the source format.
initial.check_empty_tables
Indicates whether InitialSync checks mapped target tables to determine if they are empty before loading data. Options are:
  • 0
    . Do not check if the target tables are empty. InitialSync loads data to all of the mapped tables on the target.
  • 1
    . Check if the target tables are empty. If InitialSync detects that one or more tables are not empty, it skips these tables and continues to load data to other tables. If InitialSync successfully loads data to all of the empty target tables, it ends without error.
  • 2
    . Check if the target tables are empty. If InitialSync detects that one or more target tables are not empty, it skips these tables and continues to load data to other tables. After InitialSync processes all of the target tables, it returns error code 57.
For configurations with multiple sources and a single target, set this parameter to 0 to materialize the target from all of the sources.
For Oracle sources, disable the DBLinks option.
Default value: 0
initial.enforce_odbc_load
Indicates whether InitialSync uses a native database load utility or the appropriate ODBC driver to load data to the target. Options are:
  • 0
    . Use a native database load utility to load data to the target, if a load utility is available for the target type.
  • 1
    . Use the appropriate ODBC driver or the Oracle Call Interface (OCI) library to load data to the target. If you defined Tcl or SQL expressions that the target native load utility does not support, select this option so that InitialSync can properly process the expressions.
    InitialSync does not support the use of ODBC drivers for loading
    Vertica and
    Netezza targets.
Default value: 0
initial.lob_prefetch_threshold
For source tables that contain LOB columns, determines the number of rows that InitialSync gets in a single fetch. Large values provide better InitialSync performance but require a lot of memory to allocate prefetched rows. Use the following formula to estimate the allocated memory size:
allocated_memory_size
=
row_size
*
number_of_rows
Default value: 0
If you use the default value of 0, InitialSync gets the following number of rows in a single fetch:
  • 1000 rows for Oracle sources
  • 1 row for DB2 for Linux, UNIX, and Windows
    , Microsoft SQL Server
    sources
For source tables that do not contain LOB columns, use the initial.select_buffer_size runtime parameter.
initial.named_transactions
For
DB2,
Microsoft SQL Server and Oracle targets, controls whether InitialSync uses a specific transaction or application name for the transactions that load data to targets during bidirectional or cascade replication. The transaction or application name can be the default name of DbSyncTransaction or an override name that you specify in the TX_NAME command line parameter. Options are:
  • 0
    . InitialSync does not use a specific transaction or application name for the transactions that insert data into targets.
  • 1
    . InitialSync uses one of the following transaction or application names:
    • For Microsoft SQL Server and Oracle targets, the default transaction name of DbSyncTransaction.
    • For DB2 targets, the default application name of DbSyncTransaction.
    • For any of these target types, an override transaction or application name that you specify with the
      TX_NAME
      command line parameter.
Default value: 1
initial.pipe_directory
Specifies the path and directory where pipes (*nix) are created when InitialSync runs. On Windows, you cannot specify this directory because pipes are created as
\\.\pipe\PipeName
. Informatica recommends that you specify this parameter for DB2 for Linux, UNIX, and Windows sources.
Default value: The Applier output directory
initial.pipe_wait_time
For
DB2,
Greenplum
, PostgreSQL
, Vertica
and Netezza targets, specifies the number of seconds that InitialSync waits for the database native load utility to connect to the pipe that InitialSync opens for loading data to the target.
Increase this parameter value if you get the following error:
Write to pipe timed out
write_size
from
bytes_to_write
by thread
thread_name
to the file
file_name
Default value: 20 seconds on Linux and UNIX, 5 seconds on Windows
initial.print_statistics_interval
For configurations with targets that use ODBC drivers, specifies the number of seconds that must elapse before Data Replication updates InitialSync processing statistics in the InitialSync output log. The statistics include the number of processed rows, the total number of rows to be processed, and the estimated time for processing to complete. If you specify 0, Data Replication does not include this information in the InitialSync output log.
Default value: 10 seconds
initial.rows_to_commit
Specifies the maximum number of Insert rows that a single InitialSync thread can commit on the target. Use this parameter to avoid large transactions that might degrade load performance. If you use the ODBC drivers to connect to targets, this parameter applies to all target database types. If you use the target native load utilities, this parameter applies to the following target types:
  • Greenplum
  • Microsoft SQL Server
  • PostgreSQL
  • Vertica
Default value: 2147483647 rows
initial.script_engine_enabled
Indicates whether to enable or disable Tcl scripts and SQL expressions during InitialSync synchronization of source and target tables. Options are:
  • 0
    . Disable Tcl scripts and SQL expressions during InitialSync synchronization. The InitialSync task does not include target columns that are mapped to virtual columns with Tcl scripts or SQL expressions in the Insert statements for loading data.
  • 1
    . Enable Tcl scripts and SQL expressions during InitialSync synchronization.
Default value: 1
initial.select_buffer_size
Specifies the buffer size, in rows, that InitialSync uses for selecting data from source tables that do not contain mapped LOB columns. For source tables that contain mapped LOB columns, use the
initial.lob_prefetch_threshold
runtime parameter.
Default value: 0
If you use the default value of 0, InitialSync uses one of the following buffer sizes:
  • 100 rows for
    Microsoft SQL Server and
    Oracle sources
  • 100 rows for DB2 sources on Windows
  • 1 row for DB2 sources on Linux and UNIX
Values greater than 200 might enhance InitialSync performance but cause higher memory consumption.
initial.verbose_level
For Oracle sources, if you use InitialSync subtask threads, specifies the verbosity level for IniitalSync output. Options are:
  • 0
    . Provides standard InitialSync output.
  • 1
    . Provides additional output for each subtask thread.
For sources other than Oracle, use the default value of 0.
Default value: 0
initialsync.abort_on_failed_table
Indicates whether the InitialSync task ends abnormally if an error occurs during initial synchronization of source and target tables. Options are:
  • 0
    . InitialSync skips processing the target table when an error occurs but continues processing the other tables.
  • 1
    . InitialSync ends with an error and does not synchronize the remaining tables.
Default value: 1
initialsync.progress_print_step
Specifies the number of rows that the Extractor, Applier, or InitialSync component must process before Data Replication updates the progress information that is displayed when you start the Extractor or InitialSync component from the Data Replication Console.
Default value: 10000
pgloader.exceptions_path
Specifies the location of the log file that contains exceptions for a failed data load operation during an InitialSync or Applier job.
Default value: The Applier output directory
pgloader.reject_max
Specifies the maximum number of rows that can fail to be loaded before the Applier or InitialSync task fails.
Default value: 0
pgloader.reject_data_path
Specifies the location of the file that contains the data that failed to be loaded during an InitialSync or Applier task.
Default value: The Applier output directory

DDL Replication Parameters

apply.skip_add_column_failed_ddl
Determines the Applier behavior when it encounters an error during replication of an ADD COLUMN operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
If the
apply.skip_all_failed_ddl
parameter is set to 1, the Applier skips the failed operation and continues processing, regardless of this parameter setting.
Default value: 0
apply.skip_all_failed_ddl
Determines the Applier behavior when it encounters an error during replication of any DDL operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
Default value: 0
apply.skip_alter_column_failed_ddl
Determines the Applier behavior when it encounters an error during replication of an ALTER COLUMN or MODIFY COLUMN operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
If the apply.skip_all_failed_ddl parameter is set to 1, the Applier skips the failed operation and continues processing, regardless of this parameter setting.
Default value: 0
apply.skip_create_index_failed_ddl
Determines the Applier behavior when it encounters an error during replication of an ADD INDEX or CREATE INDEX operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
If the
apply.skip_all_failed_ddl
parameter is set to 1, the Applier skips the failed operation and continues processing, regardless of this parameter setting.
Default value: 0
apply.skip_create_table_failed_ddl
Determines the Applier behavior when it encounters an error during replication of a CREATE TABLE operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
If the apply.skip_all_failed_ddl parameter is set to 1, the Applier skips the failed operation and continues processing, regardless of this parameter setting.
Default value: 0
apply.skip_drop_column_failed_ddl
Determines the Applier behavior when it encounters an error during replication of a DROP COLUMN operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
If the apply.skip_all_failed_ddl parameter is set to 1, the Applier skips the failed operation and continues processing, regardless of this parameter setting.
Default value: 0
apply.skip_drop_index_failed_ddl
Determines the Applier behavior when it encounters an error during replicaiton of a DROP INDEX operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
If the apply.skip_all_failed_ddl parameter is set to 1, the Applier skips the failed operation and continues processing, regardless of this parameter setting.
Default value: 0
apply.skip_drop_table_failed_ddl
Determines the Applier behavior when it encounters an error during replication of a DROP TABLE operation. Options are:
  • 0
    . Issue an error and stop Applier processing.
  • 1
    . Skip the failed operation and continue Applier processing.
If the apply.skip_all_failed_ddl parameter is set to 1, the Applier skips the failed operation and continues processing, regardless of this parameter setting.
Default value: 0

Applier Thread Distribution Parameters

apply.distribute_by_obj_size_threshold
If you enable subtask Applier threads by setting the apply.subtasks_enabled parameter to 1, this parameter determines the threshold load amount, in bytes, for the subtask Applier threads. When this threshold is reached, another subtask thread is created. The default value of 0 sets the threshold at the lowest subtask thread load.
Typically, use the default value because it allows Data Replication to calculate the threshold dynamically. However, you might need to set this value manually in some situations. For example, if a target table has very few changes and the other target tables have many changes, you might need to set the threshold to a value larger than the lowest load on the subtask thread for the table with few changes to prevent spawning multiple subtask Applier threads.
Default value: 0
apply.global_transaction_support
Indicates whether to use global transactions when applying changes to Oracle targets. Options are:
  • 0
    . Disable global transactions and commit data on each subtask thread or on each Applier thread if you do not use subtasks.
  • 1
    . Enable global transactions and perform a single commit for all Applier threads or subtask threads at the end of the apply cycle. This option might degrade performance.
  • 2
    . Enable global transactions and perform a commit on each Applier thread.
Default value: 1
apply.max_subtasks
If you set the apply.subtasks_enabled parameter to 1 to enable creation of subtask threads for each Applier task, this parameter specifies the maximum number of subtask threads that can be created. Ensure that the number of apply subtask threads is not larger than the number of target database connections.
Default value: 4
apply.subtasks_enabled
Indicates whether Data Replication can create subtask threads for each Applier task thread. The Applier can create subtasks only when loading change data to target tables in SQL Apply mode. If you enable subtask threads, Data Replication distributes changes across threads by object rather than by row. Options are:
  • 0
    . Do not use subtask threads.
  • 1
    . Use subtask threads. Select this option to provide more parallelism in Applier processing in SQL Apply mode. For example, use this option if you have few tables to which to apply many changes. In this case, enabling subtask threads for each Applier task results in better performance because Data Replication uses more threads to process the tables.
Multiple subtask threads can reduce available CPU resources for other applications.
Default value: 0

0 COMMENTS

We’d like to hear from you!