Host
IP address of the source application database server. When the source application server is in an Oracle Real Application Cluster, enter the IP address of node1 and the IP address of node 2, separated by a comma.
Port
Port of the source application database server. When the source application server is in an Oracle RAC, enter the port number of node 1 and port number of node 2, separated by a comma.
Service Name
Unique identifier or system identifier for the source application database server.
When you define an Oracle RAC source database server, enter the service name.
Admin Schema Name
Default administration database user for the source database server, such as SYSTEM.
The administration database user has DBA rights to the database, including the ability to run DDL and access system-level objects.
Admin Login Name
Admin Login Name is the same as the schema name.
Password
Password for the administration login name.
Apps Schema Name
Database user that owns the application tables that you want archive, such as APPS for Oracle applications.
The production application user is either the owner of the tables that you want to archive or partition, or the user that has full rights to the tables that you want to archive or partition. During the archive process, you can use the production application user to delete from the source. Alternatively, you can configure the source connection to use the staging user to delete data from the source.
Use the default database user for the application or provide another value if you do not use the default user. If you provide a user that does not own or has full access to the application tables, then the archive job fails because the job cannot find the table or view.
The archive job uses the number of schemas in the source application to determine schema names. The archive job uses this attribute value for applications that have one schema, such as PeopleSoft and Siebel. The archive job uses the schema name from the mined application for applications that have multiple schemas, such as Oracle applications.
The ILM engine uses this field to generate the SQL SELECT statement to read data from the application tables. For example, if you enter SYSADMIN, the archive job generates the following SQL:
SELECT * from SYSADMIN.<OBJECT NAME>
Application Login Name
Login name that connects to the source database that contains the data you want to archive. This user does not require special permissions as it is only used for the database connection. You can provide any user name, such as a database connection user or a read-only user.
Password
Password for the application login name.
Staging Schema Name
Staging database user for the source database.
The staging database user stores interim tables for business rule validation and generates candidates during the archive and partitioning processes. If you enable staging for the archive source connection, the staging user also creates staging tables. The staging tables temporarily store a copy of the archived data before the archive job moves the data to the archive or history database. The staging user typically resides in the source database.
Staging Login Name
Login name for the staging database.
Password
Password for the staging login name.
Staging Tablespace
Tablespace in the staging database user that stores the interim and staging tables when you run an archive cycle. The archive job always stores the interim tables in this tablespace. The archive job stores staging tables if you enable staging.
Use Staging
Determines whether the archive job temporarily stores data in a staging area on the archive source before it moves data to the target. The archive job includes heavy select and delete operations on the tables that are selected in the archive job. The operations may impact performance on any user operations on the tables. For example, if you run an archive job that archives purchase orders, users may see a performance impact when they create a purchase order. The main benefit of staging is to reduce the number of operations on the production tables.
If enabled, the archive job copies data to the staging area after the archive job evaluates the business rules. The archive job copies data that meets the business rules from the archive source to the staging area. Then, the job deletes data from the source and moves data from the staging area to the target system.
If disabled, the job extracts data directly from the archive source to the archive target. The extraction and deletion occur simultaneously. You may want to disable staging if you do not want to create additional storage space on the production database.
Use Row ID for Delete
Determines how the archive job removes data from the source database. This attribute improves performance when the archive job deletes data from the source.
If enabled, the system uses the row ID and the primary key to delete records from the source database. The system creates temporary tables in the staging database user to store the row ID and the row number or primary key. To create temporary tables in the staging database user, you must enable staging for the connection. This option is the fastest method for deletion.
If disabled, the system deletes records based on the primary key or the WHERE clause from the archive job. If the Enterprise Data Manager has a primary key defined for the table, the system uses the primary key to delete records for the table. If the Enterprise Data Manager does not have a primary key defined for the table, the system uses the same WHERE clause that the archive job generated to extract the archive data from the source.
JDBC Fetch Size
Controls the number of rows that the system fetches in one batch from one table if you use JDBC to connect to the source database.
Only change the default value if you receive an out-of-memory error when the archive job copies data to the destination. You may get the out-of-memory error if the source table has large objects in each row, such as CLOBS or BLOBS data types. If you receive this error, then reduce the fetch size.
Use Row ID for File Gen
Determines how the archive job generates the file from the source database. If you provide multiple threads to extract from the same table, this attribute improves the archive job performance when the archive job copies data to the destination.
If enabled, the system issues multiple SELECT statements against the same table. The system creates temporary tables in the staging database user to store the row ID and the row number to control what each worker thread processes. The staging database user requires write access or both the create table and create object privileges. Disable this option if you cannot grant the privileges to the staging database user.
If disabled, the system uses one worker thread for each table.
Use Oracle Parallel DML for Delete
Determines if the system uses Oracle or Java workers when the archive job deletes data from the source. This attribute only applies if you use the row ID and primary key for deletion.
If enabled, the system uses the Oracle parallel Data Manipulation Language (DML) for delete operations. Oracle parallel DML locks the source tables during deletion. During the deletion, no other insert or update operations are allowed on the tables. This option may improve performance. You may want to use this option if you archive during system downtime or lower activity periods.
If disabled, the system uses Java workers to delete data from the source. Java workers do not lock the tables during deletion.
Database Link to ILM Repository
Database link name that connects the source database to the ILM repository. The ILM repository is commonly referred to as the home schema. This attribute is valid if the entity that is included in the archive job has run procedure steps. For example, the entity may have a run procedure step to call an external routine.
When the archive job runs procedures, the system uses this attribute for the database link value. If you do not provide a value, then the system dynamically creates the database link from the source to the ILM repository. After the archive job runs procedures, the system drops the database link.
If the system needs to create the database link, then the administration or application database users need to have create and drop database link permissions.
Database Link to Production (Restore Only)
Database link name that connects the history database to the production database. This attribute is required when you create the history database as a source and you use transaction or cycle restore. For restore scenarios, the history database is the source, and the production database is the target.
During the restore job, the system uses this attribute for the database link value. If you do not provide a value, then the system dynamically creates the database link from the source to the ILM repository. After the restore job completes, the system drops the database link.
If the system needs to create the database link, then the administration or application database users need to have create and drop database link permissions.
Transactional Commit (Restore Only)
Determines when the system issues commits for restore jobs.
If enabled, the system issues a single commit after it restores all tables.
If disabled, the system issues a commit after it restores each table.
Use Staging User for Deletes
Determines which database user the archive job uses to delete data from the source database.
If enabled, the archive job uses the staging database user to delete from the source.
If disabled, the archive job uses the application database user to delete from the source.
Whether you enable this option depends on the authorizations that the application or staging database users have. For example, you may want to use the staging user for deletes if you have a read-only application database user.
Disable Triggers
Determines whether the system disables insert, update, and delete triggers when the archive job deletes rows from tables.
If enabled, the system disables triggers when the archive job deletes data from the source.
If disabled, the system retains the triggers.
Target Attachment Location
Target location for external attachments. Enter the location where you want to archive the external attachments to. You must have write access to the directory.
After the retirement job moves attachments to this location, you can use this location as the final destination that stores the attachments. Or, you can move the attachments from this location to a different file system, external storage, or the Data Vault.
Source/Staging Attachment Location
Source location of the external attachments. Enter the current location where the attachments exist. You must have read access to the directory.
For SAP application retirement, enter the location of the SAP application server file system. The retirement job uses an SAP function module to generate BCP files for data in transparent HR and STXL tables, ADK files, and attachments.
Enter the full path of the location. For example,
\\10.1.10.10\interfaces\CCO\
.
The location must be accessible to the SAP application server and in the same drive where the ILM engine is installed. If the SAP system and Data Archive are on the same operating systems, the path is the same as the
Staging Directory
property in the Data Vault target connection.
If the SAP system and Data Archive are on different operating systems, then the paths are different.
Staging Script Location
Temporary location that stores the scripts that the archive job generates to move external attachments from the source to the target.
Enter a location that you have read and write access to. For Siebel attachments, enter a location that is accessible to the Data Vault Service for External Attachments server.
This attribute only applies to external attachments.
Move Attachments in Synchronous Mode
Determines whether the archive job automatically archives external attachments or whether you run a standalone job to move the attachments after the archive job completes. If you provide a source file location for attachments, the archive job creates SQL scripts in the file server location and associates the scripts with the archive job ID.
If enabled, the archive job runs the scripts during the archive process. The run procedures configuration in the entity determines when the archive job archives the attachments.
If disabled, you must initiate the movement after the archive job completes. You can manually run the scripts in the file server location or you can run a standalone job to move the attachments. If you run the standalone job, you must provide the archive job ID. The job then looks for the scripts that are associated to the archive job ID.
This attribute only applies to external attachments.
Parallel Entity Run
Determines the interim table names in the entity if you want to run the entity in parallel archive jobs. For example, you may extract data in parallel from two different organizations in the same human resource entity. The system creates interim tables in the staging database user. However, the tables need to be differentiated so the archive jobs know which interim tables are relevant for that job.
By default, the system creates interim table names with the same name in EDM. If you enable this attribute, then the system creates the interim table names with the following syntax:
If you enable this attribute and the staging database user does not have authorization to create tables, then the archive job fails.
Create Interim Table with No Logging Mode
Determines if the archive job creates interim tables without logging.
If enabled, the archive job creates the interim tables in no logging mode. You may want to enable if you have a limited amount of logging space in the source. However, if you create tables without logging, then recovery ability is at risk.
If disabled, the archive job creates the interim tables with logging.
Maintain Source Compression on Target
When you archive to an Oracle target, determines if the archive job creates target tables with the same compression as the source.
If enabled, the archive job creates target tables with the same compression as the source tables if the source tables have COMPRESS FOR DIRECT_LOAD OPERATIONS table compression.
If disabled, the archive job does not create target tables with compression.
Maintain Source Partitions on Target
When you archive to an Oracle target, determines if the archive job creates target tables with the same partitioning as the source.
If enabled, the archive job creates target tables with partitioning if the source tables include list and range partition types.
If disabled, the archive job does not create target tables with partitioning.
SAP Fetch Size
Number of rows that the retirement job extracts at a time from the SAP cluster and pool tables to write to the BCP file.
Required for SAP application retirement only. Default is 2,000 rows.
SAP Host
Host of the SAP application that you want to retire.
Required for SAP application retirement only.
SAP Client
Client in which the user logs in. Note that all clients in the SAP application are retired.
Required for SAP application retirement only.
SAP System Number
System number in which the user logs in.
Required for SAP application retirement only.
SAP Language
Language in which the user logs in. Note that all languages in the SAP application are retired.
Required for SAP application retirement only.
SAP User
User that logs in to the SAP application. The user must be assigned to the ZINFA_RETIREMENT_PREPARATION role and include RFC connection authorizations.
Required for SAP application retirement only.
SAP User Password
Password for the SAP user.
Required for SAP application retirement only.
Use Imported Schema Name
The name of the schema when you imported the table metadata from the source.
You must select
Use Imported Schema Name
when the tables in the entity are from more than one schema.
Enable Smart Partitioning
Enables the smart partitioning feature for Oracle 10g, 11g, and 12c sources.
Required if you want to use smart partitioning to create segments from an application database.
FTP User
User name to connect to the FTP server.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Password
Password for the FTP user.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Host
Host name of the FTP server.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Port
Port number of the FTP server. Default port is 21.
When you specify a port number, enable that port number for FTP on the host machine.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Folder Location
Name of the FTP folder on the Data Archive server. For example,
ERP\
.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
Compressed
For SAP application retirement projects, if you want Data Archive to save the BCP files in a compressed file format, enable the check box. When you enable the check box, Data Archive saves the BCP files as
.gz
files in the staging area.