Host
IP address of the source application database server.
Port
Port of the source application database server.
Admin Database
Default administration database for the source database server. Default is master.
Admin Database Owner
Administration database owner that has DBA rights to the database, including the ability to execute DDL and access system-level objects.
Default is dbo.
Admin Login Name
Login name for the administration database user. This user does not require special permissions as it is only used to connect to the source database. You can provide any user name, such as a database connection user or a read-only user.
Password
Password for the administration login name.
Apps Database Owner
Application database owner that owns the application tables you want to archive.
The production application database owner is either the owner of the tables that you want to archive, or is the user that has full rights to the tables that you want to archive. During the archive process, you can use the production database owner to delete from the source. Alternatively, you can use the staging database owner to delete from the source.
Use the default database owner for the application or provide another value if you do not use the default owner. If you provide an owner that does not own or has full access to the application tables, then the archive job fails because the job cannot find the table or view.
The archive job uses the number of schemas in the source application to determine schema names. The archive job uses this attribute value for applications that have one schema. The archive job uses the schema name from the mined application for applications that have multiple schemas.
The ILM engine uses this field to generate the SQL SELECT statement to read data from the application tables.
Default is dbo.
Application Database
Database that contains the application tables that you want to archive.
Application Login Name
Login name that connects to the source database that contains the data you want to archive. This user does not require special permissions as it is only used for the database connection. You can provide any user name, such as a database connection user or a read-only user.
Password
Password for the application login name.
Staging Database
Staging database that stores interim tables for business rule validation and generates candidates during the archive process. The database also stores staging tables if you enable staging for the archive source connection.
Staging Database Owner
Staging database owner that has privileges to create tables in the staging database. The staging database owner creates interim tables. If you enable staging for the archive source connection, the staging owner also creates staging tables. The staging tables temporarily store a copy of the archived data before the archive job moves the data to the archive or history database. The staging owner can delete data from the source. Or, you can use the application owner to delete from the source.
Default is dbo.
Staging Login Name
Login name for the staging database.
Password
Password for the staging login name.
Staging Filegroup
Tablespace in the staging database user that stores the interim and staging tables when you run an archive cycle. The archive job always stores the interim tables in this tablespace. The archive job stores staging tables if you enable staging.
Use Staging
Determines whether the archive job temporarily stores data in a staging area on the archive source before it moves data to the target. The archive job includes heavy select and delete operations on the tables that are selected in the archive job. The operations may impact performance on any user operations on the tables. For example, if you run an archive job that archives purchase orders, users may see a performance impact when they create a purchase order. The main benefit of staging is to reduce the number of operations on the production tables.
If enabled, the archive job copies data to the staging area after the archive job evaluates the business rules. The archive job copies data that meets the business rules from the archive source to the staging area. Then, the job deletes data from the source and moves data from the staging area to the target system.
If disabled, the job extracts data directly from the archive source to the archive target. The extraction and deletion occur simultaneously. You may want to disable staging if you do not want to create additional storage space on the production database.
JDBC Fetch Size
Controls the number of rows that the system fetches in one batch from one table if you use JDBC to connect to the source database.
Only change the default value if you receive an out-of-memory error when the archive job copies data to the destination. You may get the out-of-memory error if the source table has large objects in each row, such as CLOBS or BLOBS data types. If you receive this error, then reduce the fetch size.
Linked Server Name to Home or Database Link to ILM Repository
Database link name that connects the source database to the ILM repository. The ILM repository is commonly referred to as the home schema. This attribute is valid if the entity that is included in the archive job has run procedure steps. For example, the entity may have a run procedure step to call an external routine.
When the archive job runs procedures, the system uses this attribute for the database link value. If you do not provide a value, then the system dynamically creates the database link from the source to the ILM repository. After the archive job runs procedures, the system drops the database link.
If the system needs to create the database link, then the administration or application database users need to have create and drop database link permissions.
Linked Server Name to Production (Restore Only) or Database Link to Production (Restore Only)
Database link name that connects the history database to the production database. This attribute is required when you create the history database as a source and you use transaction or cycle restore. For restore scenarios, the history database is the source, and the production database is the target.
During the restore job, the system uses this attribute for the database link value. If you do not provide a value, then the system dynamically creates the database link from the source to the ILM repository. After the restore job completes, the system drops the database link.
If the system needs to create the database link, then the administration or application database users need to have create and drop database link permissions.
Disable Triggers
Determines whether the system disables insert, update, and delete triggers when the archive job deletes rows from tables.
If enabled, the system disables triggers when the archive job deletes data from the source.
If disabled, the system retains the triggers.
Target Attachment Location
Target location for external attachments. Enter the location where you want to archive the external attachments to. You must have write access to the directory.
After the retirement job moves attachments to this location, you can use this location as the final destination that stores the attachments. Or, you can move the attachments from this location to a different file system, external storage, or the Data Vault.
Source/Staging Attachment Location
Source location of the external attachments. Enter the current location where the attachments exist. You must have read access to the directory.
For SAP application retirement, enter the location of the SAP application server file system. The retirement job uses an SAP function module to generate BCP files for data in transparent HR and STXL tables, ADK files, and attachments.
Enter the full path of the location. For example,
\\10.1.10.10\interfaces\CCO\
.
The location must be accessible to the SAP application server and in the same drive where the ILM engine is installed. If the SAP system and Data Archive are on the same operating systems, the path is the same as the
Staging Directory
property in the Data Vault target connection.
If the SAP system and Data Archive are on different operating systems, then the paths are different.
Staging Script Location
Temporary location that stores the scripts that the archive job generates to move external attachments from the source to the target.
Enter a location that you have read and write access to. For Siebel attachments, enter a location that is accessible to the Data Vault Service for External Attachments server.
This attribute only applies to external attachments.
Move Attachments in Synchronous Mode
Determines whether the archive job automatically archives external attachments or whether you run a standalone job to move the attachments after the archive job completes. If you provide a source file location for attachments, the archive job creates SQL scripts in the file server location and associates the scripts with the archive job ID.
If enabled, the archive job runs the scripts during the archive process. The run procedures configuration in the entity determines when the archive job archives the attachments.
If disabled, you must initiate the movement after the archive job completes. You can manually run the scripts in the file server location or you can run a standalone job to move the attachments. If you run the standalone job, you must provide the archive job ID. The job then looks for the scripts that are associated to the archive job ID.
This attribute only applies to external attachments.
Compile ILM Functions
Determines if the system compiles user-defined functions during the archive job.
By default, the staging database user needs privileges to compile user-defined functions on Microsoft SQL Server. The user-defined functions are required to use application metadata to archive data. Application retirement does not use metadata to archive data. Therefore, the staging user does not need any additional privileges besides select permissions, to compile the user-defined functions.
You must clear the
Compile ILM Functions
check box before you start a retirement job on a Microsoft SQL Server source application. This ensures that the staging database user has read-only access and will not be able to modify the source application.
Parallel Entity Run
Determines the interim table names in the entity if you want to run the entity in parallel archive jobs. For example, you may extract data in parallel from two different organizations in the same human resource entity. The system creates interim tables in the staging database user. However, the tables need to be differentiated so the archive jobs know which interim tables are relevant for that job.
By default, the system creates interim table names with the same name in EDM. If you enable this attribute, then the system creates the interim table names with the following syntax:
If you enable this attribute and the staging database user does not have authorization to create tables, then the archive job fails.
SAP Fetch Size
Number of rows that the retirement job extracts at a time from the SAP cluster and pool tables to write to the BCP file.
Required for SAP application retirement only. Default is 2,000 rows.
SAP Host
Host of the SAP application that you want to retire.
Required for SAP application retirement only.
SAP Client
Client in which the user logs in. Note that all clients in the SAP application are retired.
Required for SAP application retirement only.
SAP System Number
System number in which the user logs in.
Required for SAP application retirement only.
SAP Language
Language in which the user logs in. Note that all languages in the SAP application are retired.
Required for SAP application retirement only.
SAP User
User that logs in to the SAP application. The user must be assigned to the ZINFA_RETIREMENT_PREPARATION role and include RFC connection authorizations.
Required for SAP application retirement only.
SAP User Password
Password for the SAP user.
Required for SAP application retirement only.
Use Imported Schema Name
The name of the schema when you imported the table metadata from the source.
If you want to archive an entity that contains tables from multiple schemas, you must select the
Use Imported Schema Name
check box. If you do not, the archive job fails at the generate candidates step.
If you are upgrading Data Archive from a version before 6.2 HotFix 2, do not select the check box. Review the Application Database property to confirm that you are connecting to the correct database.
FTP User
User name to connect to the FTP server.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Password
Password for the FTP user.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Host
Host name of the FTP server.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Port
Port number of the FTP server. Default port is 21.
When you specify a port number, enable that port number for FTP on the host machine.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Folder Location
Name of the FTP folder on the Data Archive server. For example,
ERP\
.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
Compressed
For SAP application retirement projects, if you want Data Archive to save the BCP files in a compressed file format, enable the check box. When you enable the check box, Data Archive saves the BCP files as
.gz
files in the staging area.