Table of Contents

Search

  1. Preface
  2. Starting Data Archive
  3. System Configuration
  4. Database Users and Privileges
  5. Source Connections
  6. Target Connections
  7. Archive Store Configuration
  8. Datatype Mapping
  9. Database Optimization
  10. SAP Application Retirement
  11. z/OS Source Data Retirement
  12. Seamless Data Access
  13. Data Discovery Portal
  14. Security
  15. SSL Communication with Data Vault
  16. LDAP User Authentication
  17. Auditing
  18. Running Jobs from External Applications
  19. Salesforce Archiving Administrator Tasks
  20. Upgrading Oracle History Data
  21. Upgrading PeopleSoft History Data
  22. Data Archive Maintenance
  23. Appendix A: Datetime and Numeric Formatting
  24. Appendix B: Data Archive Connectivity

Administrator Guide

Administrator Guide

IBM DB2 Source Connections

IBM DB2 Source Connections

When you define an archive source connection for applications on IBM DB2 databases, you can choose from multiple connection types. The connection type that you choose depends on the database version.
Choose one of the following connection types to connect to an IBM DB2 database:
  • DB2_Adapter z/OS. Use to connect to IBM DB2 on z/OS.
  • DB2_Adapter UDB. Use to connect IBM DB2 on AIX, Linux, UNIX, or Windows (LUX).
  • DB2_AS400_Adapter. Use to connect to IBM DB2 on AS/400.
The properties that you configure depend on the connection type that you choose. Some properties are not relevant for all connection types. Property names may vary between connection types.
When you use IBM DB2 native utilities for data movement, the archive job uses the connections that you configure in IBM DB2 Connect to connect to the source and target databases. When you run an archive job, the job gets the DB2 connect connection name from the archive source and target connections. Then, the job uses connection name to get the connection details from DB2 Connect.
To use IBM DB2 native utilities for data movement, create connections to the source and target in IBM DB2 Connect first. Then, create source and target connections in Data Archive and add the DB2 Connect connection name.
Depending on the connection type, you can configure the following source connection properties:
Host
IP address of the source application database server.
Port
Port of the source application database server.
Database or Service Name
Unique identifier or system identifier for the source application database server.
If the database is on AS/400 and includes CHAR for BIT DATA data types, append the following code to the service name:
CharsetFor65535=<code page>
For example, if your service name is
AU4SNDBX
and your code page is
CP037
. Enter the following code for the service name:
AU4SNDBX;CharsetFor65535=CP037
Admin Schema or ILM Repository Administrator User
Default administration database user for the source database server, such as SYSTEM.
The administration database user has DBA rights to the database, including the ability to run DDL and access system-level objects.
Admin User Name or Admin Login Name
Login name for the administration database user. This user does not require special permissions as it is only used to connect to the source database. You can provide any user name, such as a database connection user or a read-only user.
Password
Password for the administration login name.
Apps Schema or Application User Name
Database user that owns the application tables that you want archive, such as APPS for Oracle applications.
The production application user is either the owner of the tables that you want to archive or partition, or the user that has full rights to the tables that you want to archive or partition. During the archive process, you can use the production application user to delete from the source. Alternatively, you can configure the source connection to use the staging user to delete data from the source.
Use the default database user for the application or provide another value if you do not use the default user. If you provide a user that does not own or has full access to the application tables, then the archive job fails because the job cannot find the table or view.
The archive job uses the number of schemas in the source application to determine schema names. The archive job uses this attribute value for applications that have one schema, such as PeopleSoft and Siebel. The archive job uses the schema name from the mined application for applications that have multiple schemas, such as Oracle applications.
The ILM engine uses this field to generate the SQL SELECT statement to read data from the application tables. For example, if you enter SYSADMIN, the archive job generates the following SQL:
SELECT * from SYSADMIN.<OBJECT NAME>
Apps User Name or Application Login Name
Login name that connects to the source database that contains the data you want to archive. This user does not require special permissions as it is only used for the database connection. You can provide any user name, such as a database connection user or a read-only user.
Password
Password for the application login name.
Staging Schema or Staging User Name
Staging database user for the source database.
The staging database user stores interim tables for business rule validation and generates candidates during the archive and partitioning processes. If you enable staging for the archive source connection, the staging user also creates staging tables. The staging tables temporarily store a copy of the archived data before the archive job moves the data to the archive or history database. The staging user typically resides in the source database.
Staging User Name or Staging Login Name
Login name for the staging database.
Password
Password for the staging login name.
Staging Tablespace
Tablespace in the staging database user that stores the interim and staging tables when you run an archive cycle. The archive job always stores the interim tables in this tablespace. The archive job stores staging tables if you enable staging.
For IBM DB2 databases, verify that the staging tablespace is segmented and the tables in each segment are not partitioned.
Index Storage Group or Index Tablespace
Tablespace in the staging database user that stores the interim table indexes when you run an archive cycle.
Use Staging
Determines whether the archive job temporarily stores data in a staging area on the archive source before it moves data to the target. The archive job includes heavy select and delete operations on the tables that are selected in the archive job. The operations may impact performance on any user operations on the tables. For example, if you run an archive job that archives purchase orders, users may see a performance impact when they create a purchase order. The main benefit of staging is to reduce the number of operations on the production tables.
If enabled, the archive job copies data to the staging area after the archive job evaluates the business rules. The archive job copies data that meets the business rules from the archive source to the staging area. Then, the job deletes data from the source and moves data from the staging area to the target system.
If disabled, the job extracts data directly from the archive source to the archive target. The extraction and deletion occur simultaneously. You may want to disable staging if you do not want to create additional storage space on the production database.
If your source data contains the Decfloat data type, do not enable
Use Staging
.
Default is disabled.
JDBC Fetch Size
Controls the number of rows that the system fetches in one batch from one table if you use JDBC to connect to the source database.
Only change the default value if you receive an out-of-memory error when the archive job copies data to the destination. You may get the out-of-memory error if the source table has large objects in each row, such as CLOBS or BLOBS data types. If you receive this error, then reduce the fetch size.
Default is 1000.
Database Link to Production (Restore Only)
Database link name that connects the history database to the production database. This attribute is required when you create the history database as a source and you use transaction or cycle restore. For restore scenarios, the history database is the source, and the production database is the target.
During the restore job, the system uses this attribute for the database link value. If you do not provide a value, then the system dynamically creates the database link from the source to the ILM repository. After the restore job completes, the system drops the database link.
If the system needs to create the database link, then the administration or application database users need to have create and drop database link permissions.
Use Staging User for Deletes
Determines which database user the archive job uses to delete data from the source database.
If enabled, the archive job uses the staging database user to delete from the source.
If disabled, the archive job uses the application database user to delete from the source.
Whether you enable this option depends on the authorizations that the application or staging database users have. For example, you may want to use the staging user for deletes if you have a read-only application database user.
Default is disabled.
Target Attachment Location
Target location for external attachments. Enter the location where you want to archive the external attachments to. You must have write access to the directory.
After the retirement job moves attachments to this location, you can use this location as the final destination that stores the attachments. Or, you can move the attachments from this location to a different file system, external storage, or the Data Vault.
Source/Staging Attachment Location
Source location of the external attachments. Enter the current location where the attachments exist. You must have read access to the directory.
For SAP application retirement, enter the location of the SAP application server file system. The retirement job uses an SAP function module to generate BCP files for data in transparent HR and STXL tables, ADK files, and attachments.
Enter the full path of the location. For example,
\\10.1.10.10\interfaces\CCO\
.
The location must be accessible to the SAP application server and in the same drive where the ILM engine is installed. If the SAP system and Data Archive are on the same operating systems, the path is the same as the
Staging Directory
property in the Data Vault target connection.
If the SAP system and Data Archive are on different operating systems, then the paths are different.
Staging Script Location
Temporary location that stores the scripts that the archive job generates to move external attachments from the source to the target.
Enter a location that you have read and write access to. For Siebel attachments, enter a location that is accessible to the Data Vault Service for External Attachments server.
This attribute only applies to external attachments.
Move Attachments in Synchronous Mode
Determines whether the archive job automatically archives external attachments or whether you run a standalone job to move the attachments after the archive job completes. If you provide a source file location for attachments, the archive job creates SQL scripts in the file server location and associates the scripts with the archive job ID.
If enabled, the archive job runs the scripts during the archive process. The run procedures configuration in the entity determines when the archive job archives the attachments.
If disabled, you must initiate the movement after the archive job completes. You can manually run the scripts in the file server location or you can run a standalone job to move the attachments. If you run the standalone job, you must provide the archive job ID. The job then looks for the scripts that are associated to the archive job ID.
This attribute only applies to external attachments.
SAP Fetch Size
Number of rows that the retirement job extracts at a time from the SAP cluster and pool tables to write to the BCP file.
Required for SAP application retirement only. Default is 2,000 rows.
SAP Host
Host of the SAP application that you want to retire.
Required for SAP application retirement only.
SAP Client
Client in which the user logs in. Note that all clients in the SAP application are retired.
Required for SAP application retirement only.
SAP System Number
System number in which the user logs in.
Required for SAP application retirement only.
SAP Language
Language in which the user logs in. Note that all languages in the SAP application are retired.
Required for SAP application retirement only.
SAP User
User that logs in to the SAP application. The user must be assigned to the ZINFA_RETIREMENT_PREPARATION role and include RFC connection authorizations.
Required for SAP application retirement only.
SAP User Password
Password for the SAP user.
Required for SAP application retirement only.
Parallel Entity Run
Determines the interim table names in the entity if you want to run the entity in parallel archive jobs. For example, you may extract data in parallel from two different organizations in the same human resource entity. The system creates interim tables in the staging database user. However, the tables need to be differentiated so the archive jobs know which interim tables are relevant for that job.
By default, the system creates interim table names with the same name in EDM. If you enable this attribute, then the system creates the interim table names with the following syntax:
<table name>_<job ID>
If you enable this attribute and the staging database user does not have authorization to create tables, then the archive job fails.
Default is disabled.
IBM DB2 Connect Data Source Name
Data source name that you configured for the source connection in IBM DB2 Connect. Required if you use IBM DB2 native utilities for data movement. The archive job uses the connections that you configure in IBM DB2 Connect to connect to the source and target databases. When you run an archive job, the job gets the IBM DB2 connect connection name from the archive source and target connections. Then, the job uses the connection name to get the connection details from IBM DB2 Connect.
If you do not configure a data source name and you use IBM DB2 native utilities for data movement, the archive job fails. The job can only connect to the database from IBM DB2 Connect.
Use Imported Schema Name
The name of the schema when you imported the table metadata from the source.
You must select
Use Imported Schema Name
when the tables in the entity are from more than one schema.
FTP User
User name to connect to the FTP server.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Password
Password for the FTP user.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Host
Host name of the FTP server.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Port
Port number of the FTP server. Default port is 21.
When you specify a port number, enable that port number for FTP on the host machine.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
FTP Folder Location
Name of the FTP folder on the Data Archive server. For example,
ERP\
.
Required for SAP application retirement if you set up an FTP connection between the SAP application server and Data Archive.
Compressed
For SAP application retirement projects, if you want Data Archive to save the BCP files in a compressed file format, enable the check box. When you enable the check box, Data Archive saves the BCP files as
.gz
files in the staging area.

0 COMMENTS

We’d like to hear from you!