Table of Contents

Search

  1. Preface
  2. Data Replication Overview
  3. Understanding Data Replication
  4. Sources - Preparation and Replication Considerations
  5. Targets - Preparation and Replication Considerations
  6. Starting the Server Manager
  7. Getting Started with the Data Replication Console
  8. Defining and Managing Server Manager Main Servers and Subservers
  9. Creating and Managing User Accounts
  10. Creating and Managing Connections
  11. Creating Replication Configurations
  12. Materializing Targets with InitialSync
  13. Scheduling and Running Replication Tasks
  14. Implementing Advanced Replication Topologies
  15. Monitoring Data Replication
  16. Managing Replication Configurations
  17. Handling Replication Environment Changes and Failures
  18. Troubleshooting
  19. Data Replication Files and Subdirectories
  20. Data Replication Runtime Parameters
  21. Command Line Parameters for Data Replication Components
  22. Updating Configurations in the Replication Configuration CLI
  23. DDL Statements for Manually Creating Recovery Tables
  24. Sample Scripts for Enabling or Disabling SQL Server Change Data Capture
  25. Glossary

Preparing Cloudera and Hortonworks Target Systems

Preparing Cloudera and Hortonworks Target Systems

To replicate change data to Cloudera and Hortonworks targets on a Hadoop Distributed File System (HDFS), you must complete several prerequisite tasks to prepare the systems where the Applier and Data Replication Console run.
  1. Install the 64-bit Java Development Kit (JDK) 1.7 or 1.8 if you have not done so already.
    For Cloudera or Hortonworks targets that use Kerberos authentication, ensure that the JDK 1.7u65 or later is installed.
  2. Define the JAVA_HOME environment variable to point to the root Java installation directory.
  3. Add a Java library to the system path.
    • On Windows, add the directory that contains the jvm.dll library to the PATH environment variable. For example, use the following command:
      PATH=%PATH%;%JAVA_HOME%\jre\bin\server
    • On Linux and UNIX, add the directory that contains the libjvm.so library to the library path environment variable for your operating system. The library path environment variables are:
      • LD_LIBRARY_PATH for HP-UX and Linux systems
      • LD_LIBRARY_PATH_64 for Solaris systems
      • LIBPATH for AIX systems
      For example, use the following command:
      LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$JAVA_HOME/jre/lib/amd64/server
  4. On AIX, add the network API library, libnet.so, to the LIBPATH environment variable. Use the following command:
    LIBPATH=$LIBPATH:$JAVA_HOME/jre/lib/ppc_64
  5. On Windows, install WinUtils by performing the following substeps:
    1. Download WinUtils from the following web site:
    2. Extract the .zip file.
    3. Add the bin subdirectory that contains the WinUtils executable file and required .dll libraries to the HADOOP_HOME environment variable.
  6. On Windows, for Cloudera or Hortonworks targets that use Kerberos authentication, define the DBSYNC_KERBEROS_CACHE_NAME environment variable. The environment variable points to the file that contains Kerberos credential cache.
    You can get the path to the Kerberos credential cache folder from the KRB5CCNAME environment variable.
  7. Download the hadoop_libs.zip file that Data Replication provides and that contains the .jar files. Extract this zip file into the
    DataReplication_installation
    directory.
  8. Verify that the
    DataReplication_installation
    /lib directory contains the hadoop subdirectory.
  9. For Cloudera and Hortonworks targets, copy the following configuration files to the
    DataReplication_installation
    /lib/hadoop/
    hadoop_distribution
    directory:
    • hdfs-site.xml
    • core-site.xml
    • yarn-site.xml
      The yarn-site.xml file is required only if the target uses HDFS high availability.

0 COMMENTS

We’d like to hear from you!