Table of Contents

Search

  1. Preface
  2. Installation Overview
  3. Before You Begin
  4. Pre-Installation Tasks
  5. Data Integration Hub Installation
  6. Post-Installation Tasks
  7. Upgrading Data Integration Hub
  8. Starting and Stopping Data Integration Hub
  9. Optional Data Integration Hub Configuration
  10. Installing and Configuring the Data Integration Hub Accelerator for Data Archive
  11. Troubleshooting the Data Integration Hub Installation
  12. Uninstallation

Installation and Configuration Guide

Installation and Configuration Guide

Troubleshooting the Data Integration Hub Hadoop Service

Troubleshooting the Data Integration Hub Hadoop Service

Consider the following troubleshooting tips when you install the
Data Integration Hub
Hadoop Service:
When I test the DIH__STAGING__HADOOP connection the test succeeds but when I try to run a publication the running fails.
The following example shows a sample error message:
Wrong FS: hdfs://10.40.40.96:8020/user/infa_user/TEST_FILTER_87003, expected: hdfs://tavdxcdh53n1:8020
In the
Data Integration Hub
Operation Console, in the Connections page, edit the DIH__STAGING__HADOOP connection so that
NameNode URI
is identical to the setting of the property
fs.default.name
or the property
fs.defaultFS
in the
core-site.xml
file.
When I test the DIH__STAGING__HADOOP connection the test fails with the following error:
Data Access Connection test failed with the following error:DXServerException: Cannot establish connection to Apache Spark
.
Perform the following actions:
  1. Check the
    Data Integration Hub
    Hadoop Service log file for additional details. For more information, see Default Log Files.
  2. Verify that Apache Spark is running.
  3. On the machine where the
    Data Integration Hub
    Hadoop Service is installed, open the
    dx-configuration.properties
    file in a text editor from the following location:
    <
    DIH
    InstallationDir>/DataIntegrationHub/tomcat/shared/classes/
When I test the DIH__STAGING__HADOOP connection the test fails with the following error:
Data Access Connection test failed with the following error:DXServerException: Exception : org/apache/hadoop/fs/FileSystem
.
The definition of the classpath of the
Data Integration Hub
Hadoop Service is incorrect.
Perform the following actions:
  1. On the machine where the
    Data Integration Hub
    Hadoop Service is installed, open the
    dih-hadoop-service.xml
    file in a text editor from the following location:
    <
    DIH
    InstallationDir>/DataIntegrationHub/tomcat/conf/Catalina/localhost
  2. Configure the correct locations of the
    JAR
    files.
When I test the DIH__STAGING__HADOOP connection the test fails with the following error:
Data Access Connection test failed with the following error:DXServerException: Cannot establish connection to Apache Hive
.
Perform the following actions:
  1. Check the
    Data Integration Hub
    Hadoop Service log file for additional details. For more information, see Default Log Files.
  2. Verify that Apache Hive is running.
  3. On the machine where the
    Data Integration Hub
    Hadoop Service is installed, open the
    dx-configuration.properties
    file in a text editor from the following location:
    <
    DIH
    InstallationDir>/DataIntegrationHub/tomcat/shared/classes/
    Verify that the value of
    dih.hadoop.service.hive.url
    is correct. For more information, see Configuring the Environment for a Hadoop Publication Repository.
  4. If the message
    java.lang.ClassNotFoundException: org.apache.hive.jdbc.HiveDriver
    appears in the log file, or if a similar message appears in the file, this is an indication that the definition of the classpath of the
    Data Integration Hub
    Hadoop Service is incorrect. Perform the following actions:
    1. On the machine where the
      Data Integration Hub
      Hadoop Service is installed, open the
      dih-hadoop-service.xml
      file in a text editor from the following location:
      <
      DIH
      InstallationDir>/DataIntegrationHub/tomcat/conf/Catalina/localhost
    2. Configure the classpath correctly. For more information, see Configuring the Environment for a Hadoop Publication Repository.
When I test the DIH__STAGING__HADOOP connection the test fails with the following error:
Data Access Connection test failed with the following error:DXServerException: The Hadoop file system is not available
.
Perform the following actions:
  1. Verify that Apache Hadoop is running.
  2. On the machine where the
    Data Integration Hub
    Hadoop Service is installed, open the
    dih-hadoop-service.xml
    file in a text editor from the following location:
    <
    DIH
    InstallationDir>/DataIntegrationHub/tomcat/conf/Catalina/localhost
  3. Verify that all settings in the file are correct, including the correct locations of the
    JAR
    files. For more information, see Configuring the Environment for a Hadoop Publication Repository.
When I test the DIH__STAGING__HADOOP connection the test fails with the following error:
Data Access Connection test failed with the following error:DXServerException: ResourceAccessException: The connection is not valid. The Data Integration Hub Hadoop service is not running
.
The
Data Integration Hub
Hadoop Service is not running.
Log in to the machine where the
Data Integration Hub
Hadoop Service is installed and run the service. For more information, see Starting and Stopping Data Integration Hub on Linux.
Publication or subscription events stay in Processing status indefinitely
Verify that the configuration of the
Data Integration Hub
Hadoop Service is correct.
Perform the following actions:
  1. On the machine where the
    Data Integration Hub
    server is installed, open the
    dx-configuration.properties
    file in a text editor from the following location:
    <
    DIH
    InstallationDir>/DataIntegrationHub/tomcat/shared/classes/
    Verify that the value of
    dx.server.url
    is correct.
  2. If you installed the
    Data Integration Hub
    Hadoop Service on a different machine than the machine where you installed
    Data Integration Hub
    , open the
    dx-configuration.properties
    file in a text editor from the same location on the machine where the
    Data Integration Hub
    Hadoop Service is installed and verify that the value of
    dx.server.url
    is correct.
  3. On the machine where the
    Data Integration Hub
    Hadoop Service is installed, ping the URL that is defined in
    dx.server.url
    and verify that it is accessible.
The running of publications and subscriptions fails. The
Data Integration Hub
Hadoop Service log shows that the service repeatedly tries to access
localhost:8020
and fails each time.
In Cloudera Manager, enable the option
Bind NameNode to Wildcard Address
and then restart the HDFS service.

Back to Top

0 COMMENTS

We’d like to hear from you!