Install Python for the Python transformation on Hadoop

Install Python for the Python transformation on Hadoop

Steps

Steps

This article uses the following method to install Python on each Data Integration Service machine:
  1. Install Python in a directory that we will set to an environment variable called
    $PYTHONHOME
    .
  2. Install Python third-party libraries in the same location.
  3. Copy the contents under
    $PYTHONHOME
    to
    $INFA_HOME/services/shared/spark/python/
    where
    $INFA_HOME
    is the location of the Informatica domain.
  4. Repeat steps 1 to 3 for each Data Integration Service machine.
The final location
$INFA_HOME/services/shared/spark/python/
is the location that the Data Integration Service uses to push the Python installation to the Hadoop cluster nodes where the Spark engine runs the Python transformation.

0 COMMENTS

We’d like to hear from you!