This article uses the following method to install Python on each Data Integration Service machine:
Install Python in a directory that we will set to an environment variable called
$PYTHONHOME
.
Install Python third-party libraries in the same location.
Copy the contents under
$PYTHONHOME
to
$INFA_HOME/services/shared/spark/python/
where
$INFA_HOME
is the location of the Informatica domain.
Repeat steps 1 to 3 for each Data Integration Service machine.
The final location
$INFA_HOME/services/shared/spark/python/
is the location that the Data Integration Service uses to push the Python installation to the Hadoop cluster nodes where the Spark engine runs the Python transformation.