Table of Contents


  1. Preface
  2. Part 1: Hadoop Integration
  3. Part 2: Databricks Integration
  4. Appendix A: Connections

Configure the Files on S3 on the Spark Engine

Configure the Files on S3 on the Spark Engine

To run mappings on S3 on the Spark engine, you need to configure the files from the master node to the Data Integration Service machine.
Perform this task in the following situations:
  • You are integrating for the first time.
  • You upgraded from any Informatica version and changed the distribution version.
You can perform the following steps to configure the files to integrate with EMR 5.20:
Copy the .jar file
To integrate with EMR 5.20, get
from the Hadoop administrator. Copy
the file to the following locations on each Data Integration Service machine:
/<Informatica installation directory>/services/shared/hadoop/EMR_5.16/lib
Create a file
Create a
on the Data Integration Service machine. The file must contain the AWS location.
For example,
[default] region=us-west-2
Create an environment variable
Create the
environment variable on the Data Integration Service machine. Set the value to
Copy and replace a file
Copy hadoop-common-2.8.5-amzn-1.jar from the following location in the EMR 5.20 cluster:
Replace the file in the following location:
<Informatica installation directory>/services/shared/hadoop/EMR_5.16/lib
Recycle the Data Integration Service
You must recycle the Data Integration Service to reflect the changes.


We’d like to hear from you!