Table of Contents

Search

  1. Preface
  2. Introduction to PowerExchange for Amazon S3
  3. PowerExchange for Amazon S3 Configuration Overview
  4. Amazon S3 Connections
  5. PowerExchange for Amazon S3 Data Objects
  6. PowerExchange for Amazon S3 Mappings
  7. PowerExchange for Amazon S3 Lookups
  8. Appendix A: Amazon S3 Data Type Reference
  9. Appendix B: Troubleshooting

PowerExchange for Amazon S3 User Guide

PowerExchange for Amazon S3 User Guide

Prerequisites

Prerequisites

Before you can use PowerExchange for Amazon S3, perform the following tasks:
  • Ensure that PowerExchange for Amazon S3 license is activated.
  • Create an Access Key ID and Secret Access Key in AWS. You can provide these key values when you create an Amazon S3 connection
  • Ensure that you have the
    sts:AssumeRole
    permission and a trust relationship established within the AWS accounts to use the temporary security credentials to access the AWS resources.
  • Create the temporary security credentials policy to use the temporary security credentials to access the AWS resources.
  • Verify that you have write permissions on all the directories within the
    <INFA_HOME>
    directory.
  • To run mappings on Amazon EMR distributions that use non-Kerberos authentication, configure user impersonation.
    For information about configuring user impersonation, see the
    Data Engineering Integration Guide
    .
  • To run mappings on MapR secure clusters, configure the MapR secure clusters on all the nodes.
    For information about configuring MapR secure clusters, see the
    Data Engineering Integration Guide
    .
  • To run mappings on the Spark engine and in Amazon EMR distributions that use server-side encryption with KMS, enable the KMS policy for the Amazon S3 bucket.
  • Verify that the Hadoop Distribution Directory property in the developerCore.ini file is set based on the Hadoop distribution that you use.
  • To run a mapping to process complex files, you must configure the INFA_PARSER_HOME environment variable for the Data Integration Service in Informatica Administrator. Set the value of the environment variable to the absolute path of the Hadoop distribution directory on the machine that runs the Data Integration Service.
  • To run mappings on Spark or Databricks Spark engine or to test connections through Administration tool using a secure domain, you must perform the following steps:
    • Download the
      Baltimore CyberTrust Root
      certificate file.
    • Provide the read, write, and execute permissions to the certificate file.
    • Run the following command to import the certificate file into the Informatica TrustStore location:
      <INFA_HOME>/java/jre/bin/keytool -keystore <infa_trust_store_location> -importcert -alias <Alias_Name> -file <BaltimoreCyberTrustRoot certificate file path>/<certificate_filename> -storepass <Truststore_Password>
    • Restart the Data Integration Service.

0 COMMENTS

We’d like to hear from you!