Table of Contents

Search

  1. Preface
  2. Introduction to PowerExchange for Amazon S3
  3. PowerExchange for Amazon S3 Configuration Overview
  4. Amazon S3 Connections
  5. PowerExchange for Amazon S3 Data Objects
  6. PowerExchange for Amazon S3 Mappings
  7. Appendix A: Amazon S3 Data Type Reference
  8. Appendix B: Troubleshooting

PowerExchange for Amazon S3 User Guide

PowerExchange for Amazon S3 User Guide

Prerequisites

Prerequisites

Before you can use PowerExchange for Amazon S3, perform the following tasks:
  • Ensure that PowerExchange for Amazon S3 license is activated.
  • Create an Access Key ID and Secret Access Key in AWS. You can provide these key values when you create an Amazon S3 connection
  • Ensure that you have the
    sts:AssumeRole
    permission and a trust relationship established within the AWS accounts to use the temporary security credentials to access the AWS resources.
  • Create the temporary security credentials policy to use the temporary security credentials to access the AWS resources.
  • Verify that you have write permissions on all the directories within the
    <INFA_HOME>
    directory.
  • To run mappings on Hortonworks and Amazon EMR distributions that use non-Kerberos authentication, configure user impersonation.
    For information about configuring user impersonation, see the
    Data Engineering Integration Guide
    .
  • To run mappings on MapR secure clusters, configure the MapR secure clusters on all the nodes.
    For information about configuring MapR secure clusters, see the
    Data Engineering Integration Guide
    .
  • To run mappings on the Spark engine and in EMR 5.20 or later distributions that use server-side encryption with KMS, enable the KMS policy for the Amazon S3 bucket.
  • To successfully preview data from complex files, you must configure the INFA_PARSER_HOME property for the Data Integration Service in Informatica Administrator. Perform the following steps to configure the INFA_PARSER_HOME property:
    • Log in to Informatica Administrator.
    • Click the Data Integration Service and then click the
      Processes
      tab on the right pane.
    • Click
      Edit
      in the
      Environment Variables
      section.
    • Click
      New
      to add an environment variable.
    • Enter the name of the environment variable as
      INFA_PARSER_HOME
      .
    • Set the value of the environment variable to the absolute path of the Hadoop distribution directory on the machine that runs the Data Integration Service. Verify that the version of the Hadoop distribution directory that you define in the INFA_PARSER_HOME property is the same as the version you defined in the cluster configuration.
  • To run mappings on Spark or Databricks Spark engine or to test connections through Administration tool using a secure domain, you must perform the following steps:
    • Download the
      Baltimore CyberTrust Root
      certificate file.
    • Provide the read, write, and execute permissions to the certificate file.
    • Run the following command to import the certificate file into the Informatica TrustStore location:
      <INFA_HOME>/java/jre/bin/keytool -keystore <infa_trust_store_location> -importcert -alias <Alias_Name> -file <BaltimoreCyberTrustRoot certificate file path>/<certificate_filename> -storepass <Truststore_Password>
    • Restart the Data Integration Service.

0 COMMENTS

We’d like to hear from you!