Table of Contents

Search

  1. Abstract
  2. PowerExchange for Amazon Redshift
  3. PowerExchange for Amazon S3
  4. PowerExchange for Cassandra
  5. PowerExchange for Cassandra JDBC
  6. PowerExchange for DataSift
  7. PowerExchange for Facebook
  8. PowerExchange for Google Analytics
  9. PowerExchange for Google BigQuery
  10. PowerExchange for Google Cloud Spanner
  11. PowerExchange for Google Cloud Storage
  12. PowerExchange for Greenplum
  13. PowerExchange for HBase
  14. PowerExchange for HDFS
  15. PowerExchange for Hive
  16. PowerExchange for JD Edwards EnterpriseOne
  17. PowerExchange for JDBC V2
  18. PowerExchange for LDAP
  19. PowerExchange for LinkedIn
  20. PowerExchange for MapR-DB
  21. PowerExchange for Microsoft Azure Blob Storage
  22. PowerExchange for Microsoft Azure Cosmos DB SQL API
  23. PowerExchange for Microsoft Azure Data Lake Storage Gen1
  24. PowerExchange for Microsoft Azure Data Lake Storage Gen2
  25. PowerExchange for Microsoft Azure SQL Data Warehouse
  26. PowerExchange for Microsoft Dynamics CRM
  27. PowerExchange for MongoDB
  28. PowerExchange for Netezza
  29. PowerExchange for OData
  30. PowerExchange for Salesforce
  31. PowerExchange for Salesforce Marketing Cloud
  32. PowerExchange for SAP NetWeaver
  33. PowerExchange for Snowflake
  34. PowerExchange for Tableau
  35. PowerExchange for Tableau V3
  36. PowerExchange for Teradata Parallel Transporter API
  37. PowerExchange for Twitter
  38. PowerExchange for Web Content-Kapow Katalyst
  39. Informatica Global Customer Support

PowerExchange Adapters for Informatica Release Notes (10.4.1.2)

PowerExchange Adapters for Informatica Release Notes (10.4.1.2)

PowerExchange for HDFS Known Issues (10.4.1)

PowerExchange for HDFS Known Issues (10.4.1)

The following table describes known issues:
Bug
Description
OCON-25683
When you set the
-DINFA_HADOOP_DIST_DIR=hadoop\<Distro>
option in the
developerCore.ini
file and import a Parquet file , the format of the imported metadata differs based on the distribution. For Cloudera CDP 7.1, the metadata is imported as string and for other supported distributions, the metadata is imported as UTF8.
OCON-25655
When you run a mapping in the native environment to write data to a complex file target and choose to overwrite the target data and the target filename does not contain the file format extension such as ".avro or .parquet", the Data Integration Service does not delete the target data before writing data.
OCON-25649
When you run a mapping on the Blaze engine to read data from and write data to sequence complex file data objects using a Kerberos connection, the mapping fails with the following exception:
"GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos target)"
OCON-24348
When you run a mapping on the Spark engine to read data from an empty JSON complex file source and write data to a complex file target, the mapping should fail, but the mapping runs successfully and the Data Integration Service generates an empty target file.