Table of Contents

Search

  1. Abstract
  2. PowerExchange for Amazon Redshift
  3. PowerExchange for Amazon S3
  4. PowerExchange for Cassandra
  5. PowerExchange for Google Analytics
  6. PowerExchange for Google BigQuery
  7. PowerExchange for Google Cloud Spanner
  8. PowerExchange for Google Cloud Storage
  9. PowerExchange for Greenplum
  10. PowerExchange for HBase
  11. PowerExchange for HDFS
  12. PowerExchange for Hive
  13. PowerExchange for JDBC V2
  14. PowerExchange for JD Edwards EnterpriseOne
  15. PowerExchange for Kudu
  16. PowerExchange for LDAP
  17. PowerExchange for Microsoft Azure Blob Storage
  18. PowerExchange for Microsoft Azure Cosmos DB SQL API
  19. PowerExchange for Microsoft Azure Data Lake Storage Gen1
  20. PowerExchange for Microsoft Azure Data Lake Storage Gen2
  21. PowerExchange for Microsoft Azure SQL Data Warehouse
  22. PowerExchange for Microsoft Dynamics CRM
  23. PowerExchange for MongoDB
  24. PowerExchange for Netezza
  25. PowerExchange for OData
  26. PowerExchange for Salesforce
  27. PowerExchange for SAP NetWeaver
  28. PowerExchange for Snowflake
  29. PowerExchange for Tableau
  30. PowerExchange for Tableau V3
  31. PowerExchange for Teradata Parallel Transporter API
  32. Informatica Global Customer Support

PowerExchange Adapters for Informatica Release Notes

PowerExchange Adapters for Informatica Release Notes

PowerExchange for Microsoft Azure Data Lake Storage Gen1 (10.5)

PowerExchange for Microsoft Azure Data Lake Storage Gen1 (10.5)

Known Issues

The following table describes known issues:
Bug
Description
OCON-27080
When you run a mapping to read or write complex files on the SUSE Linux Enterprise Server version 15, the mapping fails with the following error:
"Data Transport Error, Origin :[IProxyDTMService :: stopDTM RPC"
Workaround: Update the third-party libraries to
wildfly-openssl-java-1.0.10.Final.jar
in the distribution where you want to run the mapping.
OCON-20605
When you run a mapping, in the native environment, to read a flat file that has unicode characters, a space, null values, single quotes, or a value that starts with a dollar sign, the Data Integration Service adds double quotes to the values when writing data to the target.
OCON-17443
When you use the
Create Target
option to create a Microsoft Azure Blob Storage target and select Flat as the Resource Format, fields are not getting propagated to the target.
Workaround: Enable column projection and create fields manually in the target file and run the mapping.
OCON-10244
When you read data from or write data to Microsoft Azure Data Lake Store, the entire file gets downloaded in the staging directory even if you cancel the mapping.

0 COMMENTS

We’d like to hear from you!