Table of Contents

Search

  1. Preface
  2. Introduction to PowerExchange for Teradata Parallel Transporter API
  3. PowerExchange for Teradata Parallel Transporter API Configuration
  4. PowerExchange for Teradata Parallel Transporter Connections
  5. PowerExchange for Teradata Parallel Transporter API Data Objects
  6. PowerExchange for Teradata Parallel Transporter API Mappings
  7. Teradata Parallel Transporter API Run-time Processing
  8. Data Type Reference

PowerExchange for Teradata Parallel Transporter API User Guide

PowerExchange for Teradata Parallel Transporter API User Guide

PowerExchange for Teradata Parallel Transporter API Overview

PowerExchange for Teradata Parallel Transporter API Overview

You can use PowerExchange for Teradata Parallel Transporter API to read data from or write data to a Teradata database. You can run Teradata mappings in the native or Hadoop environment. If you choose the Hadoop environment, you can run Teradata mappings on the Blaze or Hive engines.
PowerExchange for Teradata Parallel Transporter API uses different integration methods based on the environment in which you run the mappings.

Integration with Teradata Parallel Transporter API

When you run Teradata mappings in the native environment or on the Hive engine to read or load data in bulk, PowerExchange for Teradata Parallel Transporter API integrates with Teradata Parallel Transporter API (Teradata PT API). The Data Integration Service uses the Teradata PT API infrastructure to connect to Teradata.
The Data Integration Service uses the Export system operator to read data. The Export operator exports large amounts of data from Teradata tables or views.
The Data Integration Service uses one of the following Teradata PT API system operators to load data in bulk:
Load
Loads data in bulk into an empty Teradata table.
Stream
Performs insert, update, upsert, and delete operations against Teradata tables in near real time.
Update
Performs insert, update, upsert, and delete operations against Teradata tables.

Integration with Sqoop and Hortonworks Connector for Teradata

When you run Teradata mappings on a Hortonworks cluster and on the Blaze engine, the Data Integration Service invokes the Hortonworks Connector for Teradata at run time. The Data Integration Service then runs the mapping through Sqoop.

Integration with Sqoop and Cloudera Connector Powered by Teradata

When you run Teradata mappings on a Cloudera cluster and on the Blaze engine, the Data Integration Service invokes the Cloudera Connector Powered by Teradata at run time. The Data Integration Service then runs the mapping through Sqoop.

Integration with Teradata Connector for Hadoop

When you run Teradata mappings on the Hive engine, you can enable Teradata Connector for Hadoop (TDCH) to run the mapping and increase performance. TDCH is a set of API and tools that Teradata Corporation provides for parallel processing of data between Teradata databases and the Hadoop ecosystem of products.

Example

In a native environment, you can store payroll information of an organization for the past five years. You can use PowerExchange for Teradata Parallel Transporter API in the native environment to read payroll information and then write the data to Teradata tables for storage.
In a Hadoop environment, you can gather, store, and analyze large volumes of unstructured data such as web logs. You can process these large amounts of data in a Hadoop environment and use PowerExchange for Teradata Parallel Transporter API to write meaningful data to Teradata tables for analysis.


Updated October 05, 2018