Network
Data Engineering
Data Engineering Integration
Enterprise Data Catalog
Enterprise Data Preparation
Cloud Integration
Cloud Application Integration
Cloud Data Integration
Cloud Customer 360
DiscoveryIQ
Cloud Data Wizard
Informatica for AWS
Informatica for Microsoft
Cloud Integration Hub
Complex Event Processing
Proactive Healthcare Decision Management
Proactive Monitoring
Real-Time Alert Manager
Rule Point
Data Integration
B2B Data Exchange
B2B Data Transformation
Data Integration Hub
Data Replication
Data Services
Data Validation Option
Fast Clone
Informatica Platform
Metadata Manager
PowerCenter
PowerCenter Express
PowerExchange
PowerExchange Adapters
Data Quality
Axon Data Governance
Data as a Service
Data Explorer
Data Quality
Data Security Group (Formerly ILM)
Data Archive
Data Centric Security
Secure@Source
Secure Testing
Master Data Management
Identity Resolution
MDM - Relate 360
Multidomain MDM
MDM Registry Edition
Process Automation
ActiveVOS
Process Automation
Product Information Management
Informatica Procurement
MDM - Product 360
Ultra Messaging
Ultra Messaging Options
Ultra Messaging Persistence Edition
Ultra Messaging Queuing Edition
Ultra Messaging Streaming Edition
Edge Data Streaming
Knowledge Base
Resources
PAM (Product Availability Matrices)
Support TV
Velocity (Best Practices)
Mapping Templates
Debugging Tools
User Groups
Documentation
English
English
English
Español
Spanish
Deutsch
German
Français
French
日本語
Japanese
한국어
Korean
Português
Portuguese
中文
Chinese
Log Out
Log In
Sign Up
Data Engineering Integration
10.0
H2L
10.4.1
10.4.0
10.2.2 HotFix 1
10.2.2 Service Pack 1
10.2.2
10.2.1
10.2 HotFix 2
10.2 HotFix 1
10.2
10.1.1 Update 2
10.1.1 HotFix 1
10.1.1
10.1
10.0 Update 1
10.0
User Guide
Data Engineering Integration
All Products
Table of Contents
Search
No Results
Preface
Introduction to Informatica Big Data Management
Informatica Big Data Management Overview
Example
Big Data Management Tasks
Read from and Write to Big Data Sources and Targets
Perform Data Discovery
Perform Data Lineage on Big Data Sources
Stream Machine Data
Manage Big Data Relationships
Big Data Process
Step 1. Collect the Data
Step 2. Cleanse the Data
Step 3. Transform the Data
Step 4. Process the Data
Step 5. Monitor Jobs
Big Data Management Component Architecture
Clients and Tools
Application Services
Repositories
Third-Party Applications
Big Data Management Connectivity Architecture
Hadoop Ecosystem Architecture
Connections
Connections Overview
Hadoop Connection Properties
HDFS Connection Properties
HBase Connection Properties
Hive Connection Properties
Creating a Connection to Access Sources or Targets
Creating a Hadoop Connection
Mappings in a Hadoop Environment
Mappings in a Hadoop Environment Overview
Data Warehouse Optimization Mapping Example
Hive Engine Architecture
Informatica Blaze Engine Architecture
High-Level Steps to Run a Mapping in the Hadoop Environment
Sources in a Hadoop Environment
Flat File Sources
Hive Sources
Relational Sources
Targets in a Hadoop Environment
Flat File Targets
HDFS Flat File Targets
Hive Targets
Relational Targets
Transformations in a Hadoop Environment
Variable Ports in a Hadoop Environment
Functions in a Hadoop Environment
Mappings in a Hadoop Environment
Data Types in a Hadoop Environment
Parameters in a Hadoop Environment
Parameter Usage
Create and Use Hadoop Parameters
Workflows that Run Mappings in a Hadoop Environment
Configuring a Mapping to Run in a Hadoop Environment
Mapping Execution Plans
Hive Engine Execution Plan Details
Blaze Engine Execution Plan Details
Viewing the Execution Plan for a Mapping in the Developer Tool
Monitor Jobs
Accessing the Monitoring URL
Monitor Blaze Engine Jobs
Monitoring a Mapping
Hadoop Environment Logs
Blaze Engine Logs
Hive Engine Logs
Viewing Hadoop Environment Logs in the Administrator Tool
Viewing Logs in the Blaze Job Monitor
Optimization for the Hadoop Environment
Truncating Partitions in a Hive Target
Enabling Data Compression on Temporary Staging Tables
Step 1. Configure the Hive Connection to Enable Data Compression on Temporary Staging Tables
Step 2. Configure the Hadoop Cluster to Enable Compression on Temporary Staging Tables
Parallel Sorting
Troubleshooting a Mapping in a Hadoop Environment
Mappings in the Native Environment
Mappings in the Native Environment Overview
Data Processor Mappings
HDFS Mappings
HDFS Data Extraction Mapping Example
Hive Mappings
Hive Mapping Example
Social Media Mappings
Twitter Mapping Example
Profiles
Profiles Overview
Native and Hadoop Environments
Run-time Environment and Profile Performance
Profile Types on Hadoop
Column Profiles on Hadoop
Rule Profiles on Hadoop
Data Domain Discovery on Hadoop
Running a Profile on Hadoop in the Developer Tool
Running a Profile on Hadoop in the Analyst Tool
Running Multiple Data Object Profiles on Hadoop
Monitoring a Profile
Troubleshooting
Native Environment Optimization
Native Environment Optimization Overview
Processing Big Data on a Grid
Data Integration Service Grid
Grid Optimization
Processing Big Data on Partitions
Partitioned Model Repository Mappings
Partition Optimization
High Availability
POWERCENTERHELP
POWERCENTERHELP.CHM
Big Data Edition User Guide
Data Type Reference
Data Type Reference Overview
Hive Complex Data Types
Hive Data Types and Transformation Data Types
User Guide
User Guide
10.0
10.4.1
10.4.0
10.2.2 HotFix 1
10.2.2 Service Pack 1
10.2.2
10.2.1
10.2 HotFix 1
10.2
10.1.1 Update 2
10.1.1 HotFix 1
10.1.1
10.1
Back
Next
Native Environment Optimization
Native Environment Optimization
Native Environment Optimization Overview
Processing Big Data on a Grid
Processing Big Data on Partitions
High Availability
Updated July 03, 2018
Download Guide
Send Feedback
Explore Informatica Network
Communities
Knowledge Base
Success Portal
Back to Top
Back
Next