Table of Contents

Search

  1. Preface
  2. Introduction to Informatica Data Engineering Integration
  3. Mappings
  4. Mapping Optimization
  5. Sources
  6. Targets
  7. Transformations
  8. Python Transformation
  9. Data Preview
  10. Cluster Workflows
  11. Profiles
  12. Monitoring
  13. Hierarchical Data Processing
  14. Hierarchical Data Processing Configuration
  15. Hierarchical Data Processing with Schema Changes
  16. Intelligent Structure Models
  17. Blockchain
  18. Stateful Computing
  19. Appendix A: Connections Reference
  20. Appendix B: Data Type Reference
  21. Appendix C: Function Reference

Message Targets on Hadoop

Message Targets on Hadoop

You can write to Kafka messaging targets.
Apache Kafka is a publish-subscribe messaging system. Kafka runs as a cluster comprised of one or more servers each of which is called a broker. Kafka brokers process data in the form of messages, publishes the messages to a topic, subscribes the messages from a topic, and then writes it to the Kafka target.
You can use the Kafka data object write operation as a target. If you want to configure high availability for the mapping, ensure that the Kafka cluster is highly available.
For more information about Kafka clusters, Kafka brokers, and partitions, see https://kafka.apache.org/documentation.
Support to write the batch data to a Kafka target in Data Engineering Integration is available for technical preview.
Technical preview functionality is supported for evaluation purposes but is unwarranted and is not production-ready. Informatica recommends that you use in non-production environments only. Informatica intends to include the preview functionality in an upcoming release for production use, but might choose not to in accordance with changing market or technical circumstances. For more information, contact Informatica Global Customer Support.


Updated September 28, 2020