Table of Contents

Search

  1. Preface
  2. Part 1: Introduction to Data Discovery
  3. Part 2: Data Discovery with Informatica Analyst
  4. Part 3: Data Discovery with Informatica Developer
  5. Appendix A: Function Support Based on Profiling Warehouse Connection

Data Discovery Guide

Data Discovery Guide

Databricks Environment

Databricks Environment

You can choose the Databricks Spark option to run the profiles in the Databricks run-time environment.
After you choose the Databricks Spark option, you can select a Databricks connection. The Data Integration Service pushes the profile logic to the Spark engine on the Databricks cluster to run profiles.
When you run a profile in the Databricks environment, the Developer tool submits the profile jobs to the Profiling Service Module. The Profiling Service Module then breaks down the profile jobs into a set of mappings. The Data Integration Service pushes the mappings to the Spark engine through the Hadoop connection. The Spark engine processes the mappings and the Data Integration Service writes the profile results to the profiling warehouse.

0 COMMENTS

We’d like to hear from you!