Table of Contents


  1. Preface
  2. Part 1: Hadoop Integration
  3. Part 2: Databricks Integration
  4. Appendix A: Connections Reference

Create a Hive Staging Directory

Create a Hive Staging Directory

The Blaze and Spark engines require access to the Hive staging directory. You can use the default directory, or you can create a directory on HDFS. For example, if you create a directory, you might run the following command:
hadoop fs –mkdir /staging
If you use the default directory or create a directory, you must grant execute permission to the Hadoop impersonation user and the mapping impersonation users.


We’d like to hear from you!