Table of Contents

Search

  1. Preface
  2. Introduction to PowerExchange for Amazon S3
  3. PowerExchange for Amazon S3 Configuration Overview
  4. Amazon S3 Connections
  5. PowerExchange for Amazon S3 Data Objects
  6. PowerExchange for Amazon S3 Mappings
  7. PowerExchange for Amazon S3 Lookups
  8. Appendix A: Amazon S3 Data Type Reference
  9. Appendix B: Troubleshooting

PowerExchange for Amazon S3 User Guide

PowerExchange for Amazon S3 User Guide

Directory-Level Partitioning

Directory-Level Partitioning

When you run mappings on the Spark and Databricks Spark engines, you can read data from and write data to Avro, ORC, and Parquet files that are partitioned based on directories.

Importing a data object with partition files

Perform the following steps to import a data object to read or write from partition files:
  1. Select a project or folder in the
    Object Explorer
    view.
  2. Click
    File
    New
    Data Object
    .
  3. Select
    AmazonS3 Data Object
    and click
    Next
    .
    The
    AmazonS3 Data Object
    dialog box appears.
  4. Click
    Browse
    next to the
    Location
    option and select the target project or folder.
  5. In the
    Resource Format
    list, select Avro, Parquet, or ORC from the drop-down.
  6. Click
    Add
    next to the
    Selected Resource
    option to add a resource to the data object. The
    Add Resource
    dialog box appears. You can use the
    File Type
    column to distinguish between a directory and a file.
    The following image shows the Add resource dialogue box where you can select the file name and directory:
    The File Type tab showing files and directories.
  7. Select the check box for a directory. Click
    OK
    .
  8. Click
    Finish
    .
    The partitioned columns are displayed with the order of partitioning in the data object
    Overview
    tab.
    The following image shows the data object overview tab:
    The partition order tab shows the order of partitions.

Create target with partition files

Perform the following steps to create target with partition files:
  1. Select a project or folder in the
    Object Explorer
    view.
  2. Select a source or a transformation in the mapping.
  3. Right-click the Source transformation and select
    Create Target
    .
    The
    Create Target
    dialog box appears.
    The following image shows the
    Create Target
    option:
    You can view the option.Create Target
  4. Select
    Others
    and then select
    AmazonS3
    data object from the list in the
    Data Object Type
    section.
  5. Click
    OK
    .
    The
    New AmazonS3 Data Object
    dialog box appears.
    The following image shows the
    New AmazonS3 Data Object
    dialog box:
    The New AmazonS3 Data Object dialog box.
  6. Enter a name for the data object.
  7. Enter the partition fields.
    The following image shows the
    Edit partition fields
    dialog box:
    You can edit the partition firlds here.
  8. You can change the partition order using the up and down arrows.
    The following image shows the partitioned fields after changing the order:
    changed order of the partition fields.
  9. Click
    Finish
    .
    The partitioned columns are displayed with the order of partitioning in the data object
    Overview
    tab.

0 COMMENTS

We’d like to hear from you!