Dropbox Extract

Dropbox Extract



Dropbox Extract

The Dropbox Extract component uses the Dropbox API to retrieve and store data to be referenced by an External Table. Users can then transform their data with Matillion ETL's library of transformation components.

The Dropbox Extract component uses the Dropbox API to retrieve and store data to be loaded into a table. Users can then transform their data with Matillion ETL's library of transformation components.


Important Information

Using this component on Matillion ETL for Redshift may return structured data that requires flattening. For help flattening such data, please read our Nested Data Load Component documentation.

Using this component on Matillion ETL for Snowflake may return structured data that requires flattening. For help flattening such data, please read our Extract Nested Data Component documentation.

Using this component on Matillion ETL for BigQuery may return structured data that requires flattening. For help flattening such data, please read our Extract Nested Data Component documentation.



Redshift Properties

Property Setting Description
Name String Input the descriptive name of the component.
Data Source Select Select a Dropbox data source.
OAuth Select Select the OAuth entry. Oauth entries should be set up in advance and can be created by first clicking the Manage button. For help configuring an entry, read our Dropbox Authentication Guide.
Path String Provide the path for the file(s) you wish to retrieve.
Note: This property is only available when the Data Source property is set to either "Files" or "Shared Links".
Recursive Boolean When set to true, Matillion ETL will recursively retrieve the contents of Dropbox folders in the defined parent folder. The default setting is true.
Note: This property is only available when the Data Source property is set to "Files".
Folder ID String Provide the ID of a specific shared folder.
Note: This property is only available when the Data Source property is set to "Shared Members".
Account ID String Provide your Account ID. This is a user's account identifier used by the Dropbox API v2.
Note: This property is only available when the Data Source property is set to "Users".
Page Limit Integer Set the page limit for the amount of records to be returned and staged.
Location Storage Location Provide an S3 bucket path that will be used to store the data. Once on an S3 bucket, the data can be referenced by an external table. A folder will be created at this location with the same name as the target table.
External Schema Select Select the table's external schema. To learn more about external schemas, please read our support documentation on Getting Started With Amazon Redshift Spectrum.
For more information on using multiple schemas, see Schema Support.
Target Table String Provide a name for the external table to be used.
Warning: Upon running the job, this table will be recreated and will drop any existing table of the same name.

Snowflake Properties

Property Setting Description
Name String Input the descriptive name of the component.
Data Source Select Select a Dropbox data source.
OAuth Select Select the OAuth entry. Oauth entries should be set up in advance and can be created by first clicking the Manage button. For help configuring an entry, read our Dropbox Authentication Guide.
Path String Provide the path for the file(s) you wish to retrieve.
Note: This property is only available when the Data Source property is set to either "Files" or "Shared Links".
Recursive Boolean When set to true, Matillion ETL will recursively retrieve the contents of Dropbox folders in the defined parent folder. The default setting is true.
Note: This property is only available when the Data Source property is set to "Files".
Folder ID String Provide the ID of a specific shared folder.
Note: This property is only available when the Data Source property is set to "Shared Members".
Account ID String Provide your Account ID. This is a user's account identifier used by the Dropbox API v2.
Note: This property is only available when the Data Source property is set to "Users".
Page Limit Integer Set the page limit for the amount of records to be returned and staged.
Location Storage Location Provide an S3 bucket path, GCS bucket path, or Azure Blob Storage path that will be used to store the data. Once on an S3 bucket, GCS bucket or Azure Blob, the data can be referenced by an external table. A folder will be created at this location with the same name as the Target Table.
Integration Select (GCP only) Choose your Google Cloud Storage Integration. Integrations are required to permit Snowflake to read data from and write to a Google Cloud Storage bucket. Integrations must be set up in advance of selecting them in Matillion ETL. To learn more about setting up a storage integration, read our Storage Integration Setup Guide.
Warehouse Select Choose a Snowflake warehouse that will run the load.
Database Select Choose a database to create the new table in.
Schema Select Select the table schema. For more information on using multiple schemas, please refer to this article.
Target Table String Provide a new table name.
Warning: Upon running the job, this table will be recreated and will drop any existing table of the same name.

BigQuery Properties

Property Setting Description
Name String Input the descriptive name of the component.
Data Source Select Select a Dropbox data source.
OAuth Select Select the OAuth entry. Oauth entries should be set up in advance and can be created by first clicking the Manage button. For help configuring an entry, read our Dropbox Authentication Guide.
Path String Provide the path for the file(s) you wish to retrieve.
Note: This property is only available when the Data Source property is set to either "Files" or "Shared Links".
Recursive Boolean When set to true, Matillion ETL will recursively retrieve the contents of Dropbox folders in the defined parent folder. The default setting is true.
Note: This property is only available when the Data Source property is set to "Files".
Folder ID String Provide the ID of a specific shared folder.
Note: This property is only available when the Data Source property is set to "Shared Members".
Account ID String Provide your Account ID. This is a user's account identifier used by the Dropbox API v2.
Note: This property is only available when the Data Source property is set to "Users".
Page Limit Integer Set the page limit for the amount of records to be returned and staged.
Project Select The target BigQuery project to load data into.
Dataset Select The target BigQuery dataset to load data into.
Target Table String Provide a new table name.
Warning: Upon running the job, this table will be recreated and will drop any existing table of the same name.
Cloud Storage Staging Area String | File Structure Specify the target Google Cloud Storage bucket to be used for staging the queried data. Users can either:
  1. Input the URL string of the Cloud Storage bucket following the template provided: gs://<bucket>/<path>
  2. Navigate through the file structure to select the target bucket.
Load Options Multiple Select Clean Cloud Storage Files: Destroy staged files on Cloud Storage after loading data. Default is On.
Cloud Storage File Prefix: Give staged file names a prefix of your choice. The default setting is an empty field.
Recreate Target Table: Choose whether the component recreates its target table before the data load. If Off, the existing table will be used. Default is On.
Use Grid Variable: Check this checkbox to use a grid variable. Default is unchecked.