Data Ingestion from Amazon S3 to Snowflake

Calibo Accelerate now supports ingestion of delta data into a Snowflake data lake from an Amazon S3 data source where the file format used is xlsx. An audit table is created on folder selection, during the job creation,

Let us create a pipeline with the following nodes:

Data Ingestion for delta data from S3 to Snowflake

To create a data integration job using Amazon S3 as source and Snowflake as target

  1. Configure the Amazon S3 nodes and select a folder with a .xlsx file.

  2. Configure the Snowflake node.

  3. Click the Databricks node in the data integration stage of the pipeline and click Create Templatized Job.

  4. Complete the following steps to create the job:

What happens after the first job run?

  • After the first job run target tables are created as per the number of sheets in the source Excel file and the source data is loaded into the respective tables.

  • For subsequent job runs, the delta data is loaded into the tables.

Related Topics Link IconRecommended Topics What's next?Databricks Templatized Data Integration Jobs