Skip to main content
This guide walks you through the end-to-end process of creating a Databricks source in the Integrate module, from selecting the source category to verifying a successful integration. For detailed information about the underlying components and prerequisites, refer to Create and Implement a Databricks Source.

Step 1: Open the Sources Tab and Click Create Source

Navigate to Integrate → Sources in the Zeotap CDP App. Click Create Source in the top-right corner of the Sources listing page.

Step 2: Select Data Warehouse as the Category

In the Create Source dialog, you will be presented with a list of source categories. Select Data Warehouse to proceed with a Databricks source.
On the next screen, choose Databricks as the data source type.

Step 3: Fill in Source Details

The Source Details step collects the configuration needed to connect to your Databricks workspace. Fill in the following fields:
  • Source Name — Enter a short, descriptive name for the source.
  • Sequence — Specify the sequence number for data ordering.
  • Sync Frequency — Set how often data should be synchronised (for example, hourly, daily, or weekly).
  • Host — Enter your Databricks workspace URL.
  • Catalog Name — Provide the catalog that contains your target data.
  • Schema Name — Specify the schema within the catalog.
  • Table Name — Enter the name of the table to ingest.
  • Data Entity — Choose whether you are ingesting Customer Data or Non Customer Data.
  • Delta Queries Selection — Set to true if you want to fetch only new and updated records based on a timestamp column. Set to false to fetch all records on every run.

Step 4: Choose the Connection Type and Authentication

Under Type, select the mechanism used to pull data from Databricks:
  • JDBC — The standard connection method.
  • Job Based — The recommended option for large data volumes.
NoteIf you are ingesting more than 1 million records, use the Job Based approach. JDBC may encounter performance issues with large data volumes.
After selecting the Type, choose the Auth Type from the dropdown. The available options differ depending on the connection type selected.

Step 5: Configure JDBC Authentication

If you selected JDBC, enter the following credentials:
FieldDescription
Client IDThe OAuth client ID from your Databricks service principal.
Client SecretThe OAuth client secret associated with your service principal.
HTTP PathThe HTTP path of your Databricks SQL warehouse. Found under SQL Warehouses → Connection details.
Partition ColumnOptional. A column used to split large datasets for parallel reads. Use a unique column if no partition column exists.
For guidance on obtaining the Client ID, Client Secret, and HTTP Path, refer to the Prerequisites for JDBC section.

Step 6: Configure Job-Based Authentication

If you selected Job Based, choose the Auth Type and enter the corresponding credentials: PACKTOKEN
FieldDescription
Pack TokenA personal access token generated from your Databricks account. Navigate to Settings → User Settings → Developer → Generate new token.
Service Principal
FieldDescription
Client IDThe client ID of the Databricks service principal.
Client SecretThe client secret associated with the service principal.
Security noticeKeep all credentials — Client ID, Client Secret, and Pack Token — secure. Do not share them publicly or commit them to version control.

Step 7: Finalise and Verify the Source

After entering all required fields, click Next to proceed to column selection. Select the columns you want to ingest, then click Create Source.
Once created, the source appears in the Sources listing page with a status of Integrated.
NoteThe initial data transfer from Databricks to Zeotap CDP may take time depending on data volume. For assistance with Databricks source setup, contact the Zeotap support team at [email protected].

Create and Implement a Databricks Source

Create a Snowflake Source

Create a BigQuery Source

Last modified on April 8, 2026