Skip to content

Databricks

Uploads data to Databricks as Parquet files via the Files API, with optional COPY INTO.

Configuration

yaml
warehouse:
  kind: databricks
  workspace_url: "https://my-workspace.cloud.databricks.com"
  token: ${DATABRICKS_TOKEN}
  warehouse_id: "abc123"
  catalog: main
  schema: default
FieldDefaultDescription
workspace_url(required)Databricks workspace URL
token(required)Personal access token
warehouse_idSQL warehouse ID (enables COPY INTO)
catalogmainUnity Catalog name
schemadefaultSchema name

CLI

bash
skippr connect warehouse databricks \
  --workspace-url https://dbc-xxxxxxxx.cloud.databricks.com \
  --token "$DATABRICKS_TOKEN" \
  --warehouse-id abc123 \
  --catalog main \
  --schema default
FlagDescription
--workspace-urlDatabricks workspace URL (or set DATABRICKS_HOST)
--tokenPersonal access token (or set DATABRICKS_TOKEN)
--warehouse-idSQL warehouse ID (or set DATABRICKS_WAREHOUSE_ID)
--catalogUnity Catalog name (default: main)
--schemaTarget schema for bronze/raw data (default: default)

CDC

Databricks supports CDC with Unity Catalog MERGE on Delta tables. Skippr automatically creates _skippr_order_token columns and tombstone tables.

See CDC Destinations -- Databricks for details.