Databricks
Uploads data to Databricks as Parquet files via the Files API, with optional COPY INTO.
Configuration
yaml
warehouse:
kind: databricks
workspace_url: "https://my-workspace.cloud.databricks.com"
token: ${DATABRICKS_TOKEN}
warehouse_id: "abc123"
catalog: main
schema: default| Field | Default | Description |
|---|---|---|
workspace_url | (required) | Databricks workspace URL |
token | (required) | Personal access token |
warehouse_id | SQL warehouse ID (enables COPY INTO) | |
catalog | main | Unity Catalog name |
schema | default | Schema name |
CLI
bash
skippr connect warehouse databricks \
--workspace-url https://dbc-xxxxxxxx.cloud.databricks.com \
--token "$DATABRICKS_TOKEN" \
--warehouse-id abc123 \
--catalog main \
--schema default| Flag | Description |
|---|---|
--workspace-url | Databricks workspace URL (or set DATABRICKS_HOST) |
--token | Personal access token (or set DATABRICKS_TOKEN) |
--warehouse-id | SQL warehouse ID (or set DATABRICKS_WAREHOUSE_ID) |
--catalog | Unity Catalog name (default: main) |
--schema | Target schema for bronze/raw data (default: default) |
CDC
Databricks supports CDC with Unity Catalog MERGE on Delta tables. Skippr automatically creates _skippr_order_token columns and tombstone tables.
See CDC Destinations -- Databricks for details.
