Ridge AI's Databricks connector enables direct access to your Unity Catalog tables and Delta Lake data. Query your Databricks data using either simple table references or custom SQL, with automated ingestion to Ridge AI's analytics platform.
You'll need a Databricks workspace with Unity Catalog enabled, and catalog access to the tables you want to query.
You'll need four pieces of information to connect:
1. Access Token
Generate a personal access token or service principal token:
2. Workspace Endpoint
Your workspace URL in the format:https://dbc-xxxxx-xxxx.cloud.databricks.com
Find this in your browser's address bar when logged into Databricks, or in your workspace settings.
3. AWS Region
The AWS region hosting your workspace (e.g., us-west-2, us-east-1, eu-west-1).
Check your workspace URL or Databricks account console for region information.
4. Catalog Name
The Unity Catalog name you want to access (e.g., main, production, analytics).
Unity Catalog uses a three-level namespace:catalog.schema.table


Connections manage authorization and authentication to a data source, such as Databricks.
Once you have a Connection, create a Data Set. Data Sets point to specific tables and may specify queries and refresh schedules. To set up a Data Set,
Option 1: Table Reference (simplest)
catalog.schema.table_name
Example: main.sales.transactions
Option 2: Custom SQL (for filtering or transformations)
SELECT * FROM catalog.schema.table_name WHERE date > '2024-01-01'
Example:
SELECT
customer_id,
SUM(amount) as total_amount
FROM main.sales.transactions
WHERE year = 2024
GROUP BY customer_id
catalog.schema.table0 0 * * *0 * * * *0 8 * * 1-5When ingestion runs:
Use service principals for production: Personal access tokens expire and are tied to individual users. Service principals provide better security and stability for automated ingestion.
Scope queries appropriately: Query only the data you need. Use WHERE clauses to filter by date ranges or other criteria to improve performance.
Leverage Unity Catalog permissions: Use Databricks' built-in access controls rather than sharing credentials broadly.
Monitor ingestion schedules: Set schedules based on how frequently your source data updates. Over-scheduling wastes resources; under-scheduling delays insights.
Test queries in Databricks first: Before configuring a data source in Ridge AI, run your SQL query in the Databricks SQL editor to verify it works correctly.
TABLE_DB_STORAGE ❌
TABLE_EXTERNAL ✅
TABLE_DELTA_EXTERNAL ✅
Symptom: "Invalid token" or "Authentication failed"
Solutions:
Symptom: "Access denied" or "Insufficient privileges"
Solutions:
GRANT USE CATALOG ON CATALOG catalog_name TO userGRANT USE SCHEMA ON SCHEMA schema_name TO userGRANT SELECT ON TABLE table_name TO userSymptom: "Cannot connect" or "Endpoint unreachable"
Solutions:
https://dbc-xxxxx-xxxx.cloud.databricks.comSymptom: "Table not found" or "SQL syntax error"
Solutions:
catalog.schema.table (not just table)