Skip to content

263 tools for Databricks via MCP. SDK-first, covers Unity Catalog, SQL, Compute, Jobs, Serving, Vector Search, Apps, Lakebase, and more.

License

Notifications You must be signed in to change notification settings

pramodbhatofficial/databricks-mcp-server

Repository files navigation

Databricks MCP Server

PyPI CI Python 3.10+ License: Apache 2.0

A comprehensive Model Context Protocol (MCP) server for Databricks, built on the official Databricks Python SDK.

Provides 263 tools and 8 prompt templates across 28 service domains, giving AI assistants full access to the Databricks platform.

Features

  • SDK-first: Uses databricks-sdk for type safety and automatic API freshness
  • Comprehensive: Covers Unity Catalog, SQL, Compute, Jobs, Pipelines, Serving, Vector Search, Apps, Lakebase, Dashboards, Genie, Secrets, IAM, Connections, Experiments, and Delta Sharing
  • Zero custom auth: Delegates authentication entirely to the SDK (PAT, OAuth, Azure AD, service principal -- all automatic)
  • Selective loading: Include/exclude tool modules via environment variables
  • MCP Resources: Read-only workspace context (URL, current user, auth type)

Quick Start

Installation

pip install databricks-sdk-mcp

Or run with Docker:

docker run -i -e DATABRICKS_HOST=... -e DATABRICKS_TOKEN=... databricks-mcp

Or install from source:

git clone https://github.com/pramodbhatofficial/databricks-mcp-server.git
cd databricks-mcp-server
pip install -e ".[dev]"

Authentication

Authentication is handled by the Databricks SDK. Set one of:

Personal Access Token (simplest):

export DATABRICKS_HOST=https://your-workspace.databricks.com
export DATABRICKS_TOKEN=dapi...

OAuth (M2M):

export DATABRICKS_HOST=https://your-workspace.databricks.com
export DATABRICKS_CLIENT_ID=...
export DATABRICKS_CLIENT_SECRET=...

Other methods: Azure AD, Databricks CLI profile, Azure Managed Identity -- all auto-detected by the SDK.

Running

databricks-mcp

This starts the MCP server using stdio transport.

Integrations

Claude Code (Terminal)

Add to ~/.claude/settings.json or your project's .claude/settings.json:

{
  "mcpServers": {
    "databricks": {
      "command": "databricks-mcp",
      "env": {
        "DATABRICKS_HOST": "https://your-workspace.databricks.com",
        "DATABRICKS_TOKEN": "dapi..."
      }
    }
  }
}

Then restart Claude Code. Verify with /mcp to see the registered tools.

Claude Desktop

Add to your Claude Desktop config file:

  • macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
  • Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "databricks": {
      "command": "databricks-mcp",
      "env": {
        "DATABRICKS_HOST": "https://your-workspace.databricks.com",
        "DATABRICKS_TOKEN": "dapi..."
      }
    }
  }
}

Restart Claude Desktop. The Databricks tools will appear in the tool picker.

Cursor

Add to .cursor/mcp.json in your project root (or ~/.cursor/mcp.json for global):

{
  "mcpServers": {
    "databricks": {
      "command": "databricks-mcp",
      "env": {
        "DATABRICKS_HOST": "https://your-workspace.databricks.com",
        "DATABRICKS_TOKEN": "dapi..."
      }
    }
  }
}

Open Cursor Settings > MCP to verify the server is connected.

Windsurf

Add to ~/.codeium/windsurf/mcp_config.json:

{
  "mcpServers": {
    "databricks": {
      "command": "databricks-mcp",
      "env": {
        "DATABRICKS_HOST": "https://your-workspace.databricks.com",
        "DATABRICKS_TOKEN": "dapi..."
      }
    }
  }
}

VS Code (Copilot)

Add to .vscode/mcp.json in your project:

{
  "servers": {
    "databricks": {
      "command": "databricks-mcp",
      "env": {
        "DATABRICKS_HOST": "https://your-workspace.databricks.com",
        "DATABRICKS_TOKEN": "dapi..."
      }
    }
  }
}

Zed

Add to Zed's settings (~/.config/zed/settings.json):

{
  "context_servers": {
    "databricks": {
      "command": {
        "path": "databricks-mcp",
        "env": {
          "DATABRICKS_HOST": "https://your-workspace.databricks.com",
          "DATABRICKS_TOKEN": "dapi..."
        }
      }
    }
  }
}

Any MCP Client (Generic stdio)

The server uses stdio transport. Connect from any MCP-compatible client:

# Set auth env vars
export DATABRICKS_HOST=https://your-workspace.databricks.com
export DATABRICKS_TOKEN=dapi...

# Start the server (communicates via stdin/stdout)
databricks-mcp

Tip: Load Only What You Need

If your MCP client struggles with many tools, use selective loading to reduce the tool count:

{
  "mcpServers": {
    "databricks": {
      "command": "databricks-mcp",
      "env": {
        "DATABRICKS_HOST": "https://your-workspace.databricks.com",
        "DATABRICKS_TOKEN": "dapi...",
        "DATABRICKS_MCP_TOOLS_INCLUDE": "unity_catalog,sql,compute,jobs"
      }
    }
  }
}

Tool Modules

Module Tools Description
unity_catalog 23 Catalogs, schemas, tables, volumes, functions, registered models
sql 14 Warehouses, SQL execution, queries, alerts, history
workspace 10 Notebooks, files, repos
compute 18 Clusters, instance pools, policies, node types, Spark versions
jobs 13 Jobs, runs, tasks, repair, cancel all
pipelines 8 DLT / Lakeflow pipelines
serving 10 Serving endpoints, model versions, OpenAPI
vector_search 10 Vector search endpoints, indexes, sync
apps 10 Databricks Apps lifecycle
database 10 Lakebase PostgreSQL instances
dashboards 9 Lakeview AI/BI dashboards, published views
genie 5 Genie AI/BI conversations
secrets 8 Secret scopes and secrets
iam 16 Users, groups, service principals, permissions, current user
connections 5 External connections
experiments 14 MLflow experiments, runs, artifacts, metrics, params
sharing 11 Delta Sharing shares, recipients, providers
files 12 DBFS and UC Volumes file operations
grants 3 Unity Catalog permission grants (GRANT/REVOKE)
storage 10 Storage credentials and external locations
metastores 8 Unity Catalog metastore management
online_tables 3 Online tables for low-latency serving
global_init_scripts 5 Workspace-wide init scripts
tokens 5 Personal access token management
git_credentials 5 Git credential management for repos
quality_monitors 8 Data quality monitoring and refreshes
command_execution 4 Interactive command execution on clusters
workflows 5 Composite multi-step operations (workspace status, schema setup, query preview)

Selective Tool Loading

With 263 tools, it's recommended to load only the modules you need. This improves agent performance and tool selection accuracy.

Role-Based Presets (Recommended)

Pick a preset that matches your role:

Preset Modules Tools Config
Data Engineer unity_catalog, sql, compute, jobs, pipelines, files, quality_monitors ~100 DATABRICKS_MCP_TOOLS_INCLUDE=unity_catalog,sql,compute,jobs,pipelines,files,quality_monitors
ML Engineer serving, vector_search, experiments, compute, unity_catalog, online_tables, files ~98 DATABRICKS_MCP_TOOLS_INCLUDE=serving,vector_search,experiments,compute,unity_catalog,online_tables,files
Platform Admin iam, secrets, tokens, metastores, compute, global_init_scripts, grants, storage ~85 DATABRICKS_MCP_TOOLS_INCLUDE=iam,secrets,tokens,metastores,compute,global_init_scripts,grants,storage
App Developer apps, database, sql, files, serving, secrets ~64 DATABRICKS_MCP_TOOLS_INCLUDE=apps,database,sql,files,serving,secrets
Data Analyst sql, unity_catalog, dashboards, genie, workspace ~61 DATABRICKS_MCP_TOOLS_INCLUDE=sql,unity_catalog,dashboards,genie,workspace
Minimal sql, unity_catalog ~37 DATABRICKS_MCP_TOOLS_INCLUDE=sql,unity_catalog

Example using a preset in Claude Code:

{
  "mcpServers": {
    "databricks": {
      "command": "databricks-mcp",
      "env": {
        "DATABRICKS_HOST": "https://your-workspace.databricks.com",
        "DATABRICKS_TOKEN": "dapi...",
        "DATABRICKS_MCP_TOOLS_INCLUDE": "unity_catalog,sql,compute,jobs,pipelines,files,quality_monitors"
      }
    }
  }
}

Custom Filtering

# Only include specific modules
export DATABRICKS_MCP_TOOLS_INCLUDE=unity_catalog,sql,serving

# Exclude specific modules (cannot combine with INCLUDE)
export DATABRICKS_MCP_TOOLS_EXCLUDE=iam,sharing,experiments

If INCLUDE is set, only those modules load. If EXCLUDE is set, everything except those modules loads. INCLUDE takes precedence if both are set.

Tool Discovery (For AI Agents)

The server includes built-in tool discovery to help AI agents find the right tools:

MCP Resources

URI Description
databricks://workspace/info Workspace URL, current user, auth type
databricks://tools/guide Tool catalog with module descriptions, use cases, and role presets

Agents can read databricks://tools/guide at connection time to understand what's available.

Discovery Tool

The databricks_tool_guide tool helps agents find the right tools during a conversation:

# Find tools for a specific task
databricks_tool_guide(task="run a SQL query")
databricks_tool_guide(task="deploy an ML model")
databricks_tool_guide(task="create a user")

# Get role-based recommendations
databricks_tool_guide(role="data_engineer")
databricks_tool_guide(role="ml_engineer")

This returns matching modules with descriptions and usage hints, so the agent knows exactly which databricks_* tools to call.

MCP Prompts (Guided Workflows)

The server includes 8 prompt templates that guide AI agents through multi-step Databricks workflows:

Prompt Description
explore_data_catalog Browse Unity Catalog structure (catalogs → schemas → tables)
query_data Find a warehouse, execute SQL, and format results
debug_failing_job Investigate a failing job: status, logs, error analysis
setup_ml_experiment Create an MLflow experiment and configure tracking
deploy_model Deploy a model to a serving endpoint
setup_data_pipeline Create a DLT pipeline with scheduling
workspace_health_check Audit clusters, warehouses, jobs, and endpoints
manage_permissions Review and update permissions on workspace objects

Prompts appear automatically in MCP clients that support them (e.g., Claude Desktop's prompt picker).

Docker

Run the MCP server in a container:

# Build
docker build -t databricks-mcp .

# Run with stdio
docker run -i \
  -e DATABRICKS_HOST=https://your-workspace.databricks.com \
  -e DATABRICKS_TOKEN=dapi... \
  databricks-mcp

# Run with SSE transport
docker run -p 8080:8080 \
  -e DATABRICKS_HOST=https://your-workspace.databricks.com \
  -e DATABRICKS_TOKEN=dapi... \
  databricks-mcp --transport sse --port 8080

# Run with selective modules
docker run -i \
  -e DATABRICKS_HOST=https://your-workspace.databricks.com \
  -e DATABRICKS_TOKEN=dapi... \
  -e DATABRICKS_MCP_TOOLS_INCLUDE=sql,unity_catalog \
  databricks-mcp

SSE Transport (Remote Server)

The server supports SSE transport for remote connections:

# Start as SSE server
databricks-mcp --transport sse --port 8080

# Custom host/port
databricks-mcp --transport sse --host 127.0.0.1 --port 3000

Connect from any MCP client that supports SSE:

{
  "mcpServers": {
    "databricks": {
      "url": "http://localhost:8080/sse"
    }
  }
}

Development

# Install with dev dependencies
pip install -e ".[dev]"

# Lint
ruff check databricks_mcp/

# Test
pytest tests/ -v

Author

Pramod Bhat

License

Apache 2.0 -- see LICENSE.

About

263 tools for Databricks via MCP. SDK-first, covers Unity Catalog, SQL, Compute, Jobs, Serving, Vector Search, Apps, Lakebase, and more.

Topics

Resources

License

Contributing

Stars

Watchers

Forks

Packages

No packages published

Contributors 2

  •  
  •