Connect your AI Agents to Databricks in minutes

Merge lets you securely connect your agents to Databricks and thousands of tools instantly

Available tools

list_clusters

List all clusters in the Databricks workspace. Returns cluster IDs, names, states, and configurations. Use this to find cluster IDs for other operations.

get_cluster

Get detailed information about a specific Databricks cluster including state, configuration, and resource allocation. Use list_clusters first to find the cluster ID.

create_cluster

Create a new Databricks cluster. Requires cluster name, Spark version, and node type. Specify num_workers for fixed size or autoscale_min/max_workers for autoscaling.

start_cluster

Start a terminated Databricks cluster. The cluster must be in TERMINATED state. Use list_clusters to find clusters and their states.

terminate_cluster

Terminate a running Databricks cluster. This stops the cluster but preserves its configuration for restarting. Use list_clusters to find cluster IDs.

list_jobs

List jobs in the Databricks workspace with optional name filter and pagination. Returns job IDs, names, and settings. Use page_token from response for next page.

get_job

Get detailed information about a specific Databricks job including tasks, schedule, and configuration. Use list_jobs first to find the job ID.

create_job

Create a new Databricks job with one or more tasks. Each task needs a task_key and type (notebook_task, spark_python_task, sql_task, etc). Supports scheduling with cron expressions.

delete_job

Permanently delete a Databricks job. This also cancels any active runs. Use list_jobs to find the job ID.

run_job_now

Trigger an immediate run of a Databricks job. Optionally pass notebook_params or python_named_params to override defaults. Use list_jobs to find the job ID.

list_job_runs

List job runs in the Databricks workspace. Filter by job_id, active_only, or completed_only. Supports offset/limit pagination. Returns run IDs, states, and timing info.

get_job_run

Get detailed information about a specific job run including state, timing, and task details. Use list_job_runs to find the run ID.

How to set up Merge Agent Handler

In an mcp.json file, add the configuration below, and restart Cursor.

Learn more in the official documentation ↗

1{
2  "mcpServers": {
3    "agent-handler": {
4      "url": "https://ah-api-develop.merge.dev/api/v1/tool-packs/{TOOL_PACK_ID}/registered-users/{REGISTERED_USER_ID}/mcp",
5      "headers": {
6        "Authorization": "Bearer yMt*****"
7      }
8    }
9  }
10}
11
Copy Code

Open your Claude Desktop configuration file and add the server configuration below. You'll also need to restart the application for the changes to take effect.

Make sure Claude is using the Node v20+.

Learn more in the official documentation ↗

1{
2  "mcpServers": {
3    "agent-handler": {
4      "command": "npx",
5      "args": [
6        "-y",
7        "mcp-remote@latest",
8        "https://ah-api-develop.merge.dev/api/v1/tool-packs/{TOOL_PACK_ID}/registered-users/{REGISTERED_USER_ID}/mcp",
9        "--header",
10        "Authorization: Bearer ${AUTH_TOKEN}"
11      ],
12      "env": {
13        "AUTH_TOKEN": "yMt*****"
14      }
15    }
16  }
17}
Copy Code
Copied!

Open your Windsurf MCP configuration file and add the server configuration below.
Click on the refresh button in the top right of the Manage MCP server page or in the top right of the chat box in the box icon.

Learn more in the official documentation ↗

1{
2    "mcpServers": {
3      "agent-handler": {
4        "command": "npx",
5        "args": [
6          "-y",
7          "mcp-remote@latest",
8          "https://ah-api.merge.dev/api/v1/tool-packs/<tool-pack-id>/registered-users/<registered-user-id>/mcp",
9          "--header",
10          "Authorization: Bearer ${AUTH_TOKEN}"
11        ],
12        "env": {
13          "AUTH_TOKEN": "<ah-production-access-key>"
14        }
15      }
16    }
17  }
Copy Code

In Command Palette (Cmd+Shift+P on macOS, Ctrl+Shift+P on Windows), run "MCP: Open User Configuration".

You can then add the configuration below and press "start" right under servers. Enter the auth token when prompted.

Learn more in the official documentation ↗

1{
2  "inputs": [
3    {
4      "type": "promptString",
5      "id": "agent-handler-auth",
6      "description": "Agent Handler AUTH_TOKEN", // "yMt*****" when prompt
7      "password": true
8    }
9  ],
10  "servers": {
11    "agent-handler": {
12      "type": "stdio",
13      "command": "npx",
14      "args": [
15        "-y",
16        "mcp-remote@latest",
17        "https://ah-api-develop.merge.dev/api/v1/tool-packs/{TOOL_PACK_ID}/registered-users/{REGISTERED_USER_ID}/mcp",
18        "--header",
19        "Authorization: Bearer ${input:agent-handler-auth}"
20      ]
21    }
22  }
23}
Copy Code

FAQs on using Merge's Databricks  MCP server

FAQs on using Merge's Databricks  MCP server

What is a Databricks MCP server?

It’s an MCP server that exposes data and functionality from a Databricks instance via tools. Your agents can invoke these tools to create new clusters, list job runs in a given workspace, delete a job, and more.

Databricks offers a suite of “official” MCP servers:

  • Vector Search retrieves relevant documents from vector indexes 
  • Genie queries structured data via natural language 
  • Databricks SQL executes AI-generated SQL for data operations 
  • Unity Catalog Functions run predefined SQL functions as tools

You can also use a hosted MCP platform that offers dozens of tools out of the box, such as Merge Agent Handler.

How can I use the Databricks MCP server?

Here are some popular use cases:

  • Natural-language analytics over Lakehouse data: Ask questions in plain English, translate into safe SQL/Spark, run in Databricks, and return results with pointers to the underlying tables/views
  • Data quality monitoring and incident automation: Run scheduled freshness/volume/schema checks, detect anomalies, and automatically open an incident or ticket with the failing queries and impact summary
  • Pipeline debugging and root-cause analysis: When jobs fail or metrics drift, inspect recent runs/logs and propose likely causes and next-step fixes
  • Cost and performance optimization: Identify expensive queries and inefficient cluster usage, then recommend prioritized changes (e.g., query rewrites, partitioning strategy, cluster sizing) with measurable savings targets
  • Governed self-serve data discovery and access: Help teams find the right datasets, explain definitions, and streamline access requests/approvals while keeping an audit trail

What makes Merge Agent Handler’s Databricks MCP server better than alternative Databricks MCP servers?

Here are just a few reasons:

  • Enterprise-grade connector infrastructure: Merge Agent Handler provides fully managed connectors that handle authentication, rate limiting, error handling, and lifecycle management so developers can focus on building agent logic instead of maintaining integrations
  • Built-in security gateway with DLP controls: Agent Handler includes a Security Gateway that scans tool inputs and outputs for sensitive data and enforces security policies. Teams can configure default or custom rules to block, redact, or mask data before it reaches an agent or external system.
  • Flexible authentication and credential management: Authentication is handled directly by the platform through registered users and connector authentication flows. And if a user hasn’t authenticated a connector yet, Agent Handler signals the application to prompt authentication via Merge Link
  • Real-time observability and searchable audit logs: Agent Handler logs every tool call and underlying API request so teams can monitor agent activity, debug issues, and track system behavior. These logs are fully searchable and form a complete audit trail for compliance, debugging, and operational monitoring
  • Customizable connectors and Tool Packs: With Connector Studio, teams can combine Databricks with other enterprise connectors, modify tool definitions, or import additional MCP servers. This enables building Tool Packs tailored to specific agent workflows rather than relying on a single-purpose Databricks MCP implementation

Can I set custom security rules for Databricks tool calls in Merge Agent Handler?

Yes. Merge’s platform materials say Agent Handler includes default security rules and custom security rules, and the Security Gateway scans tool inputs and responses for sensitive data before taking action based on platform configuration. 

Here are a few examples:

  • Block PII patterns (inputs + outputs): Block any Databricks tool call if payloads contain SSNs or credit card numbers (regex-based)
  • Block destructive SQL / admin operations: Block queries containing <code class="blog_inline-code">DROP</code>, <code class="blog_inline-code">TRUNCATE</code>, <code class="blog_inline-code">GRANT</code>, <code class="blog_inline-code">ALTER</code>, or <code class="blog_inline-code">DELETE</code> without a <code class="blog_inline-code">WHERE</code>
  • Prod boundary rule: Block calls that target prod workspaces/catalogs/schemas unless the Tool Pack (or Registered User group) is explicitly allowed
  • Sensitive table/column allowlist: Block queries that touch specified tables or fields (e.g., salary) unless running under a privileged Tool Pack
  • Redact secrets: Allow execution, but redact API keys, bearer tokens, and connection strings from requests/logs to prevent credential leakage

How can I start using Merge Agent Handler’s Databricks MCP server?

You can follow these steps:

1. Create an Agent Handler account and grab your API key from the Agent Handler dashboard. This API key is what your MCP client uses to authenticate to the Agent Handler MCP endpoint.

2. Create a Tool Pack for the Databricks workflow (e.g., “Databricks Analytics”). Tool Packs define which connectors and tools your agent is allowed to call.

3. Create (or select) a Registered User that your agent will act “as” when making tool calls. Registered Users are the identity containers that hold credentials (user-level or shared auth).

4. Authenticate Databricks for that Registered User (or configure shared auth if you want one org-level Databricks credential reused across users). In practice, this is typically done via Agent Handler Link in your product so end users can connect without you handling credentials directly.

5. Connect your MCP client to the Agent Handler MCP endpoint for that Tool Pack + Registered User. Then your client can <code class="blog_inline-code">list_tools</code> and <code class="blog_inline-code">call_tool</code> for the Databricks tools you enabled in the Tool Pack.

6. Run a few real prompts and monitor the results using Agent Handler’s logs (and add rules/alerts as needed) to verify the tool calls and outputs look right before pushing the connector to production.

Explore other MCP servers built and managed by Merge

activecampaign
ActiveCampaign
airtable
Airtable
amadeus
Amadeus
amazon_s3
Amazon S3
anaplan
Anaplan
apollo
Apollo
arize
Arize
articulate
Articulate Reach 360
asana
Asana
attio
Attio
aviationstack
Aviationstack
bamboohr
BambooHR
basecamp
Basecamp
bitbucket
Bitbucket
bitly
Bitly
box
Box
cms_coverage
CMS Coverage
calendly
Calendly
canva
Canva
clickup
ClickUp
clinicaltrials
ClinicalTrials.gov
cloudflare
Cloudflare
coda
Coda
confluence
Confluence
contentful
Contentful
crustdata
Crustdata
datadog
Datadog
docusign
DocuSign
doordash
DoorDash
dropbox
Dropbox
duffel
Duffel
dynamics365
Dynamics 365 Sales
exa
Exa
expensify
Expensify
factset
FactSet
figma
Figma
firehydrant
FireHydrant
firecrawl
Firecrawl
fireflies
Fireflies
foursquare
Foursquare
freshbooks
FreshBooks
freshdesk
Freshdesk
freshservice
Freshservice
front
Front
gamma
Gamma
github
GitHub
gitlab
GitLab
gmail
Gmail
gong
Gong
google_bigquery
Google BigQuery
google_calendar
Google Calendar
google_docs
Google Docs
google_drive
Google Drive
google_maps
Google Maps
google_meet
Google Meet
google_sheets
Google Sheets
google_slides
Google Slides
greenhouse
Greenhouse
hibob
HiBob
hubspot
HubSpot
intercom
Intercom
jira
Jira
jira_service_management
Jira Service Management
klaviyo
Klaviyo
linear
Linear
linkedin
LinkedIn
make
Make
microsoft_teams
Microsoft Teams
microsoft_teams_gcc
Microsoft Teams GCC High
miro
Miro
monday
Monday.com
npi_registry
NPI Registry
notion
Notion
onedrive
OneDrive
onenote
OneNote
oracle_hcm
Oracle HCM
oura
Oura
outlook
Outlook
pagerduty
PagerDuty
paypal
PayPal
pipedrive
Pipedrive
pubmed
PubMed
pylon
Pylon
quartr
Quartr
quickbooks_online
QuickBooks Online
ramp
Ramp
rootly
Rootly
sapsf
SAP SuccessFactors
sabre
Sabre
salesforce
Salesforce
sendgrid
SendGrid
sentry
Sentry
servicenow
ServiceNow
sharepoint
SharePoint
shopify
Shopify
slack
Slack
snowflake
Snowflake
spotify
Spotify
square
Square
straker
Straker

Ready to try it out?

Whether you're an engineer experimenting with agents or a product manager looking to add tools, you can get started for free now