Connect your AI Agents to Databricks in minutes

Available tools
list_clusters
List all clusters in the Databricks workspace. Returns cluster IDs, names, states, and configurations. Use this to find cluster IDs for other operations.
get_cluster
Get detailed information about a specific Databricks cluster including state, configuration, and resource allocation. Use list_clusters first to find the cluster ID.
create_cluster
Create a new Databricks cluster. Requires cluster name, Spark version, and node type. Specify num_workers for fixed size or autoscale_min/max_workers for autoscaling.
start_cluster
Start a terminated Databricks cluster. The cluster must be in TERMINATED state. Use list_clusters to find clusters and their states.
terminate_cluster
Terminate a running Databricks cluster. This stops the cluster but preserves its configuration for restarting. Use list_clusters to find cluster IDs.
list_jobs
List jobs in the Databricks workspace with optional name filter and pagination. Returns job IDs, names, and settings. Use page_token from response for next page.
get_job
Get detailed information about a specific Databricks job including tasks, schedule, and configuration. Use list_jobs first to find the job ID.
create_job
Create a new Databricks job with one or more tasks. Each task needs a task_key and type (notebook_task, spark_python_task, sql_task, etc). Supports scheduling with cron expressions.
delete_job
Permanently delete a Databricks job. This also cancels any active runs. Use list_jobs to find the job ID.
run_job_now
Trigger an immediate run of a Databricks job. Optionally pass notebook_params or python_named_params to override defaults. Use list_jobs to find the job ID.
list_job_runs
List job runs in the Databricks workspace. Filter by job_id, active_only, or completed_only. Supports offset/limit pagination. Returns run IDs, states, and timing info.
get_job_run
Get detailed information about a specific job run including state, timing, and task details. Use list_job_runs to find the run ID.

How to set up Merge Agent Handler
In an mcp.json file, add the configuration below, and restart Cursor.
Learn more in the official documentation ↗
1{
2 "mcpServers": {
3 "agent-handler": {
4 "url": "https://ah-api-develop.merge.dev/api/v1/tool-packs/{TOOL_PACK_ID}/registered-users/{REGISTERED_USER_ID}/mcp",
5 "headers": {
6 "Authorization": "Bearer yMt*****"
7 }
8 }
9 }
10}
11Open your Claude Desktop configuration file and add the server configuration below. You'll also need to restart the application for the changes to take effect.
Make sure Claude is using the Node v20+.
Learn more in the official documentation ↗
1{
2 "mcpServers": {
3 "agent-handler": {
4 "command": "npx",
5 "args": [
6 "-y",
7 "mcp-remote@latest",
8 "https://ah-api-develop.merge.dev/api/v1/tool-packs/{TOOL_PACK_ID}/registered-users/{REGISTERED_USER_ID}/mcp",
9 "--header",
10 "Authorization: Bearer ${AUTH_TOKEN}"
11 ],
12 "env": {
13 "AUTH_TOKEN": "yMt*****"
14 }
15 }
16 }
17}Open your Windsurf MCP configuration file and add the server configuration below.
Click on the refresh button in the top right of the Manage MCP server page or in the top right of the chat box in the box icon.
Learn more in the official documentation ↗
1{
2 "mcpServers": {
3 "agent-handler": {
4 "command": "npx",
5 "args": [
6 "-y",
7 "mcp-remote@latest",
8 "https://ah-api.merge.dev/api/v1/tool-packs/<tool-pack-id>/registered-users/<registered-user-id>/mcp",
9 "--header",
10 "Authorization: Bearer ${AUTH_TOKEN}"
11 ],
12 "env": {
13 "AUTH_TOKEN": "<ah-production-access-key>"
14 }
15 }
16 }
17 }In Command Palette (Cmd+Shift+P on macOS, Ctrl+Shift+P on Windows), run "MCP: Open User Configuration".
You can then add the configuration below and press "start" right under servers. Enter the auth token when prompted.
Learn more in the official documentation ↗
1{
2 "inputs": [
3 {
4 "type": "promptString",
5 "id": "agent-handler-auth",
6 "description": "Agent Handler AUTH_TOKEN", // "yMt*****" when prompt
7 "password": true
8 }
9 ],
10 "servers": {
11 "agent-handler": {
12 "type": "stdio",
13 "command": "npx",
14 "args": [
15 "-y",
16 "mcp-remote@latest",
17 "https://ah-api-develop.merge.dev/api/v1/tool-packs/{TOOL_PACK_ID}/registered-users/{REGISTERED_USER_ID}/mcp",
18 "--header",
19 "Authorization: Bearer ${input:agent-handler-auth}"
20 ]
21 }
22 }
23}FAQs on using Merge's Databricks MCP server
FAQs on using Merge's Databricks MCP server
What is a Databricks MCP server?
It’s an MCP server that exposes data and functionality from a Databricks instance via tools. Your agents can invoke these tools to create new clusters, list job runs in a given workspace, delete a job, and more.
Databricks offers a suite of “official” MCP servers:
- Vector Search retrieves relevant documents from vector indexes
- Genie queries structured data via natural language
- Databricks SQL executes AI-generated SQL for data operations
- Unity Catalog Functions run predefined SQL functions as tools
You can also use a hosted MCP platform that offers dozens of tools out of the box, such as Merge Agent Handler.
How can I use the Databricks MCP server?
Here are some popular use cases:
- Natural-language analytics over Lakehouse data: Ask questions in plain English, translate into safe SQL/Spark, run in Databricks, and return results with pointers to the underlying tables/views
- Data quality monitoring and incident automation: Run scheduled freshness/volume/schema checks, detect anomalies, and automatically open an incident or ticket with the failing queries and impact summary
- Pipeline debugging and root-cause analysis: When jobs fail or metrics drift, inspect recent runs/logs and propose likely causes and next-step fixes
- Cost and performance optimization: Identify expensive queries and inefficient cluster usage, then recommend prioritized changes (e.g., query rewrites, partitioning strategy, cluster sizing) with measurable savings targets
- Governed self-serve data discovery and access: Help teams find the right datasets, explain definitions, and streamline access requests/approvals while keeping an audit trail
What makes Merge Agent Handler’s Databricks MCP server better than alternative Databricks MCP servers?
Here are just a few reasons:
- Enterprise-grade connector infrastructure: Merge Agent Handler provides fully managed connectors that handle authentication, rate limiting, error handling, and lifecycle management so developers can focus on building agent logic instead of maintaining integrations
- Built-in security gateway with DLP controls: Agent Handler includes a Security Gateway that scans tool inputs and outputs for sensitive data and enforces security policies. Teams can configure default or custom rules to block, redact, or mask data before it reaches an agent or external system.
- Flexible authentication and credential management: Authentication is handled directly by the platform through registered users and connector authentication flows. And if a user hasn’t authenticated a connector yet, Agent Handler signals the application to prompt authentication via Merge Link
- Real-time observability and searchable audit logs: Agent Handler logs every tool call and underlying API request so teams can monitor agent activity, debug issues, and track system behavior. These logs are fully searchable and form a complete audit trail for compliance, debugging, and operational monitoring
- Customizable connectors and Tool Packs: With Connector Studio, teams can combine Databricks with other enterprise connectors, modify tool definitions, or import additional MCP servers. This enables building Tool Packs tailored to specific agent workflows rather than relying on a single-purpose Databricks MCP implementation
Can I set custom security rules for Databricks tool calls in Merge Agent Handler?
Yes. Merge’s platform materials say Agent Handler includes default security rules and custom security rules, and the Security Gateway scans tool inputs and responses for sensitive data before taking action based on platform configuration.
Here are a few examples:
- Block PII patterns (inputs + outputs): Block any Databricks tool call if payloads contain SSNs or credit card numbers (regex-based)
- Block destructive SQL / admin operations: Block queries containing <code class="blog_inline-code">DROP</code>, <code class="blog_inline-code">TRUNCATE</code>, <code class="blog_inline-code">GRANT</code>, <code class="blog_inline-code">ALTER</code>, or <code class="blog_inline-code">DELETE</code> without a <code class="blog_inline-code">WHERE</code>
- Prod boundary rule: Block calls that target prod workspaces/catalogs/schemas unless the Tool Pack (or Registered User group) is explicitly allowed
- Sensitive table/column allowlist: Block queries that touch specified tables or fields (e.g., salary) unless running under a privileged Tool Pack
- Redact secrets: Allow execution, but redact API keys, bearer tokens, and connection strings from requests/logs to prevent credential leakage
How can I start using Merge Agent Handler’s Databricks MCP server?
You can follow these steps:
1. Create an Agent Handler account and grab your API key from the Agent Handler dashboard. This API key is what your MCP client uses to authenticate to the Agent Handler MCP endpoint.
2. Create a Tool Pack for the Databricks workflow (e.g., “Databricks Analytics”). Tool Packs define which connectors and tools your agent is allowed to call.
3. Create (or select) a Registered User that your agent will act “as” when making tool calls. Registered Users are the identity containers that hold credentials (user-level or shared auth).
4. Authenticate Databricks for that Registered User (or configure shared auth if you want one org-level Databricks credential reused across users). In practice, this is typically done via Agent Handler Link in your product so end users can connect without you handling credentials directly.
5. Connect your MCP client to the Agent Handler MCP endpoint for that Tool Pack + Registered User. Then your client can <code class="blog_inline-code">list_tools</code> and <code class="blog_inline-code">call_tool</code> for the Databricks tools you enabled in the Tool Pack.
6. Run a few real prompts and monitor the results using Agent Handler’s logs (and add rules/alerts as needed) to verify the tool calls and outputs look right before pushing the connector to production.
Explore other MCP servers built and managed by Merge


























































































Ready to try it out?
Whether you're an engineer experimenting with agents or a product manager looking to add tools, you can get started for free now

