A

Azure ADX

community
search

Query and analyze Azure Data Explorer databases.

Azure Data Explorer MCP Server

<a href="https://glama.ai/mcp/servers/1yysyd147h"> <img width="380" height="200" src="https://glama.ai/mcp/servers/1yysyd147h/badge" /> </a>
A Model Context Protocol (MCP) server for Azure Data Explorer/Eventhouse in Microsoft Fabric.
This provides access to your Azure Data Explorer/Eventhouse clusters and databases through standardized MCP interfaces, allowing AI assistants to execute KQL queries and explore your data.

Features

  • Execute KQL queries against Azure Data Explorer
  • Discover and explore database resources
    • List tables in the configured database
    • View table schemas
    • Sample data from tables
    • Get table statistics/details
  • Authentication support
    • Token credential support (Azure CLI, MSI, etc.)
  • Docker containerization support
  • Provide interactive tools for AI assistants
The list of tools is configurable, so you can choose which tools you want to make available to the MCP client. This is useful if you don't use certain functionality or if you don't want to take up too much of the context window.

Usage

  1. Login to your Azure account which has the permission to the ADX cluster using Azure CLI.
  2. Configure the environment variables for your ADX cluster, either through a
    .env
    file or system environment variables:
# Required: Azure Data Explorer configuration
ADX_CLUSTER_URL=https://yourcluster.region.kusto.windows.net
ADX_DATABASE=your_database
  1. Add the server configuration to your client configuration file. For example, for Claude Desktop:
{
  "mcpServers": {
    "adx": {
      "command": "uv",
      "args": [
        "--directory",
        "<full path to adx-mcp-server directory>",
        "run",
        "src/adx_mcp_server/main.py"
      ],
      "env": {
        "ADX_CLUSTER_URL": "https://yourcluster.region.kusto.windows.net",
        "ADX_DATABASE": "your_database"
      }
    }
  }
}
Note: if you see
Error: spawn uv ENOENT
in Claude Desktop, you may need to specify the full path to
uv
or set the environment variable
NO_UV=1
in the configuration.

Docker Usage

This project includes Docker support for easy deployment and isolation.

Building the Docker Image

Build the Docker image using:
docker build -t adx-mcp-server .

Running with Docker

You can run the server using Docker in several ways:

Using docker run directly:

docker run -it --rm \
  -e ADX_CLUSTER_URL=https://yourcluster.region.kusto.windows.net \
  -e ADX_DATABASE=your_database \
  adx-mcp-server

Using docker-compose:

Create a
.env
file with your Azure Data Explorer credentials and then run:
docker-compose up

Running with Docker in Claude Desktop

To use the containerized server with Claude Desktop, update the configuration to use Docker with the environment variables:
{
  "mcpServers": {
    "adx": {
      "command": "docker",
      "args": [
        "run",
        "--rm",
        "-i",
        "-e", "ADX_CLUSTER_URL",
        "-e", "ADX_DATABASE",
        "adx-mcp-server"
      ],
      "env": {
        "ADX_CLUSTER_URL": "https://yourcluster.region.kusto.windows.net",
        "ADX_DATABASE": "your_database"
      }
    }
  }
}
This configuration passes the environment variables from Claude Desktop to the Docker container by using the
-e
flag with just the variable name, and providing the actual values in the
env
object.

Using as a Dev Container / GitHub Codespace

This repository can also be used as a development container for a seamless development experience. The dev container setup is located in the
devcontainer-feature/adx-mcp-server
folder.
For more details, check the devcontainer README.

Development

Contributions are welcome! Please open an issue or submit a pull request if you have any suggestions or improvements.
This project uses
uv
to manage dependencies. Install
uv
following the instructions for your platform:
curl -LsSf https://astral.sh/uv/install.sh | sh
You can then create a virtual environment and install the dependencies with:
uv venv
source .venv/bin/activate  # On Unix/macOS
.venv\Scripts\activate     # On Windows
uv pip install -e .

Project Structure

The project has been organized with a
src
directory structure:
adx-mcp-server/
├── src/
│   └── adx_mcp_server/
│       ├── __init__.py      # Package initialization
│       ├── server.py        # MCP server implementation
│       ├── main.py          # Main application logic
├── Dockerfile               # Docker configuration
├── docker-compose.yml       # Docker Compose configuration
├── .dockerignore            # Docker ignore file
├── pyproject.toml           # Project configuration
└── README.md                # This file

Testing

The project includes a comprehensive test suite that ensures functionality and helps prevent regressions.
Run the tests with pytest:
# Install development dependencies
uv pip install -e ".[dev]"

# Run the tests
pytest

# Run with coverage report
pytest --cov=src --cov-report=term-missing
Tests are organized into:
  • Configuration validation tests
  • Server functionality tests
  • Error handling tests
  • Main application tests
When adding new features, please also add corresponding tests.

Tools

ToolCategoryDescription
execute_query
QueryExecute a KQL query against Azure Data Explorer
list_tables
DiscoveryList all tables in the configured database
get_table_schema
DiscoveryGet the schema for a specific table
sample_table_data
DiscoveryGet sample data from a table with optional sample size

License

MIT

Related Servers

B

Brave Search

reference

Web and local search using Brave's Search API

View Details
G

Git

reference

Tools to read, search, and manipulate Git repositories

View Details
G

Google Drive

reference

File access and search capabilities for Google Drive

View Details
Aiven logo

Aiven

official

Navigate your [Aiven projects](https://go.aiven.io/mcp-server) and interact with the PostgreSQL®, Apache Kafka®, ClickHouse® and OpenSearch® services

View Details
Apify logo

Apify

official

[Actors MCP Server](https://apify.com/apify/actors-mcp-server): Use 3,000+ pre-built cloud tools to extract data from websites, e-commerce, social media, search engines, maps, and more

View Details