Michal 588b2a9e65
Some checks failed
CI/CD / lint (push) Successful in 4m0s
CI/CD / typecheck (push) Successful in 2m38s
CI/CD / test (push) Successful in 3m52s
CI/CD / build (push) Successful in 5m22s
CI/CD / publish-rpm (push) Failing after 1m7s
CI/CD / publish-deb (push) Successful in 39s
CI/CD / smoke (push) Successful in 8m25s
fix: correlate upstream discovery events to client requests in console
Fan-out discovery methods (tools/list, prompts/list, resources/list)
used synthetic request IDs that couldn't be looked up in the
correlation map. This caused upstream_response events to have no
correlationId, making the console unable to find upstream content
for replay ("No content to replay").

Fix: pass correlationId through RouteContext → discovery methods →
onUpstreamCall callback, so the handler can use it directly.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-10 15:21:05 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00
2026-02-21 03:10:39 +00:00

mcpctl

kubectl for MCP servers. A management system for Model Context Protocol servers — define, deploy, and connect MCP servers to Claude using familiar kubectl-style commands.

mcpctl get servers
NAME             TRANSPORT   REPLICAS   DOCKER IMAGE                              DESCRIPTION
grafana          STDIO       1          grafana/mcp-grafana:latest                Grafana MCP server
home-assistant   SSE         1          ghcr.io/homeassistant-ai/ha-mcp:latest    Home Assistant MCP
docmost          SSE         1          10.0.0.194:3012/michal/docmost-mcp:latest Docmost wiki MCP

What is this?

mcpctl manages MCP servers the same way kubectl manages Kubernetes pods. You define servers declaratively in YAML, group them into projects, and connect them to Claude Code or any MCP client through a local proxy.

The architecture:

Claude Code  <--STDIO-->  mcplocal (local proxy)  <--HTTP-->  mcpd (daemon)  <--Docker-->  MCP servers
  • mcpd — the daemon. Runs on a server, manages MCP server containers (Docker/Podman), stores configuration in PostgreSQL.
  • mcplocal — local proxy. Runs on your machine, presents a single MCP endpoint to Claude that merges tools from all your servers. Handles namespacing (grafana/search_dashboards), plugin execution (gating, content pipelines), and prompt delivery.
  • mcpctl — the CLI. Talks to mcpd (via mcplocal or directly) to manage everything.

Quick Start

1. Install

# From RPM repository (Fedora/RHEL)
sudo tee /etc/yum.repos.d/mcpctl.repo <<'EOF'
[mcpctl]
name=mcpctl
baseurl=https://mysources.co.uk/api/packages/michal/rpm
enabled=1
gpgcheck=0
EOF
sudo dnf install mcpctl

# Or build from source
git clone https://mysources.co.uk/michal/mcpctl.git
cd mcpctl
pnpm install
pnpm build
pnpm rpm:build   # requires bun and nfpm

2. Connect to a daemon

# Login to an mcpd instance
mcpctl login --mcpd-url http://your-server:3000

# Check connectivity
mcpctl status

3. Create your first secret

Secrets store credentials that servers need — API tokens, passwords, etc.

mcpctl create secret grafana-creds \
  --data GRAFANA_URL=http://grafana.local:3000 \
  --data GRAFANA_SERVICE_ACCOUNT_TOKEN=glsa_xxxxxxxxxxxx

4. Create your first server

Browse available templates, then create a server from one:

mcpctl get templates                  # List available server blueprints
mcpctl describe template grafana      # See required env vars, health checks, etc.

mcpctl create server my-grafana \
  --from-template grafana \
  --env-from-secret grafana-creds

mcpd pulls the image, starts a container, and keeps it running. Check on it:

mcpctl get instances            # See running containers
mcpctl logs my-grafana          # View server logs
mcpctl describe server my-grafana  # Full details

5. Create a project

A project groups servers together and configures how Claude interacts with them.

mcpctl create project monitoring \
  --description "Grafana dashboards and alerting" \
  --server my-grafana \
  --proxy-model content-pipeline

6. Connect Claude Code

Generate the .mcp.json config for Claude Code:

mcpctl config claude --project monitoring

This writes a .mcp.json that tells Claude Code to connect through mcplocal. Restart Claude Code and your Grafana tools appear:

mcpctl console monitoring   # Preview what Claude sees

Declarative Configuration

Everything can be defined in YAML and applied with mcpctl apply:

# infrastructure.yaml
secrets:
  - name: grafana-creds
    data:
      GRAFANA_URL: "http://grafana.local:3000"
      GRAFANA_SERVICE_ACCOUNT_TOKEN: "glsa_xxxxxxxxxxxx"

servers:
  - name: my-grafana
    description: "Grafana dashboards and alerting"
    fromTemplate: grafana
    envFrom:
      - secretRef:
          name: grafana-creds

projects:
  - name: monitoring
    description: "Infrastructure monitoring"
    proxyModel: content-pipeline
    servers:
      - my-grafana
mcpctl apply -f infrastructure.yaml

Round-trip works too — export, edit, re-apply:

mcpctl get all --project monitoring -o yaml > state.yaml
# edit state.yaml...
mcpctl apply -f state.yaml

Plugin System (ProxyModel)

ProxyModel is mcpctl's plugin system. Each project is assigned a plugin that controls how Claude interacts with its servers.

There are two layers:

  • Plugins — TypeScript hooks that intercept MCP requests/responses (gating, tool filtering, etc.)
  • Pipelines — YAML-defined content transformation stages (pagination, summarization, etc.)

Built-in Plugins

Plugins compose through inheritance. A plugin can extend another plugin and inherit all its hooks:

gate                   → gating only (begin_session + prompt delivery)
content-pipeline       → content transformation only (pagination, section-split)
default                → extends both gate AND content-pipeline (inherits all hooks from both)
Plugin Gating Content pipeline Description
gate Yes No begin_session gate with prompt delivery
content-pipeline No Yes Content transformation (paginate, section-split)
default Yes Yes Extends both — gate + content pipeline combined

The default plugin doesn't reimplement anything — it inherits the gating hooks from gate and the content hooks from content-pipeline. Custom plugins can extend built-in ones the same way.

Gating means Claude initially sees only a begin_session tool. After calling it with a task description, relevant prompts are delivered and the full tool list is revealed. This keeps Claude's context focused.

# Gated with content pipeline (default — extends gate + content-pipeline)
mcpctl create project home --server my-ha --proxy-model default

# Ungated, content pipeline only
mcpctl create project tools --server my-grafana --proxy-model content-pipeline

# Gated only, no content transformation
mcpctl create project docs --server my-docs --proxy-model gate

Plugin Hooks

Plugins intercept MCP requests/responses at specific lifecycle points. When a plugin extends another, it inherits all the parent's hooks. If both parent and child define the same hook, the child's version wins.

Hook When it fires
onSessionCreate New MCP session established
onSessionDestroy Session ends
onInitialize MCP initialize request — can inject instructions
onToolsList tools/list — can filter/modify tool list
onToolCallBefore Before forwarding a tool call — can intercept
onToolCallAfter After receiving tool result — can transform
onResourcesList resources/list — can filter resources
onResourceRead resources/read — can intercept resource reads
onPromptsList prompts/list — can filter prompts
onPromptGet prompts/get — can intercept prompt reads

When multiple parents define the same hook, lifecycle hooks (onSessionCreate, onSessionDestroy) chain sequentially. All other hooks require the child to override — otherwise it's a conflict error.

Content Pipelines

Content pipelines transform tool results through ordered stages before delivering to Claude:

Pipeline Stages Use case
default passthroughpaginate (8KB pages) Safe pass-through with pagination for large responses
subindex section-splitsummarize-tree Splits large content into sections, returns a summary index

How subindex Works

  1. Upstream returns a large tool result (e.g., 50KB of device states)
  2. section-split divides content into logical sections (2KB-15KB each)
  3. summarize-tree generates a compact index with section summaries (~200 tokens each)
  4. Client receives the index and can request specific sections via _section parameter

Configuration

Set per-project:

kind: project
name: home-automation
proxyModel: default
servers:
  - home-assistant
  - node-red

Via CLI:

mcpctl create project monitoring --server grafana --proxy-model content-pipeline

Custom ProxyModels

Place YAML files in ~/.mcpctl/proxymodels/ to define custom pipelines:

kind: ProxyModel
metadata:
  name: my-pipeline
spec:
  stages:
    - type: section-split
      config:
        minSectionSize: 1000
        maxSectionSize: 10000
    - type: summarize-tree
      config:
        maxTokens: 150
        maxDepth: 2
  appliesTo: [toolResult, prompt]
  cacheable: true

Inspect available plugins and pipelines:

mcpctl get proxymodels              # List all plugins and pipelines
mcpctl describe proxymodel default  # Pipeline details (stages, controller)
mcpctl describe proxymodel gate     # Plugin details (hooks, extends)

Custom Stages

Drop .js or .mjs files in ~/.mcpctl/stages/ to add custom transformation stages. Each file must export default an async function matching the StageHandler contract:

// ~/.mcpctl/stages/redact-keys.js
export default async function(content, ctx) {
  // ctx provides: contentType, sourceName, projectName, sessionId,
  //               originalContent, llm, cache, log, config
  const redacted = content.replace(/([A-Z_]+_KEY)=\S+/g, '$1=***');
  ctx.log.info(`Redacted ${content.length - redacted.length} chars of secrets`);
  return { content: redacted };
}

Stages loaded from disk appear as local source. Use them in a custom ProxyModel YAML:

kind: ProxyModel
metadata:
  name: secure-pipeline
spec:
  stages:
    - type: redact-keys        # matches filename without extension
    - type: section-split
    - type: summarize-tree

Stage contract reference:

Field Type Description
content string Input content (from previous stage or raw upstream)
ctx.contentType 'toolResult' | 'prompt' | 'resource' What kind of content is being processed
ctx.sourceName string Tool name, prompt name, or resource URI
ctx.originalContent string The unmodified content before any stage ran
ctx.llm LLMProvider Call ctx.llm.complete(prompt) for LLM summarization
ctx.cache CacheProvider Call ctx.cache.getOrCompute(key, fn) to cache expensive results
ctx.log StageLogger debug(), info(), warn(), error()
ctx.config Record<string, unknown> Config values from the ProxyModel YAML

Return value:

{ content: string; sections?: Section[]; metadata?: Record<string, unknown> }

If sections is returned, the framework stores them and presents a table of contents to the client. The client can drill into individual sections via _resultId + _section parameters on subsequent tool or prompt calls.

Section Drill-Down

When a stage (like section-split) produces sections, the pipeline automatically:

  1. Replaces the full content with a compact table of contents
  2. Appends a _resultId for subsequent drill-down
  3. Stores the full sections in memory (5-minute TTL)

Claude then calls the same tool (or prompts/get) again with _resultId and _section parameters to retrieve a specific section. This works for both tool results and prompt responses.

# What Claude sees (tool result):
3 sections (json):
[users] Users (4K chars)
[config] Config (1K chars)
[logs] Logs (8K chars)

_resultId: pm-abc123 — use _resultId and _section parameters to drill into a section.

# Claude drills down:
→ tools/call: grafana/query  { _resultId: "pm-abc123", _section: "logs" }
← [full 8K content of the logs section]

Hot-Reload

Stages and ProxyModels reload automatically when files change — no restart needed.

  • Stages (~/.mcpctl/stages/*.js): File watcher with 300ms debounce. Add, edit, or remove stage files and they take effect on the next tool call.
  • ProxyModels (~/.mcpctl/proxymodels/*.yaml): Re-read from disk on every request, so changes are always picked up.

Force a manual reload via the HTTP API:

curl -X POST http://localhost:3200/proxymodels/reload
# {"loaded": 3}

curl http://localhost:3200/proxymodels/stages
# [{"name":"passthrough","source":"built-in"},{"name":"redact-keys","source":"local"},...]

Built-in Stages Reference

Stage Description Key Config
passthrough Returns content unchanged
paginate Splits large content into numbered pages pageSize (default: 8000 chars)
section-split Splits content into named sections by structure (headers, JSON keys, code boundaries) minSectionSize (500), maxSectionSize (15000)
summarize-tree Generates LLM summaries for each section maxTokens (200), maxDepth (2)

section-split detects content type automatically:

Content Type Split Strategy
JSON array One section per array element, using name/id/label as section ID
JSON object One section per top-level key
YAML One section per top-level key
Markdown One section per ## header
Code One section per function/class boundary
XML One section per top-level element

Pause Queue (Model Studio)

The pause queue lets you intercept pipeline results in real-time — inspect what the pipeline produced, edit it, or drop it before Claude receives the response.

# Enable pause mode
curl -X PUT http://localhost:3200/pause -d '{"paused":true}'

# View queued items (blocked tool calls waiting for your decision)
curl http://localhost:3200/pause/queue

# Release an item (send transformed content to Claude)
curl -X POST http://localhost:3200/pause/queue/<id>/release

# Edit and release (send your modified content instead)
curl -X POST http://localhost:3200/pause/queue/<id>/edit -d '{"content":"modified content"}'

# Drop an item (send empty response)
curl -X POST http://localhost:3200/pause/queue/<id>/drop

# Release all queued items at once
curl -X POST http://localhost:3200/pause/release-all

# Disable pause mode
curl -X PUT http://localhost:3200/pause -d '{"paused":false}'

The pause queue is also available as MCP tools via mcpctl console --stdin-mcp, which gives Claude direct access to pause, get_pause_queue, and release_paused tools for self-monitoring.

LLM Providers

ProxyModel stages that need LLM capabilities (like summarize-tree) use configurable providers. Configure in ~/.mcpctl/config.yaml:

llm:
  - name: vllm-local
    type: openai-compatible
    baseUrl: http://localhost:8000/v1
    model: Qwen/Qwen3-32B
  - name: anthropic
    type: anthropic
    model: claude-sonnet-4-20250514
    # API key from: mcpctl create secret llm-keys --data ANTHROPIC_API_KEY=sk-...

Providers support tiered routing (fast for quick summaries, heavy for complex analysis) and automatic failover — if one provider is down, the next is tried.

# Check active providers
mcpctl status   # Shows LLM provider status

# View provider details
curl http://localhost:3200/llm/providers

Pipeline Cache

ProxyModel pipelines cache LLM-generated results (summaries, section indexes) to avoid redundant API calls. The cache is persistent across mcplocal restarts.

Namespace Isolation

Each combination of LLM provider + model + ProxyModel gets its own cache namespace:

~/.mcpctl/cache/openai--gpt-4o--content-pipeline/
~/.mcpctl/cache/anthropic--claude-sonnet-4-20250514--content-pipeline/
~/.mcpctl/cache/vllm--qwen-72b--subindex/

Switching LLM providers or models automatically uses a fresh cache — no stale results from a different model.

CLI Management

# View cache statistics (per-namespace breakdown)
mcpctl cache stats

# Clear all cache entries
mcpctl cache clear

# Clear a specific namespace
mcpctl cache clear openai--gpt-4o--content-pipeline

# Clear entries older than 7 days
mcpctl cache clear --older-than 7

Size Limits

The cache enforces a configurable maximum size (default: 256MB). When exceeded, the oldest entries are evicted (LRU). Entries older than 30 days are automatically expired.

Size can be specified as bytes, human-readable units, or a percentage of the filesystem:

new FileCache('ns', { maxSize: '512MB' })   // fixed size
new FileCache('ns', { maxSize: '1.5GB' })   // fractional units
new FileCache('ns', { maxSize: '10%' })     // 10% of partition

Resources

Resource What it is Example
server MCP server definition Docker image + transport + env vars
instance Running container (immutable) Auto-created from server replicas
secret Key-value credentials API tokens, passwords
template Reusable server blueprint Community server configs
project Workspace grouping servers "monitoring", "home-automation"
prompt Curated content for Claude Instructions, docs, guides
promptrequest Pending prompt proposal LLM-submitted, needs approval
rbac Access control bindings Who can do what
serverattachment Server-to-project link Virtual resource for apply

Commands

# List resources
mcpctl get servers
mcpctl get instances
mcpctl get projects
mcpctl get prompts --project myproject

# Detailed view
mcpctl describe server grafana
mcpctl describe project monitoring

# Create resources
mcpctl create server <name> [flags]
mcpctl create secret <name> --data KEY=value
mcpctl create project <name> --server <srv> [--proxy-model <plugin>]
mcpctl create prompt <name> --project <proj> --content "..."

# Modify resources
mcpctl edit server grafana          # Opens in $EDITOR
mcpctl patch project myproj proxyModel=default
mcpctl apply -f config.yaml         # Declarative create/update

# Delete resources
mcpctl delete server grafana

# Logs and debugging
mcpctl logs grafana                 # Container logs
mcpctl console monitoring           # Interactive MCP console
mcpctl console --inspect            # Traffic inspector
mcpctl console --audit              # Audit event timeline
mcpctl console --stdin-mcp          # Claude monitor (MCP tools for Claude)

# Backup (git-based)
mcpctl backup                       # Status and SSH key
mcpctl backup log                   # Commit history
mcpctl backup restore list          # Available restore points
mcpctl backup restore diff abc1234  # Preview a restore
mcpctl backup restore to abc1234 --force  # Restore to a commit

# Project management
mcpctl --project monitoring get servers       # Project-scoped listing
mcpctl --project monitoring attach-server grafana
mcpctl --project monitoring detach-server grafana

Templates

Templates are reusable server configurations. Create a server from a template without repeating all the config:

# Register a template
mcpctl create template home-assistant \
  --docker-image "ghcr.io/homeassistant-ai/ha-mcp:latest" \
  --transport SSE \
  --container-port 8086

# Create a server from it
mcpctl create server my-ha \
  --from-template home-assistant \
  --env-from-secret ha-secrets

Gated Sessions

Projects using the default or gate plugin are gated. When Claude connects to a gated project:

  1. Claude sees only a begin_session tool initially
  2. Claude calls begin_session with a description of its task
  3. mcplocal matches relevant prompts and delivers them
  4. The full tool list is revealed

This keeps Claude's context focused — instead of dumping 100+ tools and pages of docs upfront, only the relevant ones are delivered based on the task at hand.

# Gated (default)
mcpctl create project monitoring --server grafana --proxy-model default

# Ungated (direct tool access)
mcpctl create project tools --server grafana --proxy-model content-pipeline

Prompts

Prompts are curated content delivered to Claude through the MCP protocol. They can be plain text or linked to external MCP resources (like wiki pages).

# Create a text prompt
mcpctl create prompt deployment-guide \
  --project monitoring \
  --content-file docs/deployment.md \
  --priority 7

# Create a linked prompt (content fetched live from an MCP resource)
mcpctl create prompt wiki-page \
  --project monitoring \
  --link "monitoring/docmost:docmost://pages/abc123" \
  --priority 5

Claude can also propose prompts during a session. These appear as prompt requests that you can review and approve:

mcpctl get promptrequests
mcpctl approve promptrequest proposed-guide

Interactive Console

The console lets you see exactly what Claude sees — tools, resources, prompts — and call tools interactively:

mcpctl console monitoring

The traffic inspector watches MCP traffic from other clients in real-time:

mcpctl console --inspect

Claude Monitor (stdin-mcp)

Connect Claude itself as a monitor via the inspect MCP server:

mcpctl console --stdin-mcp

This exposes MCP tools that let Claude observe and control traffic:

Tool Description
list_models List configured LLM providers and their status
list_stages List all available pipeline stages (built-in + custom)
switch_model Change the active LLM provider for pipeline stages
get_model_info Get details about a specific LLM provider
reload_stages Force reload custom stages from disk
pause Toggle pause mode (intercept pipeline results)
get_pause_queue List items held in the pause queue
release_paused Release, edit, or drop a paused item

Architecture

┌──────────────┐           ┌─────────────────────────────────────────┐
│ Claude Code  │   STDIO   │           mcplocal (proxy)              │
│              │◄─────────►│                                         │
│ (or any MCP  │           │  Namespace-merging MCP proxy            │
│  client)     │           │  Gated sessions + prompt delivery       │
│              │           │  Per-project endpoints                  │
└──────────────┘           │  Traffic inspection                     │
                           └──────────────┬──────────────────────────┘
                                          │ HTTP (REST + MCP proxy)
                                          │
                           ┌──────────────┴──────────────────────────┐
                           │              mcpd (daemon)              │
                           │                                         │
                           │  REST API (/api/v1/*)                   │
                           │  MCP proxy (routes tool calls)          │
                           │  PostgreSQL (Prisma ORM)                │
                           │  Docker/Podman container management     │
                           │  Health probes (STDIO, SSE, HTTP)       │
                           │  RBAC enforcement                       │
                           │                                         │
                           │  ┌───────────────────────────────────┐  │
                           │  │     MCP Server Containers         │  │
                           │  │                                   │  │
                           │  │  grafana/  home-assistant/  ...   │  │
                           │  │  (managed + proxied by mcpd)      │  │
                           │  └───────────────────────────────────┘  │
                           └─────────────────────────────────────────┘

Clients never connect to MCP server containers directly — all tool calls go through mcplocal → mcpd, which proxies them to the right container via STDIO/SSE/HTTP. This keeps containers unexposed and lets mcpd enforce RBAC and health checks.

Tool namespacing: When Claude connects to a project with servers grafana and slack, it sees tools like grafana/search_dashboards and slack/send_message. mcplocal routes each call through mcpd to the correct upstream server.

Project Structure

mcpctl/
├── src/
│   ├── cli/          # mcpctl command-line interface (Commander.js)
│   ├── mcpd/         # Daemon server (Fastify 5, REST API)
│   ├── mcplocal/     # Local MCP proxy (namespace merging, gating)
│   ├── db/           # Database schema (Prisma) and migrations
│   └── shared/       # Shared types and utilities
├── deploy/           # Docker Compose for local development
├── stack/            # Production deployment (Portainer)
├── scripts/          # Build, release, and deploy scripts
├── examples/         # Example YAML configurations
└── completions/      # Shell completions (fish, bash)

Development

# Prerequisites: Node.js 20+, pnpm 9+, Docker/Podman

# Install dependencies
pnpm install

# Start local database
pnpm db:up

# Generate Prisma client
cd src/db && npx prisma generate && cd ../..

# Build all packages
pnpm build

# Run tests
pnpm test:run

# Development mode (mcpd with hot-reload)
cd src/mcpd && pnpm dev

License

MIT

Description
No description provided
Readme 4.7 MiB
Languages
TypeScript 96.4%
Shell 2.9%
JavaScript 0.7%