Stop paying 30,000 tokens per query. This bridge implements Anthropic's discovery pattern with rootless security—reducing MCP context from 30K to 200 tokens while proxying any stdio server.
This bridge implements the "Code Execution with MCP" pattern, a convergence of ideas from industry leaders:
- Apple's CodeAct: "Your LLM Agent Acts Better when Generating Code."
- Anthropic's Code execution with MCP: "Building more efficient agents."
- Cloudflare's Code Mode: "LLMs are better at writing code to call MCP, than at calling MCP directly."
- Docker's Dynamic MCPs: "Stop Hardcoding Your Agents’ World."
Instead of exposing hundreds of individual tools to the LLM (which consumes massive context and confuses the model), this bridge exposes one tool: run_python. The LLM writes Python code to discover, call, and compose other tools.
While there are JavaScript-based alternatives (like universal-tool-calling-protocol/code-mode), this project is built for Data Science and Security:
| Feature | This Project (Python) | JS Code Mode (Node.js) |
|---|---|---|
| Native Language | Python (The language of AI/ML) | TypeScript/JavaScript |
| Data Science | Native (pandas, numpy, scikit-learn) |
Impossible / Hacky |
| Isolation | Hard (Podman/Docker Containers) | Soft (Node.js VM) |
| Security | Enterprise (Rootless, No Net, Read-Only) | Process-level |
| Philosophy | Infrastructure (Standalone Bridge) | Library (Embeddable) |
Choose this if: You want your agent to analyze data, generate charts, use scientific libraries, or if you require strict container-based isolation for running untrusted code.
Connect Claude to 11 MCP servers with ~100 tools = 30,000 tokens of tool schemas loaded into every prompt. That's $0.09 per query before you ask a single question. Scale to 50 servers and your context window breaks.
- Docker MCP Gateway: Manages containers beautifully, but still streams all tool schemas into Claude's context. No token optimization.
- Cloudflare Code Mode: V8 isolates are fast, but you can't proxy your existing MCP servers (Serena, Wolfram, custom tools). Platform lock-in.
- Academic Papers: Describe Anthropic's discovery pattern, but provide no hardened implementation.
- Proofs of Concept: Skip security (no rootless), skip persistence (cold starts), skip proxying edge cases.
- Constant 200-token overhead regardless of server count
- Proxy any stdio MCP server into rootless containers
- Fuzzy search across servers without preloading schemas
- Production-hardened with capability dropping and security isolation
Traditional MCP (Context-Bound)
┌─────────────────────────────┐
│ LLM Context (30K tokens) │
│ - serverA.tool1: {...} │
│ - serverA.tool2: {...} │
│ - serverB.tool1: {...} │
│ - … (dozens more) │
└─────────────────────────────┘
↓
LLM picks tool
↓
Tool executes
This Bridge (Discovery-First)
┌─────────────────────────────┐
│ LLM Context (≈200 tokens) │
│ “Use discovered_servers(), │
│ query_tool_docs(), │
│ search_tool_docs()” │
└─────────────────────────────┘
↓
LLM discovers servers
↓
LLM hydrates schemas
↓
LLM writes Python
↓
Bridge proxies execution
Result: constant overhead. Whether you manage 10 or 1000 tools, the system prompt stays right-sized and schemas flow only when requested.
| Capability | Docker MCP Gateway | Cloudflare Code Mode | Research Patterns | This Bridge |
|---|---|---|---|---|
| Solves token bloat | ❌ Manual preload | ❌ Fixed catalog | ❌ Theory only | ✅ Discovery runtime |
| Universal MCP proxying | ✅ Containers | ❌ Not provided | ✅ Any stdio server | |
| Rootless security | ✅ V8 isolate | ❌ Not addressed | ✅ Cap-dropped sandbox | |
| Auto-discovery | ❌ N/A | ❌ Not implemented | ✅ 9 config paths | |
| Tool doc search | ❌ | ❌ | ✅ search_tool_docs() |
|
| Production hardening | ✅ Managed service | ❌ Prototype | ✅ Tested bridge |
-
Two-stage discovery –
discovered_servers()reveals what exists;query_tool_docs(name)loads only the schemas you need. -
Fuzzy search across servers – let the model find tools without memorising catalog names:
from mcp import runtime matches = await runtime.search_tool_docs("calendar events", limit=5) for hit in matches: print(hit["server"], hit["tool"], hit.get("description", ""))
-
Zero-copy proxying – every tool call stays within the sandbox, mirrored over stdio with strict timeouts.
-
Rootless by default – Podman/Docker containers run with
--cap-drop=ALL, read-only root, no-new-privileges, and explicit memory/PID caps. -
Compact + TOON output – minimal plain-text responses for most runs, with deterministic TOON blocks available via
MCP_BRIDGE_OUTPUT_MODE=toon.
- Teams juggling double-digit MCP servers who cannot afford context bloat.
- Agents that orchestrate loops, retries, and conditionals rather than single tool invocations.
- Security-conscious operators who need rootless isolation for LLM-generated code.
- Practitioners who want to reuse existing MCP catalogs without hand-curating manifests.
This server aligns with the philosophy that you might not need MCP at all for every little tool. Instead of building rigid MCP servers for simple tasks, you can use this server to give your agent raw, sandboxed access to Bash and Python.
- Ad-Hoc Tools: Need a script to scrape a site or parse a file? Just write it and run it. No need to deploy a new MCP server.
- Composability: Pipe outputs between commands, save intermediate results to files, and use standard Unix tools.
- Safety: Unlike giving an agent raw shell access to your machine, this server runs everything in a secure, rootless container. You get the power of "Bash/Code" without the risk.
- Lazy Runtime Detection: Starts up instantly even if Podman/Docker isn't ready. Checks for runtime only when code execution is requested.
- Self-Reference Prevention: Automatically detects and skips configurations that would launch the bridge recursively.
- Noise Filtering: Ignores benign JSON parse errors (like blank lines) from chatty MCP clients.
- Smart Volume Sharing: Probes Podman VMs to ensure volume sharing works, even on older versions.
- Rootless containers - No privileged helpers required
- Network isolation - No network access
- Read-only filesystem - Immutable root
- Dropped capabilities - No system access
- Unprivileged user - Runs as UID 65534
- Resource limits - Memory, PIDs, CPU, time
- Auto-cleanup - Temporary IPC directories
- Persistent clients - MCP servers stay warm
- Context efficiency - 95%+ reduction vs traditional MCP
- Async execution - Proper resource management
- Single tool - Only
run_pythonin Claude's context
-
Multiple access patterns:
mcp_servers["server"] # Dynamic lookup mcp_server_name # Attribute access from mcp.servers.server import * # Module import
-
Top-level await - Modern Python patterns
-
Type-safe - Proper signatures and docs
-
Compact responses - Plain-text output by default with optional TOON blocks when requested
- Default (compact) – responses render as plain text plus a minimal
structuredContentpayload containing only non-empty fields.stdout/stderrlines stay intact, so prompts remain lean without sacrificing content. - Optional TOON – set
MCP_BRIDGE_OUTPUT_MODE=toonto emit Token-Oriented Object Notation blocks. We still drop empty fields and mirror the same structure instructuredContent; TOON is handy when you want deterministic tokenisation for downstream prompts. - Fallback JSON – if the TOON encoder is unavailable we automatically fall back to pretty JSON blocks while preserving the trimmed payload.
SANDBOX_HELPERS_SUMMARYin the tool schema only advertises the discovery helpers (discovered_servers(),list_servers(),query_tool_docs(),search_tool_docs(), etc.). It never includes individual server or tool documentation.- On first use the LLM typically calls
discovered_servers()(orlist_servers_sync()for the cached list) to enumerate MCP servers, thenquery_tool_docs(server)/query_tool_docs_sync(server)orsearch_tool_docs("keyword")/search_tool_docs_sync("keyword")to fetch the relevant subset of documentation. - Tool metadata is streamed on demand, keeping the system prompt at roughly 200 tokens regardless of how many servers or tools are installed.
- Once the LLM has the docs it needs, it writes Python that uses the generated
mcp_<alias>proxies ormcp.runtimehelpers to invoke tools.
Need a short description without probing the helpers? Call runtime.capability_summary() to print a one-paragraph overview suitable for replying to questions such as “what can the code-execution MCP do?”
- Check version:
python3 --version - If needed, install Python 3.14 via package manager or python.org
- macOS:
brew install podmanorbrew install --cask docker - Ubuntu/Debian:
sudo apt-get install -y podmanorcurl -fsSL https://get.docker.com | sh
curl -LsSf https://astral.sh/uv/install.sh | shpodman pull python:3.14-slim
# or
docker pull python:3.14-slimNote on Pydantic compatibility (Python 3.14):
- If you use Python 3.14, ensure you have a modern Pydantic release installed (for example,
pydantic >= 2.12.0). Some older Pydantic versions or environments that install a separatetypingpackage from PyPI may raise errors such as:
TypeError: _eval_type() got an unexpected keyword argument 'prefer_fwd_module'
If you see this error, run:
pip install -U pydantic
pip uninstall typing # if present; the stdlib's typing should be usedAnd re-run the project setup (e.g. remove .venv/ and uv sync).
Use uv to sync the project environment:
uv syncuvx --from git+https://github.com/elusznik/mcp-server-code-execution-mode mcp-server-code-execution-mode runIf you prefer to run from a local checkout, the equivalent command is:
uv run python mcp_server_code_execution_mode.pyAdd the following server configuration to your agent's MCP settings file (e.g., mcp_config.json, claude_desktop_config.json, etc.):
{
"mcpServers": {
"mcp-server-code-execution-mode": {
"command": "uvx",
"args": [
"--from",
"git+https://github.com/elusznik/mcp-server-code-execution-mode",
"mcp-server-code-execution-mode",
"run"
],
"env": {
"MCP_BRIDGE_RUNTIME": "podman"
}
}
}
}# Use MCP tools in sandboxed code
result = await mcp_filesystem.read_file(path='/tmp/test.txt')
# Complex workflows
data = await mcp_search.search(query="TODO")
await mcp_github.create_issue(repo='owner/repo', title=data.title)run_python only loads the MCP servers you request. Pass them via the servers array when you invoke the tool so proxies such as mcp_serena or mcp_filesystem become available inside the sandbox:
{
"code": "print(await mcp_serena.search(query='latest AI papers'))",
"servers": ["serena", "filesystem"]
}If you omit the list the discovery helpers still enumerate everything, but any RPC call that targets an unloaded server returns Server '<name>' is not available.
Note: The servers array only controls which proxies are generated for a sandbox invocation. It does not set server configuration fields such as cwd. The cwd property is part of the host/server config and LLMs should call runtime.describe_server(name) or inspect runtime.list_loaded_server_metadata() to discover the configured cwd before assuming the server's working directory.
Note: server configurations can include an optional cwd property. If present the bridge will start the host MCP server process in that working directory; agents should check runtime.describe_server(name) to discover a server's configured cwd before making assumptions.
Project environments target CPython 3.14. Ensure your local environment uses Python 3.14+:
uv python pin 3.14
uv syncRuntime dependencies stay lean, so pytest is fetched on demand when needed:
uv run --with pytest pytestPrefer a persistent install? Add a dev extra and sync it once:
[project.optional-dependencies]
test = ["pytest>=9"]uv sync --group test
uv run pytest┌─────────────┐
│ MCP Client │ (Your Agent)
└──────┬──────┘
│ stdio
▼
┌──────────────┐
│ MCP Code Exec │ ← Discovers, proxies, manages
│ Bridge │
└──────┬──────┘
│ container
▼
┌─────────────┐
│ Container │ ← Executes with strict isolation
│ Sandbox │
└─────────────┘
Unlike traditional MCP servers that preload every tool definition (sometimes 30k+ tokens), this bridge pins its system prompt to roughly 200 tokens and trains the LLM to discover what it needs on demand:
- LLM calls
discovered_servers()→ learns which bridges are available without loading schemas. - LLM calls
query_tool_docs("serena")→ hydrates just that server's tool docs, optionally filtered per tool. - LLM writes orchestration code → invokes helpers like
mcp_serena.search()ormcp.runtime.call_tool().
Result: context usage stays effectively constant no matter how many MCP servers you configure.
Process:
- Client calls
run_python(code, servers, timeout) - Bridge loads requested MCP servers
- Prepares a sandbox invocation: collects MCP tool metadata, writes an entrypoint into a shared
/ipcvolume, and exportsMCP_AVAILABLE_SERVERS - Generated entrypoint rewires stdio into JSON-framed messages and proxies MCP calls over the container's stdin/stdout pipe
- Runs container with security constraints
- Host stream handler processes JSON frames, forwards MCP traffic, enforces timeouts, and cleans up
| Variable | Default | Description |
|---|---|---|
MCP_BRIDGE_RUNTIME |
auto | Container runtime (podman/docker) |
MCP_BRIDGE_IMAGE |
python:3.14-slim | Container image |
MCP_BRIDGE_TIMEOUT |
30s | Default timeout |
MCP_BRIDGE_MAX_TIMEOUT |
120s | Max timeout |
MCP_BRIDGE_MEMORY |
512m | Memory limit |
MCP_BRIDGE_PIDS |
128 | Process limit |
MCP_BRIDGE_CPUS |
- | CPU limit |
MCP_BRIDGE_CONTAINER_USER |
65534:65534 | Run as UID:GID |
MCP_BRIDGE_RUNTIME_IDLE_TIMEOUT |
300s | Shutdown delay |
MCP_BRIDGE_STATE_DIR |
~/MCPs |
Host directory for IPC sockets and temp state |
MCP_BRIDGE_OUTPUT_MODE |
compact |
Response text format (compact or toon) |
MCP_BRIDGE_LOG_LEVEL |
INFO |
Bridge logging verbosity |
Primary Location:
~/MCPs/*.json(Recommended)
Note: Support for scanning individual agent configuration files (e.g.,
.claude.json,.vscode/mcp.json) is currently postponed. Please place all your MCP server definitions .jsons in the~/MCPsdirectory to ensure they are discovered.
Example Server (~/MCPs/filesystem.json):
{
"mcpServers": {
"filesystem": {
"command": "npx",
"args": ["-y", "@modelcontextprotocol/server-filesystem", "/tmp"]
}
}
}
> **Note:** To prevent recursive launches, the bridge automatically skips any config entry that appears to start `mcp-server-code-execution-mode` again (including `uvx … mcp-server-code-execution-mode run`). Set `MCP_BRIDGE_ALLOW_SELF_SERVER=1` if you intentionally need to expose the bridge as a nested MCP server.When you rely on docker mcp gateway run to expose third-party MCP servers, the bridge simply executes the gateway binary. The gateway is responsible for pulling tool images and wiring stdio transports, so make sure the host environment is ready:
- Run
docker loginfor every registry referenced in the gateway catalog (e.g. Docker Hubmcp/*images,ghcr.io/github/github-mcp-server). Without cached credentials the pull step fails before any tools come online. - Provide required secrets for those servers—
github-officialneedsgithub.personal_access_token, others may expect API keys or auth tokens. Usedocker mcp secret set <name>(or whichever mechanism your gateway is configured with) so the container sees the values at start-up. - Mirror any volume mounts or environment variables that the catalog expects (filesystem paths, storage volumes, etc.). Missing mounts or credentials commonly surface as
failed to connect: calling "initialize": EOFduring the stdio handshake. - If
list_toolsonly returns the internal management helpers (mcp-add,code-mode, …), the gateway never finished initializing the external servers—check the gateway logs for missing secrets or registry access errors.
- Runtime artifacts (including the generated
/ipc/entrypoint.pyand related handshake metadata) live under~/MCPs/by default. SetMCP_BRIDGE_STATE_DIRto relocate them. - When the selected runtime is Podman, the bridge automatically issues
podman machine set --rootful --now --volume <state_dir>:<state_dir>so the VM can mount the directory. On olderpodman machinebuilds that do not support--volume, the bridge now probes the VM withpodman machine ssh test -d <state_dir>and proceeds if the share is already available. - Docker Desktop does not expose a CLI for file sharing; ensure the chosen state directory is marked as shared in Docker Desktop → Settings → Resources → File Sharing before running the bridge.
- To verify a share manually, run
docker run --rm -v ~/MCPs:/ipc alpine ls /ipc(or the Podman equivalent) and confirm the files are visible.
# List and filter files
files = await mcp_filesystem.list_directory(path='/tmp')
for file in files:
content = await mcp_filesystem.read_file(path=file)
if 'TODO' in content:
print(f"TODO in {file}")# Extract data
transcript = await mcp_google_drive.get_document(documentId='abc123')
# Process
summary = transcript[:500] + "..."
# Store
await mcp_salesforce.update_record(
objectType='SalesMeeting',
recordId='00Q5f000001abcXYZ',
data={'Notes': summary}
)# Jira → GitHub migration
issues = await mcp_jira.search_issues(project='API', status='Open')
for issue in issues:
details = await mcp_jira.get_issue(id=issue.id)
if 'bug' in details.description.lower():
await mcp_github.create_issue(
repo='owner/repo',
title=f"Bug: {issue.title}",
body=details.description
)from mcp import runtime
print("Discovered:", runtime.discovered_servers())
print("Cached servers:", runtime.list_servers_sync())
print("Loaded metadata:", runtime.list_loaded_server_metadata())
print("Selectable via RPC:", await runtime.list_servers())
# Peek at tool docs for a server that's already loaded in this run
loaded = runtime.list_loaded_server_metadata()
if loaded:
first = runtime.describe_server(loaded[0]["name"])
for tool in first["tools"]:
print(tool["alias"], "→", tool.get("description", ""))
# Ask for summaries or full schemas only when needed
if loaded:
summaries = await runtime.query_tool_docs(loaded[0]["name"])
detailed = await runtime.query_tool_docs(
loaded[0]["name"],
tool=summaries[0]["toolAlias"],
detail="full",
)
print("Summaries:", summaries)
print("Cached tools:", runtime.list_tools_sync(loaded[0]["name"]))
print("Detailed doc:", detailed)
# Fuzzy search across loaded servers without rehydrating every schema
results = await runtime.search_tool_docs("calendar events", limit=3)
for result in results:
print(result["server"], result["tool"], result.get("description", ""))
# Synchronous helpers for quick answers without extra awaits
print("Capability summary:", runtime.capability_summary())
print("Docs from cache:", runtime.query_tool_docs_sync(loaded[0]["name"]) if loaded else [])
print("Search from cache:", runtime.search_tool_docs_sync("calendar"))Example output seen by the LLM when running the snippet above with the stub server:
Discovered: ('stub',)
Loaded metadata: ({'name': 'stub', 'alias': 'stub', 'tools': [{'name': 'echo', 'alias': 'echo', 'description': 'Echo the provided message', 'input_schema': {...}}]},)
Selectable via RPC: ('stub',)
Clients that prefer listMcpResources can skip executing the helper snippet and instead request the
resource://mcp-server-code-execution-mode/capabilities resource. The server advertises it via
resources/list, and reading it returns the same helper summary plus a short checklist for loading
servers explicitly.
| Constraint | Setting | Purpose |
|---|---|---|
| Network | --network none |
No external access |
| Filesystem | --read-only |
Immutable base |
| Capabilities | --cap-drop ALL |
No system access |
| Privileges | no-new-privileges |
No escalation |
| User | 65534:65534 |
Unprivileged |
| Memory | --memory 512m |
Resource cap |
| PIDs | --pids-limit 128 |
Process cap |
| Workspace | tmpfs, noexec | Safe temp storage |
| Action | Allowed | Details |
|---|---|---|
| Import stdlib | ✅ | Python standard library |
| Access MCP tools | ✅ | Via proxies |
| Memory ops | ✅ | Process data |
| Write to disk | ✅ | Only /tmp, /workspace |
| Network | ❌ | Completely blocked |
| Host access | ❌ | No system calls |
| Privilege escalation | ❌ | Prevented by sandbox |
| Container escape | ❌ | Rootless + isolation |
- README.md - This file, quick start
- GUIDE.md - Comprehensive user guide
- ARCHITECTURE.md - Technical deep dive
- HISTORY.md - Evolution and lessons
- STATUS.md - Current state and roadmap
- Code Execution with MCP (Anthropic)
- Code Mode (Cloudflare)
- Model Context Protocol
- Dynamic MCPs with Docker
- CodeAct: Your LLM Agent Acts Better when Generating Code (Apple)
- Rootless container sandbox
- Single
run_pythontool - MCP server proxying
- Persistent clients
- Comprehensive docs
- Automated testing
- Observability (logging, metrics)
- Policy controls
- Runtime diagnostics
- Connection pooling
- Web UI
- Multi-language support
- Workflow orchestration
- Agent-visible discovery channel (host-proxied
mcp-find/mcp-add) - Execution telemetry (structured logs, metrics, traces)
- Persistent and shareable code-mode artifacts
GPLv3 License
For issues or questions, see the documentation or file an issue.
