name: add-atomic-chat-tool description: Add Atomic Chat MCP server so the container agent can call local models served by the Atomic Chat desktop app via its OpenAI-compatible API.
Add Atomic Chat Integration
This skill adds a stdio-based MCP server that exposes models running in the local Atomic Chat desktop app as tools for the container agent. Claude remains the orchestrator but can offload work to local models served by Atomic Chat on http://127.0.0.1:1337/v1 (OpenAI-compatible).
Tools exposed:
atomic_chat_list_models— list models currently available in Atomic Chat (GET /v1/models)atomic_chat_generate— send a prompt to a specified model and return the response (POST /v1/chat/completions)
Model management (download, delete) is done through the Atomic Chat desktop UI — the app is a fork of Jan and manages its own model library.
The skill ships the MCP server source in this folder and copies it into the agent-runner tree at install time, then wires it up with small edits to index.ts, providers/claude.ts, and container-runner.ts. No branch merge — all edits are additive and idempotent.
Phase 1: Pre-flight
Check if already applied
Check if container/agent-runner/src/atomic-chat-mcp-stdio.ts exists. If it does, skip to Phase 3 (Configure).
Check prerequisites
Verify Atomic Chat is installed and its local API server is running. On the host:
curl -s http://127.0.0.1:1337/v1/models | head
If the request fails:
- Install Atomic Chat from the latest release (macOS only for now —
atomic-chat.dmg). - Open the app.
- Open Settings → Local API Server and make sure it's enabled on port
1337. - Go to the Hub (or Models) tab and download at least one model (e.g. Llama 3.2 3B, Qwen 2.5 Coder 7B).
- Load the model once by sending any message in Atomic Chat's UI to warm it up.
Phase 2: Apply Code Changes
Copy the MCP server source
cp .claude/skills/add-atomic-chat-tool/atomic-chat-mcp-stdio.ts container/agent-runner/src/atomic-chat-mcp-stdio.ts
Register the MCP server in the agent-runner
Edit container/agent-runner/src/index.ts. Find the mcpServers object that currently looks like this:
const mcpServers: Record<string, { command: string; args: string[]; env: Record<string, string> }> = {
nanoclaw: {
command: 'bun',
args: ['run', mcpServerPath],
env: {},
},
};
Add an atomic_chat entry alongside nanoclaw:
const mcpServers: Record<string, { command: string; args: string[]; env: Record<string, string> }> = {
nanoclaw: {
command: 'bun',
args: ['run', mcpServerPath],
env: {},
},
atomic_chat: {
command: 'bun',
args: ['run', path.join(__dirname, 'atomic-chat-mcp-stdio.ts')],
env: {
...(process.env.ATOMIC_CHAT_HOST ? { ATOMIC_CHAT_HOST: process.env.ATOMIC_CHAT_HOST } : {}),
...(process.env.ATOMIC_CHAT_API_KEY ? { ATOMIC_CHAT_API_KEY: process.env.ATOMIC_CHAT_API_KEY } : {}),
},
},
};
Add the tool glob to the allowlist
Edit container/agent-runner/src/providers/claude.ts. Find 'mcp__nanoclaw__*', in the TOOL_ALLOWLIST array and add 'mcp__atomic_chat__*', on the following line:
'mcp__nanoclaw__*',
'mcp__atomic_chat__*',
];
Forward host env vars into the container
Edit src/container-runner.ts in buildContainerArgs. Find the TZ env line:
args.push('-e', `TZ=${TIMEZONE}`);
Add ATOMIC_CHAT forwarding right after it:
args.push('-e', `TZ=${TIMEZONE}`);
// Atomic Chat MCP tool: forward host overrides if set (default is host.docker.internal:1337).
if (process.env.ATOMIC_CHAT_HOST) {
args.push('-e', `ATOMIC_CHAT_HOST=${process.env.ATOMIC_CHAT_HOST}`);
}
if (process.env.ATOMIC_CHAT_API_KEY) {
args.push('-e', `ATOMIC_CHAT_API_KEY=${process.env.ATOMIC_CHAT_API_KEY}`);
}
Surface [ATOMIC] log lines at info level
In the same file, find the stderr logger:
container.stderr?.on('data', (data) => {
for (const line of data.toString().trim().split('\n')) {
if (line) log.debug(line, { container: agentGroup.folder });
}
});
Replace it with:
container.stderr?.on('data', (data) => {
for (const line of data.toString().trim().split('\n')) {
if (!line) continue;
if (line.includes('[ATOMIC]')) {
log.info(line, { container: agentGroup.folder });
} else {
log.debug(line, { container: agentGroup.folder });
}
}
});
Add env-var stubs to .env.example
Append to .env.example:
# Atomic Chat MCP tool (.claude/skills/add-atomic-chat-tool)
# Override the host where Atomic Chat exposes its OpenAI-compatible API.
# Default: http://host.docker.internal:1337 (with fallback to localhost)
# ATOMIC_CHAT_HOST=http://host.docker.internal:1337
# Optional API key. Leave unset for a local Atomic Chat install — it does not require auth.
# ATOMIC_CHAT_API_KEY=
Validate code changes
pnpm run build
pnpm exec tsc -p container/agent-runner/tsconfig.json --noEmit
./container/build.sh
All three must be clean before proceeding.
Phase 3: Configure
Set Atomic Chat host (optional)
By default, the MCP server connects to http://host.docker.internal:1337 (Docker Desktop) with a fallback to localhost. To use a custom host, add to .env:
ATOMIC_CHAT_HOST=http://your-atomic-chat-host:1337
Set API key (optional)
Atomic Chat does not require authentication when running locally — leave this unset. Only set it if you've put Atomic Chat behind a reverse proxy that enforces auth:
ATOMIC_CHAT_API_KEY=sk-...
Restart the service
launchctl kickstart -k gui/$(id -u)/com.nanoclaw # macOS
# Linux: systemctl --user restart nanoclaw
Phase 4: Verify
Test inference
Tell the user:
Send a message like: "use atomic chat to tell me the capital of France"
The agent should use
atomic_chat_list_modelsto find available models, thenatomic_chat_generateto get a response.
Check logs if needed
tail -f logs/nanoclaw.log | grep -i atomic
Look for:
[ATOMIC] Listing models...— list request started[ATOMIC] Found N models— models discovered[ATOMIC] >>> Generating with <model>— generation started[ATOMIC] <<< Done: <model> | Xs | N tokens | M chars— generation completed
Troubleshooting
Agent says "Atomic Chat is not installed" or tries to run a CLI
The agent is looking for a CLI that doesn't exist instead of using the MCP tools. This means:
- The MCP server wasn't copied — check
container/agent-runner/src/atomic-chat-mcp-stdio.tsexists - The MCP server wasn't registered — check
container/agent-runner/src/index.tshas theatomic_chatentry inmcpServers - The allowlist wasn't updated — check
container/agent-runner/src/providers/claude.tsincludesmcp__atomic_chat__*inTOOL_ALLOWLIST - The container wasn't rebuilt — run
./container/build.sh
"Failed to connect to Atomic Chat"
- Verify the host API is reachable:
curl http://127.0.0.1:1337/v1/models - Confirm the Local API Server is enabled in Atomic Chat's settings
- Check Docker can reach the host:
docker run --rm curlimages/curl curl -s http://host.docker.internal:1337/v1/models - If using a custom host, check
ATOMIC_CHAT_HOSTin.env
model not found / 404 on generate
The model ID passed to atomic_chat_generate must exactly match one of the IDs returned by atomic_chat_list_models. Ask the agent to list models first, then pick one from that list.
Slow first response
Atomic Chat lazy-loads models into memory on first use. The initial call may take longer while the model warms up. Subsequent calls against the same model are fast.
Agent doesn't use Atomic Chat tools
The agent may not know about the tools. Try being explicit: "use the atomic_chat_generate tool with llama3.2-3b-instruct to answer: ..."
Context window or output size issues
Atomic Chat respects each model's native context length. If you hit limits, pass max_tokens explicitly when calling atomic_chat_generate, or switch to a model with a larger context window in the Atomic Chat UI.