Skip to content

HarnessLab/claw-code-agent

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

12 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Claw Code Agent logo

Claw Code Agent

A Python reimplementation of the Claude Code agent architecture β€” local models, full control.

Python 3.10+ GitHub vLLM Qwen3-Coder Alpha License


πŸ“– About

This repository reimplements the Claude Code npm agent architecture entirely in Python, designed to run with local open-source models via an OpenAI-compatible API server.

Built on the public porting workspace from instructkr/claw-code, the active development lives at HarnessLab/claw-code-agent.

Goal: Not to ship the original npm source, but to reimplement the full agent flow in Python β€” prompt assembly, context building, slash commands, tool calling, session persistence, and local model execution.

Claw Code Agent demo


✨ Key Features

Feature Description
πŸ€– Agent Loop Full agentic coding loop with tool calling and iterative reasoning
🧰 Core Tools File read / write / edit, glob search, grep search, shell execution
πŸ’¬ Slash Commands Local commands: /help, /context, /tools, /memory, /status, /model, and more
🧠 Context Engine Automatic context building with CLAUDE.md discovery and usage reporting
πŸ”„ Session Persistence Save and resume agent sessions across runs
πŸ” Permission System Granular control: --allow-write, --allow-shell, --unsafe
πŸ—οΈ OpenAI-Compatible Runtime Python client targets an OpenAI-compatible API, with vLLM as the documented setup
πŸ‰ Qwen3-Coder First-class support for Qwen3-Coder-30B-A3B-Instruct via vLLM

πŸ“‹ Roadmap

Done

  • Python CLI agent loop
  • OpenAI-compatible local model backend
  • Qwen3-Coder support through vLLM with qwen3_xml tool parser
  • Core tools: list_dir, read_file, write_file, edit_file, glob_search, grep_search, bash
  • Context building and /context-style usage reporting
  • Slash commands: /help, /context, /context-raw, /prompt, /permissions, /model, /tools, /memory, /status, /clear
  • Session persistence and agent-resume flow
  • Permission system (read-only, write, shell, unsafe tiers)
  • Unit tests for the Python runtime
  • pyproject.toml packaging with setuptools

In Progress

  • Full MCP support
  • Full plugin system
  • Full slash-command parity
  • Full interactive REPL / TUI behavior
  • Exact tokenizer / context accounting
  • Hooks parity
  • Remote modes parity
  • Voice / VIM parity
  • Some deeper runtime details from the npm source
  • Cost tracking and budget limits

πŸ—οΈ Architecture

claw-code/
β”œβ”€β”€ README.md
β”œβ”€β”€ pyproject.toml
β”œβ”€β”€ .gitignore
β”œβ”€β”€ images/
β”‚   └── logo.png
β”œβ”€β”€ src/                          # Python implementation
β”‚   β”œβ”€β”€ main.py                   # CLI entry point & argument parsing
β”‚   β”œβ”€β”€ agent_runtime.py          # Core agent loop (LocalCodingAgent)
β”‚   β”œβ”€β”€ agent_tools.py            # Tool definitions & execution engine
β”‚   β”œβ”€β”€ agent_prompting.py        # System prompt assembly
β”‚   β”œβ”€β”€ agent_context.py          # Context building & CLAUDE.md discovery
β”‚   β”œβ”€β”€ agent_context_usage.py    # Context usage estimation & reporting
β”‚   β”œβ”€β”€ agent_session.py          # Session state management
β”‚   β”œβ”€β”€ agent_slash_commands.py   # Local slash command processing
β”‚   β”œβ”€β”€ agent_types.py            # Shared dataclasses & type definitions
β”‚   β”œβ”€β”€ openai_compat.py          # OpenAI-compatible API client
β”‚   β”œβ”€β”€ session_store.py          # Session serialization & persistence
β”‚   β”œβ”€β”€ permissions.py            # Tool permission filtering
β”‚   β”œβ”€β”€ tools.py                  # Mirrored tool inventory
β”‚   β”œβ”€β”€ commands.py               # Mirrored command inventory
β”‚   β”œβ”€β”€ ...                       # 75+ modules across 30+ packages
β”‚   β”œβ”€β”€ plugins/                  # Plugin subsystem (WIP)
β”‚   β”œβ”€β”€ hooks/                    # Hook system (WIP)
β”‚   β”œβ”€β”€ remote/                   # Remote runtime modes (WIP)
β”‚   β”œβ”€β”€ voice/                    # Voice mode (WIP)
β”‚   └── vim/                      # VIM mode (WIP)
└── tests/                        # Unit tests
    β”œβ”€β”€ test_agent_runtime.py
    β”œβ”€β”€ test_agent_context.py
    β”œβ”€β”€ test_agent_context_usage.py
    β”œβ”€β”€ test_agent_prompting.py
    β”œβ”€β”€ test_agent_slash_commands.py
    └── test_porting_workspace.py

πŸ“¦ Requirements

Requirement Details
🐍 Python 3.10 or higher
πŸ–₯️ Model Server vLLM, Ollama, or LiteLLM Proxy, with tool calling support
🧠 Model Qwen/Qwen3-Coder-30B-A3B-Instruct (recommended)

πŸš€ Quick Start

1. Start vLLM with Qwen3-Coder

vLLM must be started with automatic tool choice enabled. Use the qwen3_xml parser for Qwen3-Coder tool calling:

python -m vllm.entrypoints.openai.api_server \
  --model Qwen/Qwen3-Coder-30B-A3B-Instruct \
  --host 127.0.0.1 \
  --port 8000 \
  --enable-auto-tool-choice \
  --tool-call-parser qwen3_xml

Verify the server is running:

curl http://127.0.0.1:8000/v1/models

πŸ“š References: vLLM Tool Calling Docs Β· OpenAI-Compatible Server

Optional: Use Ollama Instead of vLLM

claw-code-agent can also work with Ollama because the runtime targets an OpenAI-compatible API. Use a model that supports tool calling well.

Example:

ollama serve
ollama pull qwen3

Then configure:

export OPENAI_BASE_URL=http://127.0.0.1:11434/v1
export OPENAI_API_KEY=ollama
export OPENAI_MODEL=qwen3

Notes:

  • prefer tool-capable models such as qwen3
  • plain chat-only models are not enough for full agent behavior
  • Ollama does not use the vLLM parser flags shown above

πŸ“š References: Ollama OpenAI Compatibility Β· Ollama Tool Calling

Optional: Use LiteLLM Proxy

claw-code-agent can also work through LiteLLM Proxy because the runtime targets an OpenAI-compatible chat completions API. The routed model still needs to support tool calling for full agent behavior.

Quick start example:

pip install 'litellm[proxy]'
litellm --model ollama/qwen3

LiteLLM Proxy runs on port 4000 by default. Then configure:

export OPENAI_BASE_URL=http://127.0.0.1:4000
export OPENAI_API_KEY=anything
export OPENAI_MODEL=ollama/qwen3

Notes:

  • LiteLLM Proxy gives you an OpenAI-style gateway in front of many providers
  • tool use still depends on the underlying routed model and provider behavior
  • if you configure a LiteLLM master key, use that instead of anything

πŸ“š References: LiteLLM Docs Β· LiteLLM Proxy Quick Start

2. Configure Environment

export OPENAI_BASE_URL=http://127.0.0.1:8000/v1
export OPENAI_API_KEY=local-token
export OPENAI_MODEL=Qwen/Qwen3-Coder-30B-A3B-Instruct

Use Another Model With vLLM

If you want to try another model, keep the same vLLM server setup and change the --model value when you launch vLLM.

Example:

python -m vllm.entrypoints.openai.api_server \
  --model your-model-name \
  --host 127.0.0.1 \
  --port 8000 \
  --enable-auto-tool-choice \
  --tool-call-parser your_parser

Then update:

export OPENAI_MODEL=your-model-name

Notes:

  • the documented path in this repository is vLLM
  • the model must support tool calling well enough for agent use
  • some model families require a different --tool-call-parser
  • slash commands such as /help, /context, and /tools are local and do not require the model server

3. Run the Agent

# Read-only question
python3 -m src.main agent \
  "Read src/agent_runtime.py and summarize how the loop works." \
  --cwd .

# Write-enabled task
python3 -m src.main agent \
  "Create TEST_QWEN_AGENT.md with one line: test ok" \
  --cwd . --allow-write

# Shell-enabled task
python3 -m src.main agent \
  "Run pwd and ls src, then summarize the result." \
  --cwd . --allow-shell

πŸ› οΈ Usage

Agent Commands

Command Description
agent <prompt> Run the agent with a prompt
agent-prompt Show the assembled system prompt
agent-context Show estimated context usage
agent-context-raw Show the raw context snapshot
agent-resume <id> <prompt> Resume a saved session

CLI Flags

Flag Description
--cwd <path> Set the workspace directory
--model <name> Override the model name
--base-url <url> Override the API base URL
--allow-write Allow the agent to modify files
--allow-shell Allow the agent to execute shell commands
--unsafe Allow destructive shell operations
--show-transcript Print the full message transcript
--system-prompt <text> Set a custom system prompt
--append-system-prompt <text> Append to the system prompt
--add-dir <path> Add extra directories to context

Slash Commands

These are handled locally before the model loop:

Command Aliases Description
/help /commands Show built-in slash commands
/context /usage Show estimated session context usage
/context-raw /env Show raw environment & context snapshot
/prompt /system-prompt Render the effective system prompt
/permissions β€” Show active tool permission mode
/model β€” Show or update the active model
/tools β€” List registered tools with permission status
/memory β€” Show loaded CLAUDE.md memory bundle
/status /session Show runtime/session status summary
/clear β€” Clear ephemeral runtime state
python3 -m src.main agent "/help"
python3 -m src.main agent "/context" --cwd .
python3 -m src.main agent "/tools" --cwd .
python3 -m src.main agent "/status" --cwd .

Utility Commands

python3 -m src.main summary            # Workspace summary
python3 -m src.main manifest           # Workspace manifest
python3 -m src.main commands --limit 10 # Command inventory
python3 -m src.main tools --limit 10    # Tool inventory

πŸ”§ Built-in Tools

The agent has access to 7 core tools:

Tool Description Permission
list_dir List files and directories 🟒 Always
read_file Read file contents (with line ranges) 🟒 Always
write_file Write or create files 🟑 --allow-write
edit_file Edit files via exact string matching 🟑 --allow-write
glob_search Find files by glob pattern 🟒 Always
grep_search Search file contents by regex 🟒 Always
bash Execute shell commands πŸ”΄ --allow-shell

πŸ”„ Session Persistence

Each agent run automatically saves a resumable session:

session_id=4f2c8c6f9c0e4d7c9c7b1b2a3d4e5f67
session_path=.port_sessions/agent/4f2c8c6f...

Resume a previous session:

python3 -m src.main agent-resume \
  4f2c8c6f9c0e4d7c9c7b1b2a3d4e5f67 \
  "Continue the previous task and finish the missing parts."

Inspect saved sessions:

ls -lt .port_sessions/agent

Note: Run agent-resume from the same claw-code/ directory where the session was created. A resumed session continues from the saved transcript, not from scratch.


πŸ§ͺ Testing

Run the full test suite:

python3 -m unittest discover -s tests -v

Smoke tests:

python3 -m src.main agent "/help"
python3 -m src.main agent-context --cwd .
python3 -m src.main agent \
  "Read src/agent_session.py and summarize the message flow." \
  --cwd .

πŸ” Permission Model

Claw Code Agent uses a tiered permission system to keep the agent safe by default:

Tier Capability Flag Required
Read-only List, read, glob, grep None (default)
Write + file creation and editing --allow-write
Shell + shell command execution --allow-shell
Unsafe + destructive shell operations --unsafe

⚠️ Disclaimer

  • This repository is a Python reimplementation inspired by the Claude Code npm architecture.
  • It does not ship the original npm source.
  • It is not affiliated with or endorsed by Anthropic.

Built with 🐍 Python Β· Powered by πŸ‰ HarnessLab Team.

About

Claw Code No Rust No TypeScript Only Python. Easy to work with. Fast to iterate. πŸ”₯ Zero external dependencies πŸ”₯

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages