Meta Multi-Agent Coding OS
by DSRPT.AI
Claude Code CLI, but more feature-rich.
Aider + LangGraph + Ollama + Voice + Browser UI
Fully branded Pinokio v7 Terminal Plugin
DSRPTV is a production-grade meta multi-agent coding system designed to function like Claude Code CLI but with significantly more features:
- 🧠 LangGraph Multi-Agent Core: Supervisor → Coder (Aider) → Tester → Loop with persistent SQLite checkpointing
- 🏛️ Architect Mode: Two-model planning + execution (plan with primary, edit with editor model)
- 🎙️ Voice Input: Local Whisper integration for voice-to-code
- 🌐 Web UI: Aider browser interface (always enabled by default)
- 📊 Auto-Test + Lint: Pytest/npm test + ruff/eslint loops
- 💾 Persistent Memory: Chat history + checkpoint recovery
- 🤖 Ollama Model Picker: Interactive menu to select primary + editor models from your local/remote Ollama instance
- 🚀 Windows Compatible: Direct Windows bash path support via Pinokio kernel
- Open Pinokio
- Go to Plugins → Install
- Enter:
https://github.com/DSRPT/dsrptv - Select your target project directory
- DSRPTV will auto-install dependencies and launch
git clone https://github.com/DSRPT/dsrptv.git
cd dsrptv
pip install aider-chat langgraph langchain-community langchain-ollama
# Interactive menu launcher
node menu.js
# OR direct LangGraph meta-agent
python graph.py
# OR direct aider launch
bash launch.shAfter installation, DSRPTV appears in your Pinokio Plugins list:
- Click DSRPTV → Select target project
- Choose environment variables (prompted on first run):
OLLAMA_API_BASE: Local (http://localhost:11434) or remote Ollama serverDSRPTV_MODEL: Primary model (e.g.,ollama/qwen2.5-coder:32b)DSRPTV_EDITOR_MODEL: Editor model (e.g.,ollama/qwen2.5-coder:14b)DSRPTV_MODE:architect|agent|code|ask|voice|browser
- DSRPTV launches directly in your project with Aider Architect mode
Run node menu.js for a full interactive setup:
- Ollama Host: Configure local or remote server
- Model Selection: Live-fetch models from Ollama or use recommended list
- Primary + Editor Models: Pick separate models or use same for both
- Agent Mode:
- Meta Multi-Agent — LangGraph: Supervisor + Coder + Tester (durable checkpointing)
- Native Agent — aider
--yes --mapautonomous mode - Architect — Two-model plan + execute (default)
- Voice-to-Code — Whisper input enabled
- Browser Web UI — Aider in browser
- Code Mode — Fast single-model editing
- Ask Mode — Questions only
- Extras: Voice, persistent memory, auto-test, auto-lint, web URL ingestion
Run python graph.py for the Supervisor → Coder → Tester loop:
================================================================
DSRPTV | dsrptv.co — META MULTI-AGENT CODING OS
by DSRPT.AI — https://dsrpt.ai
GitHub: https://github.com/DSRPT/dsrptv
================================================================
LangGraph Meta-Agent Core
Ollama: http://localhost:11434
Primary: qwen2.5-coder:32b
Editor: qwen2.5-coder:14b
CWD: /your/project
================================================================
[DSRPTV] LangGraph Meta-Agent Ready.
[DSRPTV] Persistent memory: ~/.dsrptv_memory.db
[DSRPTV] Type your coding task below (or 'exit'):
You: add a login route to app.py with JWT authentication
[SUPERVISOR] Analyzing task (iteration 0/5)...
[SUPERVISOR] Routing to CODER
[CODER] Running Aider to implement code changes...
[CODER] Aider succeeded.
[SUPERVISOR] Routing to TESTER
[TESTER] Running tests and lint...
[TESTER] pytest passed.
[TESTER] ruff passed.
[SUPERVISOR] Code applied and tests passed. Task complete.
[DSRPTV] Task complete. State persisted to ~/.dsrptv_memory.db
Run bash launch.sh for immediate Aider Architect mode with DSRPTV branding:
================================================================
DSRPTV | dsrptv.co — META MULTI-AGENT CODING OS
by DSRPT.AI — https://dsrpt.ai
GitHub: https://github.com/DSRPT/dsrptv
================================================================
Model: ollama/qwen2.5-coder:32b
Editor: ollama/qwen2.5-coder:14b
Mode: architect
Ollama: http://localhost:11434
================================================================
aider --model ollama/qwen2.5-coder:32b --editor-model ollama/qwen2.5-coder:14b --architect --auto-accept-architect --yes --stream --map --message-history .dsrptv-memory.json| Feature | Claude Code CLI | DSRPTV |
|---|---|---|
| Terminal Interface | ✔️ | ✔️ |
| Architect Mode | ✔️ (single model) | ✔️ (dual model: primary + editor) |
| Autonomous Agent | ✔️ | ✔️ |
| LangGraph Multi-Agent | ❌ | ✔️ (Supervisor + Coder + Tester) |
| Persistent Checkpointing | ❌ | ✔️ (SQLite) |
| Ollama Support | ❌ | ✔️ (local + remote) |
| Model Picker | ❌ | ✔️ (interactive) |
| Voice Input | ❌ | ✔️ (local Whisper) |
| Auto-Test + Lint Loops | ❌ | ✔️ (pytest/npm, ruff/eslint) |
| Web URL Ingestion | ✔️ | ✔️ |
| Pinokio Plugin | ❌ | ✔️ (fully integrated) |
| Windows Support | Limited | ✔️ (native bash path) |
| Open Source | ❌ | ✔️ |
DSRPT/dsrptv/
├── README.md # This file
├── pinokio.js # Pinokio v7 Terminal Plugin config
├── launch.sh # Direct Aider launcher with DSRPTV branding
├── menu.js # Interactive Node.js launcher (Ollama picker, mode selection, extras)
├── graph.py # LangGraph meta-agent core (Supervisor + Coder + Tester)
└── icon.png # DSRPTV icon (optional)
- Env Prompts: First-run config for Ollama host, models, mode
- Windows Bash: Direct kernel path to
miniconda/Library/bin/bash.exe - Auto-Install:
pip install aider-chat langgraph ...on first run - Direct Launch: Executes aider in target project (
{{args.cwd}})
- Ollama API: Fetches live model list from
/api/tags - Model Selection: Numbered menu with defaults
- 7 Agent Modes: Meta multi-agent, native agent, architect, voice, browser, code, ask
- Extras: Voice, memory, auto-test, auto-lint, web URLs
- Command Builder: Constructs full
aidercommand with all flags
StateGraph Workflow:
START
↓
[SUPERVISOR]
│
┌───┼───┐
│ │
[CODER] [TESTER]
│ │
└───┬───┘
│
[SUPERVISOR]
↓
FINISH/LOOP
- Supervisor Node: Routes to coder/tester based on
code_appliedandtests_passedstate - Coder Node: Runs
aider --model ollama/{EDITOR_MODEL} --yes --auto-commits --message "{task}" - Tester Node: Runs
pytest+ruff, updates state - Checkpointing: SqliteSaver persists state to
~/.dsrptv_memory.db - Max Iterations: 5 (configurable)
- ASCII Banner: DSRPTV branding
- Auto-Install: Checks for aider, installs if missing
- Direct Exec:
aider --architect --auto-accept-architect ...
Built by DSRPT.AI
Website: dsrptv.co
GitHub: DSRPT/dsrptv
- Aider — AI pair programming in terminal
- LangGraph — Multi-agent orchestration
- Ollama — Local LLM inference
- Pinokio — One-click AI app launcher
MIT License - see LICENSE for details.
- Website: dsrptv.co
- DSRPT.AI: dsrpt.ai
- GitHub: DSRPT/dsrptv
- Issues: Report a bug