QUIT: A Human-in-the-Loop Platform for AI Research Automation
Query · Understand · Implement · Tell
✦ Quit the Old Way of Doing Research ✦
QUIT is a human-in-the-loop research assistant — not a black box, but a transparent pipeline where researchers remain in full control at every step. Its artifact-driven design eliminates long-context dependency and avoids redundant token consumption.
QUIT also fully supports end-to-end mode — just set a topic and let the pipeline run to completion without any intervention. End-to-end cost with DeepSeek-V4-Pro: roughly ¥10 ($1.5) per manuscript.
| Stage | What it does |
|---|---|
| 🔍 Query | Search papers, repositories, and local literature |
| 💡 Understand | Extract evidence cards, cluster insights, generate ideas |
| 🔧 Implement | Convert the selected idea into code, run experiments, audit results |
| 📝 Tell | Draft and review paper from actual outputs |
Four specialized agents are coordinated by a central Orchestrator / State Machine:
- 🗺️ PlannerAgent — turns the user's topic into a validated
ResearchBrief - 🔬 ResearchAgent — retrieves papers, extracts evidence, and synthesizes ideas
- 🏗️ BuilderAgent — generates experiment code, runs it, and writes the paper
- 🔍 ReviewerAgent — audits ideas, code quality, and the paper draft
All coordination happens through named artifact files on disk — no agent shares conversational memory across invocations. This makes the entire pipeline traceable, reproducible, and resumable.
🙋 Human-in-the-Loop control points: researchers can stop after any state to inspect artifacts, edit intermediate files (evidence cards, BuildSpec, generated code, results), and resume from a chosen state.
Requirements: Python 3.11+, Git, LaTeX (texlive + latexmk)
git clone https://github.com/Mr-XcHan/QUIT.git
cd QUIT
bash setup.shThe script creates a .venv at the repo root and installs the agent and web UI. For local LLM inference (torch + transformers):
bash setup.sh --with-localMissing system tools (latexmk, bibtex) are reported with install hints at the end of setup.
The main config file is Quit_v0_3/config.json. Default parameters are ready to use — the only fields you typically need to set are your research topic and LLM credentials. All parameters in this file serve as defaults and can also be overridden directly in the Web UI before each run.
{
"project": {
"topic": "Flow Matching for Offline Reinforcement Learning"
},
"llm": {
"provider": "openai",
"model": "gpt-5.5",
"api_key_env": "OPENAI_API_KEY"
}
}Store your API key in a .env file next to config.json (never commit it):
OPENAI_API_KEY=sk-...
Supported providers: anthropic, openai, deepseek, local-vllm, and other OpenAI-compatible endpoints.
⚙️ Other settings you may want to adjust
| Field | Default | Description |
|---|---|---|
runtime.stop_after |
null |
Stop after a specific state (e.g. "CODE_EVAL") |
run_budget.experiment_timeout_seconds |
3600 |
Max time for generated experiments |
retrieval.sources |
["arxiv"] |
Paper search sources |
write.expected_main_pages |
8 |
Target paper length |
Activate the environment and start the server:
source .venv/bin/activate
cd Quit_v0_3_web
python server.py --port 7862Then open http://localhost:7862 in your browser.
🙋 Human-in-the-loop intervention: because all outputs are files, you can at any point:
- 👁️ Inspect artifacts (evidence cards, BuildSpec, code, results, paper) from the Artifacts panel
- ⏹️ Stop at any state, review outputs, then continue
- ✏️ Edit intermediate files and resume from a chosen state — without rerunning earlier stages
- 🔁 Rerun a specific state after corrections (e.g. edit the BuildSpec, then rerun
CODE)
📂 Browsing past runs: all runs are saved under Quit_v0_3/runs/<run_id>/. The Runs panel lists every past run with its full artifact trail, prompts, and LLM responses.
Each run produces a complete artifact trail under runs/<run_id>/:
ResearchBrief.json ← validated research plan
EvidenceCards.jsonl ← structured paper evidence
IdeaLibrary.jsonl ← candidate ideas with evidence links
BuildSpec.json ← experiment + paper contract
code/src/*.py ← generated experiment code
results/metrics.json ← experiment results
results/results_table.csv ← per-method comparison table
CodePerformanceEval.json ← LLM verdict on method vs. baselines
paper_gene/main.tex ← generated LaTeX paper
paper_gene/main.pdf ← compiled paper
run_trace.json ← full state transition log
All prompts and raw LLM responses are saved under llm/ for full reproducibility.
Papers produced end-to-end by QUIT Agent across different research domains:
We're looking for testers! 🚀 Try the pipeline with your own research topic and see how far it gets. Full auto-run results are welcome, but we're especially excited to feature human-in-the-loop works — papers where researchers stepped in to guide the agent, edit intermediate artifacts, or steer the direction. Those tend to be the most interesting results.
If you'd like your output showcased here, feel free to open an issue or pull request. 🙌
Xinchen — isxinchen.han@gmail.com
Feel free to reach out with questions, feedback, or collaboration ideas.
Apache 2.0 — free to use and build upon, including commercially, with attribution and notice of modifications required. See LICENSE for details.



