Anti-sycophancy multi-evaluator engine for LLM agents.
Empanel a grand jury of independent reviewers over the same code. Each lens returns its own verdict. Weighers score them in isolation. A deterministic synthesizer combines the findings into a ranked report — without any reviewer ever seeing another's opinion.
The thesis: LLM-as-judge workflows anchor catastrophically when one model sees another's output. Empanel enforces isolation at every stage so diverse lenses produce diverse findings, and weighers can't be flattered into consensus.
pip install empanel
# Review a file with four independent lenses, synthesize a ranked report
empanel review \
--files src/contract.sol \
--spec spec.md \
--model claude-opus-4-7 \
--output review.json \
--markdown review.mdfrom empanel import CodeReviewEngine
from empanel.lenses import SECURITY, SPEC, EDGE_CASES, ARCHITECTURE
engine = CodeReviewEngine(lenses=[SECURITY, SPEC, EDGE_CASES, ARCHITECTURE])
result = engine.run(code=Path("src/contract.sol").read_text(),
spec=Path("spec.md").read_text())
for finding in result.findings:
print(f"[{finding.confidence}] {finding.title} — {finding.location}")A grand jury is the closest real proceeding to what this tool does:
- Multiple independent reviewers hearing the same evidence
- No cross-examination between reviewers — each works in isolation
- Output is a ranked list of indictments (issues worth pursuing), not a verdict
- Evidence-enforced — every finding must cite the code
The architecture maps directly onto those properties. Adding a new lens is seating another juror; tightening weighing is tightening isolation rules.
Three phases, isolated by construction:
- Evaluate — each lens reviews the code independently. No cross-talk.
- Weigh — each weigher scores the raw findings without seeing other weighers' scores or lens identities. This is the anti-sycophancy wedge: a weigher can't be flattered into agreement with the majority.
- Synthesize — deterministic math combines the scores. Finding fingerprints deduplicate near-identical reports. Confidence tiers fall out of reviewer concurrence, not vibes.
Replay artifacts are stored as JSON so any review can be reproduced or disputed after the fact.
SECURITY— adversarial threat model framingSPEC— compares implementation against an optional specEDGE_CASES— boundaries, error paths, null/empty, overflowARCHITECTURE— coupling, leaky abstractions, hidden statePERFORMANCE— complexity, allocation, hot pathsREADABILITY— naming, flow, cognitive load
Register custom lenses by subclassing Lens and passing them to the
engine. The only contract is "return a list of Findings with evidence."
- Claude Code slash command — qa-hard uses empanel as the review backend. Runs without an API key by dispatching each evaluator through the Claude Code Agent tool.
- Standalone CLI —
empanel reviewworks with any Anthropic-API-keyed setup. Model is a flag, so anything that quacks like Claude works. - Fixtures —
empanel.fixturesbundles the regression corpus of real bugs the engine has caught. Usetests/test_fixtures.pyas a template for pinning your own.
- v0.3.0 — renamed from
independent-eval. 141 tests pass. - Self-review converged after three rounds at v0.2.x — engine reviews its own source without regressions.
- Roadmap:
ROADMAP.md(pluggable lenses, cost budgeting, cross-session replay diffing).
MIT.