Skip to content
dlorp edited this page Feb 2, 2026 · 5 revisions

Synapse Engine

Distributed orchestration platform for local language models.

Status Dashboard

What is this?

Synapse Engine coordinates multiple local LLMs into a unified inference system. Instead of running models one at a time, it orchestrates them across performance tiers-fast models for initial processing, powerful models for refinement-with sub-100ms contextual retrieval built in.

It runs entirely on your hardware. No API keys, no cloud dependencies, no data leaving your machine.

Key capabilities:

  • Multi-model orchestration across FAST/BALANCED/POWERFUL tiers
  • Query modes: Simple, Two-Stage, Council (Consensus/Debate), Benchmark
  • CGRAG: Sub-100ms contextual retrieval with FAISS
  • Metal acceleration for Apple Silicon
  • WebUI-first design - full control in the browser

Getting Started

  1. Docker Quick Start - Full installation and setup
  2. Docker Quick Reference - Command cheat sheet
  3. Model Management UI - Using the web interface

User Guides

Setup & Operations

Model Management


Features

Query Modes

  • Query Modes Overview - All available query processing modes:
    • Simple Mode - Single model, direct response
    • Two-Stage Mode - Fast tier + CGRAG retrieval, then refinement
    • Council Mode - Multi-model collaboration or debate
    • Benchmark Mode - Compare models side-by-side

Dynamic Control


Contributing


Screenshots

Query Modes Model Management
Query Modes Models
System Status Architecture
Status Topology

Clone this wiki locally