The AI Boardroom — Synthesis of Diverse Intelligence.
Lynx is not just a model router; it's a multi-modal orchestration engine designed to consult multiple LLMs simultaneously ("The Boardroom") to find patterns, reach consensus, and identify outliers.
Most orchestration tools focus on reducing costs by routing to the cheapest model. Lynx takes a different approach: Intelligence Density.
By streaming responses from multiple frontier models (e.g., GPT-4, Claude 3, Llama 3) in parallel and synthesizing them in real-time, Lynx provides:
- Consensus: High confidence when models agree.
- Nuance: Different perspectives on complex creative or reasoning tasks.
- Reliability: Fallback redundancy and outlier detection.
Lynx is built as a TypeScript Monorepo using Turborepo.
apps/api: Node.js backend using Fastify. Handles client connections, manages SSE (Server-Sent Events) streams, and invokes the orchestrator.apps/web: React + Vite frontend. Provides the "Boardroom" UI for users to select models and watch parallel streams.packages/core: The brain of Lynx. Contains theStreamOrchestrator,ModelStrategyinterfaces, and provider logic. It uses advanced async handling to merge distinct model streams into a single unified event stream.packages/shared: Shared Zod schemas (LynxRequest,LynxEvent) and TypeScript types used across the entire stack.
- Language: TypeScript throughout.
- Monorepo: Turborepo, NPM Workspaces.
- Backend: Fastify, Server-Sent Events (SSE).
- Frontend: React, Tailwind CSS, Vite.
- Validation: Zod.
- AI/LLM: Ollama (Local), Mock Providers (Remote simulation).
Lynx is configured to use local models via Ollama.
- Install Ollama.
- Pull the required models:
ollama pull gemma3 ollama pull gpt-oss # Or any other model you wish to map - Start the Ollama server:
ollama serve
# Install dependencies
npm install
# Start the full stack (API + Web)
npm run dev- API runs on
http://localhost:3000 - Web Interface runs on
http://localhost:5173(typical Vite port)
- Open
http://localhost:5173in your browser. - Command Center:
- Enter your prompt (e.g., "Analyze the impact of AI on privacy").
- Select the models you want to consult (toggle
gpt-oss,gemma3, or mocks). - Click "SUMMON BOARD".
- The Workspace:
- Watch as multiple columns stream text in parallel.
- Sync Scroll: Scrolling one column will automatically scroll the others for easy line-by-line comparison.
- Synthesis:
- The top/highlighted card will show the Aggregated Insight, a synthesized summary of the board's opinions.
Compile all packages and apps:
npm run buildTo verify the core orchestration logic independently:
# Run the demo script in packages/core
npm run test --workspace=@lynx/core
# Verify API connection to Ollama
npx ts-node apps/api/test-ollama.ts- Client sends request with
promptandmodelIds. - API validates request using
@lynx/sharedschemas. - Orchestrator initializes selected providers (Ollama or Mock).
- Parallel Streaming: Opens streams to all providers concurrently.
- Merging: Chunks are yielded instantly as they arrive from any provider via
Promise.race. - Synthesis: Aggregates results after streams finish for a final summary.
- Client receives a single server-sent event stream updating the UI in real-time.