A meeting note-taker that talks back.
OpenOats sits next to your call, transcribes both sides of the conversation in real time, and searches your own notes to surface things worth saying — right when you need them.
- Invisible to the other side — the app window is hidden from screen sharing by default, so no one knows you're using it
- Fully offline transcription — speech recognition runs entirely on your Mac; no audio ever leaves the device
- Runs 100% locally — pair with Ollama for LLM suggestions and local embeddings, and nothing touches the network at all
- Pick any LLM — use OpenRouter for cloud models (GPT-4o, Claude, Gemini) or Ollama for local ones (Llama, Qwen, Mistral)
- Live transcript — see both sides of the conversation as it happens, copy the whole thing with one click
- Auto-saved sessions — every conversation is automatically saved as a plain-text transcript and a structured session log, no manual export needed
- Knowledge base search — point it at a folder of notes and it pulls in what's relevant using Voyage AI embeddings, local Ollama embeddings, or any OpenAI-compatible endpoint (llama.cpp, llamaswap, LiteLLM, vLLM, etc.)
- You start a call and hit Live
- OpenOats transcribes both speakers locally on your Mac
- When the conversation hits a moment that matters — a question, a decision point, a claim worth backing up — it searches your notes and surfaces relevant talking points
- You sound prepared because you are
Important: OpenOats records and transcribes audio from your microphone and system audio. Many jurisdictions have laws requiring consent from some or all participants before a conversation may be recorded (e.g., two-party/all-party consent states in the U.S., GDPR in the EU).
By using this software, you acknowledge and agree that:
- You are solely responsible for determining whether recording is lawful in your jurisdiction and for obtaining any required consent from all participants before starting a session.
- The developers and contributors of OpenOats provide no legal advice and make no representations about the legality of recording in any jurisdiction.
- The developers accept no liability for any unauthorized or unlawful recording conducted using this software.
Do not use this software to record conversations without proper consent where required by law.
The app will ask you to acknowledge these obligations before your first recording session.
Grab the latest DMG from the Releases page.
Or build from source:
./scripts/build_swift_app.sh- Open the DMG and drag OpenOats to Applications
- Launch the app and grant microphone + screen capture permissions
- Open Settings (
Cmd+,) and pick your providers:- Cloud: add your OpenRouter and Voyage AI API keys
- Local: select Ollama as your LLM and embedding provider (make sure Ollama is running)
- OpenAI-compatible: select "OpenAI Compatible" as your embedding provider and point it at any
/v1/embeddingsendpoint
- Point it at a folder of
.mdor.txtfiles — that's your knowledge base - Click Idle to go live
The first run downloads the local speech model (~600 MB).
- Apple Silicon Mac, macOS 26+
- Xcode 26 / Swift 6.2
- For cloud mode: OpenRouter API key + Voyage AI API key
- For local mode: Ollama running locally with your preferred models (e.g.
qwen3:8bfor suggestions,nomic-embed-textfor embeddings) - For OpenAI-compatible embeddings: any server implementing
/v1/embeddings(llama.cpp, llamaswap, LiteLLM, vLLM, etc.)
Point the app at a folder of Markdown or plain text files. That's it. OpenOats chunks, embeds, and caches them locally. When the conversation shifts, it searches your notes and only surfaces what's actually relevant.
Works well with meeting prep docs, research notes, pitch decks, competitive analysis, customer briefs — anything you'd want at your fingertips during a call.
- Speech is transcribed locally — audio never leaves your Mac
- With Ollama: everything stays on your machine. Zero network calls.
- With cloud providers: KB chunks are sent to Voyage AI (or your chosen OpenAI-compatible endpoint) for embedding (text only, no audio), and conversation context is sent to OpenRouter for suggestions
- API keys are stored in your Mac's Keychain
- The app window is hidden from screen sharing by default
- Transcripts are saved locally to
~/Documents/OpenOats/
# Full build → sign → install to /Applications
./scripts/build_swift_app.sh
# Dev build only
cd OpenOats && swift build -c debug
# Package DMG
./scripts/make_dmg.shOptional env vars for code signing and notarization: CODESIGN_IDENTITY, APPLE_ID, APPLE_TEAM_ID, APPLE_APP_PASSWORD.
OpenOats/ SwiftUI app (Swift Package)
scripts/ Build, sign, and package scripts
assets/ Screenshot and app icon source
MIT
