Skip to content

yazinsai/OpenOats

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

63 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

OpenOats

A meeting note-taker that talks back.

Download for Mac

OpenOats sits next to your call, transcribes both sides of the conversation in real time, and searches your own notes to surface things worth saying — right when you need them.

OpenOats during a call — suggestions drawn from your own notes appear at the top, live transcript below

Features

  • Invisible to the other side — the app window is hidden from screen sharing by default, so no one knows you're using it
  • Fully offline transcription — speech recognition runs entirely on your Mac; no audio ever leaves the device
  • Runs 100% locally — pair with Ollama for LLM suggestions and local embeddings, and nothing touches the network at all
  • Pick any LLM — use OpenRouter for cloud models (GPT-4o, Claude, Gemini) or Ollama for local ones (Llama, Qwen, Mistral)
  • Live transcript — see both sides of the conversation as it happens, copy the whole thing with one click
  • Auto-saved sessions — every conversation is automatically saved as a plain-text transcript and a structured session log, no manual export needed
  • Knowledge base search — point it at a folder of notes and it pulls in what's relevant using Voyage AI embeddings, local Ollama embeddings, or any OpenAI-compatible endpoint (llama.cpp, llamaswap, LiteLLM, vLLM, etc.)

How it works

  1. You start a call and hit Live
  2. OpenOats transcribes both speakers locally on your Mac
  3. When the conversation hits a moment that matters — a question, a decision point, a claim worth backing up — it searches your notes and surfaces relevant talking points
  4. You sound prepared because you are

Recording Consent & Legal Disclaimer

Important: OpenOats records and transcribes audio from your microphone and system audio. Many jurisdictions have laws requiring consent from some or all participants before a conversation may be recorded (e.g., two-party/all-party consent states in the U.S., GDPR in the EU).

By using this software, you acknowledge and agree that:

  • You are solely responsible for determining whether recording is lawful in your jurisdiction and for obtaining any required consent from all participants before starting a session.
  • The developers and contributors of OpenOats provide no legal advice and make no representations about the legality of recording in any jurisdiction.
  • The developers accept no liability for any unauthorized or unlawful recording conducted using this software.

Do not use this software to record conversations without proper consent where required by law.

The app will ask you to acknowledge these obligations before your first recording session.

Download

Grab the latest DMG from the Releases page.

Or build from source:

./scripts/build_swift_app.sh

Quick start

  1. Open the DMG and drag OpenOats to Applications
  2. Launch the app and grant microphone + screen capture permissions
  3. Open Settings (Cmd+,) and pick your providers:
    • Cloud: add your OpenRouter and Voyage AI API keys
    • Local: select Ollama as your LLM and embedding provider (make sure Ollama is running)
    • OpenAI-compatible: select "OpenAI Compatible" as your embedding provider and point it at any /v1/embeddings endpoint
  4. Point it at a folder of .md or .txt files — that's your knowledge base
  5. Click Idle to go live

The first run downloads the local speech model (~600 MB).

What you need

  • Apple Silicon Mac, macOS 26+
  • Xcode 26 / Swift 6.2
  • For cloud mode: OpenRouter API key + Voyage AI API key
  • For local mode: Ollama running locally with your preferred models (e.g. qwen3:8b for suggestions, nomic-embed-text for embeddings)
  • For OpenAI-compatible embeddings: any server implementing /v1/embeddings (llama.cpp, llamaswap, LiteLLM, vLLM, etc.)

Knowledge base

Point the app at a folder of Markdown or plain text files. That's it. OpenOats chunks, embeds, and caches them locally. When the conversation shifts, it searches your notes and only surfaces what's actually relevant.

Works well with meeting prep docs, research notes, pitch decks, competitive analysis, customer briefs — anything you'd want at your fingertips during a call.

Privacy

  • Speech is transcribed locally — audio never leaves your Mac
  • With Ollama: everything stays on your machine. Zero network calls.
  • With cloud providers: KB chunks are sent to Voyage AI (or your chosen OpenAI-compatible endpoint) for embedding (text only, no audio), and conversation context is sent to OpenRouter for suggestions
  • API keys are stored in your Mac's Keychain
  • The app window is hidden from screen sharing by default
  • Transcripts are saved locally to ~/Documents/OpenOats/

Build

# Full build → sign → install to /Applications
./scripts/build_swift_app.sh

# Dev build only
cd OpenOats && swift build -c debug

# Package DMG
./scripts/make_dmg.sh

Optional env vars for code signing and notarization: CODESIGN_IDENTITY, APPLE_ID, APPLE_TEAM_ID, APPLE_APP_PASSWORD.

Repo layout

OpenOats/             SwiftUI app (Swift Package)
scripts/              Build, sign, and package scripts
assets/               Screenshot and app icon source

License

MIT