Skip to content
#

inference

Here are 45 public repositories matching this topic...

Docker image for a self-hosted Whisper speech-to-text server with an OpenAI-compatible transcription API. Powered by faster-whisper. Supports all Whisper models, JSON/SRT/VTT output, SSE streaming, offline mode, and multi-arch (amd64, arm64).

  • Updated Apr 22, 2026
  • Shell

The Private AI Setup Dream Guide for Demos automates the installation of the software needed for a local private AI setup, utilizing AI models (LLMs and diffusion models) for use cases such as general assistance, business ideas, coding, image generation, systems administration, marketing, planning, and more.

  • Updated Dec 20, 2025
  • Shell

Self-hosted AI gateway. 82 models across 12 providers, free-tier-first routing with automatic fallback. Local CPU inference, transcription, and TTS. Models autonomously browse the web (stealth), run agentic Claude Code, and manage object storage via MCP. Expose publicly through Cloudflare Tunnel.

  • Updated Apr 21, 2026
  • Shell

Improve this page

Add a description, image, and links to the inference topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the inference topic, visit your repo's landing page and select "manage topics."

Learn more