Skip to content

jyswee/easytyga

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

easytyga

Tunnel your local AI to the internet. Secure. One command.

Expose your local Ollama, vLLM, or any HTTP service to the internet with API key authentication. No port forwarding, no static IP, no nginx configs.

npx easytyga
  ┌─────────────────────────────────────┐
  │          easytyga v1.0.3             │
  │  Tunnel your local AI to the web     │
  └─────────────────────────────────────┘

  GPU:     NVIDIA GeForce RTX 4090
  Target:  http://localhost:11434
  Relay:   wss://easytyga.com/ws

  ┌─────────────────────────────────────┐
  │  Tunnel active                      │
  └─────────────────────────────────────┘

  Public URL:  https://easytyga.com/t/abc123
  API Key:     et_a1b2c3d4e5f6...

That's it. Your local AI is now accessible from anywhere, secured with an API key.

Why?

Ollama has no built-in authentication and no remote access. If you want to use your GPU from your phone, office, or a cloud app, you need to cobble together nginx + Cloudflare tunnels + basic auth.

easytyga solves this in one command:

  • Secure tunnel -- no port forwarding, works behind any NAT/firewall
  • API key auth -- auto-generated, every request authenticated
  • Auto-detects your GPU -- knows what hardware you're running
  • Works with anything -- Ollama, vLLM, LocalAI, ComfyUI, or any HTTP service

Usage

# Tunnel local Ollama (default port 11434)
npx easytyga

# Tunnel a different service
npx easytyga --target http://localhost:8080

# Use a specific key
npx easytyga --key et_abc123...

Options

Flag Description
--target <url> Local endpoint (default: http://localhost:11434)
--server <url> Relay server URL
--key <key> Connection key
--help Show help

How it works

  1. easytyga connects to the relay server via WebSocket
  2. The relay assigns you a public URL with API key auth
  3. Incoming requests are forwarded through the tunnel to your local service
  4. Responses stream back to the caller

Your IP stays private. No ports to open. Works from any network.

Features

  • One command -- npx easytyga, no install, no config files
  • API key auth -- every tunnel gets a unique key, no anonymous access
  • GPU detection -- auto-detects NVIDIA, AMD, and Apple Silicon
  • Model detection -- discovers installed Ollama models automatically
  • Auto-reconnect -- exponential backoff, set and forget
  • Streaming -- full support for streaming responses (chat, generate)
  • Any HTTP service -- not locked to Ollama, tunnel anything

Integrations

easytyga supports plugins for third-party services:

  • gpusmarket.com -- List your GPU on the peer-to-peer rental marketplace and earn money (--list)
  • agenticmemory.ai -- Add persistent conversation memory to your AI (--memory)

See easytyga.com for documentation on integrations and self-hosted relay servers.

License

MIT


Trademarks

"easytyga" and the easytyga logo are trademarks of Joe Wee. The name and branding may not be used in derivative works without written permission.

This software is provided under the MIT license -- you are free to use, modify, and distribute the code, but the easytyga name, branding, and relay infrastructure remain the property of the author.

Ollama is a trademark of Ollama, Inc. vLLM is a trademark of vLLM contributors. NVIDIA, GeForce, and RTX are trademarks of NVIDIA Corporation. AMD and Radeon are trademarks of AMD. Apple and Apple Silicon are trademarks of Apple Inc. All other trademarks are the property of their respective owners. This project is not affiliated with or endorsed by any of these companies.


Built by Joe Wee

About

Tunnel your local AI to the internet. Secure. One command.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors