Your multi-model AI control center
AgentK is an AI control center that unifies access to 7 major AI providers under one sleek, minimalistic interface. With AgentK, you can swap between models during chat sessions, manage multiple conversations at the same time, and interact with different AI providers individually or simultaneously — all without ever leaving the app.
Thanks to direct integration with provider APIs, you have the flexibility to use hundreds of available models across OpenAI, Anthropic, Groq, Gemini, Perplexity, Cohere, and Hugging Face. Whether you're prototyping ideas, performing research, or just exploring the capabilities of modern AI, AgentK makes it simple to experiment with the latest cutting edge AI models without spending money every month on expensive AI subscriptions.
- Features
- Prerequisites
- Installation
- Configuration
- API Keys
- Usage
- Advanced Features
- Provider Documentation
- License
🔹 Multi-AI-Provider Support
- OpenAI
- Anthropic (Claude)
- Groq
- Perplexity
- Cohere
- Google Gemini
- Hugging Face
- Hugging Face Inference Providers (gives access to 10+ more AI providers & hundreds of models)
🔹 Multi-Model Conversations
Talk to multiple models simultaneously or one at a time. Compare responses across different models for the same prompt.
🔹 Dynamic Context Sharing
Toggle shared context on/off between models. Let models see each other's responses or keep conversations isolated.
🔹 Dynamic Model Pulling
Edit the agentk_config.json
file to pull any model offered by your providers & enter API keys.
🔹 Secure API Key Management
Store your keys locally; AgentK never commits secrets as it runs locally.
🔹 Session Control
Create, switch, and delete named chat sessions.
🔹 Resource Optimization
Control token usage and response length through model configurations.
- Go 1.20+ (for the backend)
- Node.js 16+ & npm (for the frontend)
-
Clone this repo
git clone https://github.com/your-username/AgentK.git cd AgentK
-
Backend
cd backend go mod tidy (optional) go run main.go
-
Frontend
cd ../frontend npm install npm run dev
-
Open your browser to
http://localhost:5173
and enjoy.
IMPORTANT: You must sign up for API keys from each provider you wish to use: You won't be able to run inference otherwise since the request payload for all provider endpoints require it. Once you've grabbed your API keys, make sure to drop them into the AgentK_config.json file. You can skip this step for any providers you dont intend to use.
- OpenAI API Keys
- Anthropic API Keys
- Groq API Keys
- Perplexity API Keys
- Cohere API Keys
- Google AI API Keys
- Hugging Face API Keys
OpenAI and Claude allow you to create an API key for free; however, you cannot run inference on their models without purchasing API credits. Both platforms require a minimum payment of $5 to enable access. Perplexity also requires an upfront payment of at least $3 for API usage (in my opinion their Sonar models are underwhelming).
The rest of the AI providers offer both free and paid tiers, allowing you to start interacting with models at no cost. Groq and Hugging Face stand out for their generous free-tier access and offer some of the most cost-effective API credits available.
Hugging Face is particularly powerful, giving you access to hundreds of models, along with integrations from over 10 dedicated inference providers, all through a single API key. You can also visit any provider's website and try out their flagship models — just make sure to use the correct Hugging Face API URL and model name from their documentation. This unlocks a wide range of models hosted by various companies, greatly expanding your available options.
Google Gemini also provides an excellent free tier, offering access to their powerful Gemini models. In my experience, Gemini models deliver incredible response times and support very large context sizes, making them an outstanding option for demanding applications.
- Open
agentk_config.json
located at the project root folder. - Under the
"Providers"
section, enter your API keys for providers you would like access to: - Under
"Models"
, list any model IDs you wish to use. Example:"Models": { "gpt-4o-2024-08-06": { "id": "gpt-4o-2024-08-06", // the ID property needs to match the key, see default models for example "name": "GPT-4o", // model name you would like to use, can be custom "provider": "OpenAI", // AI provider, see options above "endpoint": "https://api.openai.com/v1/chat/completions", // API endpoint, see provider docs for URL endpoints (see curl example) "contextSize": 128000, // Total available context (prompt + history) for the model per request. Affects API costs. Adjustable. See provider documentation. "maxCompletionTokens": 16384 // Maximum size of the model's response. Also impacts API costs. Adjustable. See provider documentation. }, // ...more models }
- Start the backend
cd backend go mod tidy (optional) go run main.go
- Start the frontend
cd ../frontend npm run dev
- Open your browser to
http://localhost:5173
and enjoy.
- Create a new session by clicking the "+" button
- Toggle "Shared Context" on to allow models to see each other's responses
- Select multiple models from the dropdown menu
- Type your prompt and see how different models respond
- Toggle models on/off during the conversation to control which ones respond
AgentK allows you to:
- Compare Responses: Ask the same question to multiple models simultaneously
- Chain Conversations: Let one model build on another's response
- Isolate Conversations: Toggle shared context off to keep model conversations separate
AgentK allows you to control how much of the conversation history is sent to the model. By default Context Size and Max Completion Tokens are set to highest value possible according to provider docs. Feel free to play around with these values but make sure to monitor.
-
Context Size: The
contextSize
parameter in your model configuration controls the maximum number of tokens used for the conversation history. This helps manage resource usage and API costs. -
Max Completion Tokens: The
maxCompletionTokens
parameter limits how many tokens the model can generate in response. If not specified, AgentK uses a reasonable default based on the model's capabilities.
Create different variants of models with adjusted context sizes and response limits:
"llama-3.3-70b-economy": {
"id": "llama-3.3-70b-economy",
"name": "LLaMA 3.3 (Economy)",
"provider": "Groq",
"endpoint": "https://api.groq.com/openai/v1/chat/completions",
"contextSize": 32000, // Reduced from 128000
"maxCompletionTokens": 8192 // Reduced from 32768
}
This allows users to choose the resource profile that best fits their needs, balancing performance with efficiency.
Pull model IDs from any of these sources:
- 🔹 OpenAI Models
- 🔹 Anthropic Claude Models
- 🔹 Groq Models
- 🔹 Perplexity Models
- 🔹 Cohere Models
- 🔹 Google Gemini API Models
- 🔹 Hugging Face Chat-Completion
- 🔹 Hugging Face Inference Providers
This project is licensed under the MIT License. See here for details.