Fetch latest LLM models from OpenRouter and models.dev APIs.
Validates API responses with Zod schemas to detect breaking changes. Provides smart wrapper functions for model discovery, comparison, cost estimation, recommendations, statistics, and provider info. Features local caching, MCP server integration, multiple output formats, and a fluent query builder API.
brew install maxgfr/tap/llm-modelsnpm install -g llm-modelsDownload the latest binary from GitHub Releases.
| Platform | Binary |
|---|---|
| macOS ARM64 | llm-models-macos-arm64 |
| macOS Intel | llm-models-macos-x64 |
| Linux x86_64 | llm-models-linux-x64 |
| Linux ARM64 | llm-models-linux-arm64 |
| Windows x64 | llm-models-win-x64.exe |
Discover models across both APIs with smart filtering:
# Find cheapest reasoning models
llm-models find -C reasoning --sort cost_input -n 10
# Find models accepting images, sorted by cost
llm-models find -m image --sort cost_input
# Find OpenAI models with large context
llm-models find -p openai --min-context 200000
# Filter by family
llm-models find -f gpt --sort cost_input
# Filter by status
llm-models find --status active -C reasoning
# Search by name
llm-models find -s "claude" --sort cost_input
# Sort by cost-effectiveness (context per dollar)
llm-models find --sort value --desc -n 10
# Count matching models
llm-models find -C reasoning --count
# Output IDs only (for piping)
llm-models find -p openai --ids-only
# Output as CSV or Markdown
llm-models find -p anthropic --format csv
llm-models find -p anthropic --format markdown
# Output as JSON
llm-models find -p google --jsonSide-by-side model comparison with context, output limit, costs, knowledge cutoff, and capabilities:
llm-models compare openai/gpt-4o anthropic/claude-sonnet-4 google/gemini-2.0-flashDetailed information about a single model (description, costs, capabilities, parameters, tokenizer):
llm-models info openai/gpt-4oEstimate costs for specific token volumes:
# Basic cost estimation
llm-models cost openai/gpt-4o anthropic/claude-sonnet-4 -i 1M -o 100K
# With daily/monthly projections
llm-models cost openai/gpt-4o -i 1M -o 100K --daily 1000
# Using a workload profile
llm-models cost openai/gpt-4o -i 1K -o 1K -P chatbot --daily 5000Find the cheapest models matching your criteria:
llm-models cheapest -C reasoning --min-context 100000Get model recommendations for a specific use case:
# List available use cases
llm-models recommend --list
# Get recommendations
llm-models recommend reasoning
llm-models recommend code-gen --max-cost 5
llm-models recommend vision -n 5Available use cases: code-gen, vision, cheap-chatbot, reasoning, long-context, open-source, audio, tool-use.
Get aggregate statistics about the LLM model landscape:
llm-models stats
llm-models stats --jsonShows: total models/providers, top providers, capabilities distribution, modality distribution, cost percentiles, context length stats, newest models.
Show changes since last cached fetch:
llm-models diffShows: new models added, models removed, price changes, status changes.
Get provider details (SDK package, env vars, API endpoint, models):
# List all providers
llm-models provider --list
# Get details for a specific provider
llm-models provider openaillm-models openrouter # All models (JSON)
llm-models openrouter --provider google # Filter by provider
llm-models openrouter --search "claude" # Search
llm-models openrouter --count # Count onlyllm-models models-dev # All providers (JSON)
llm-models models-dev --provider openai # Filter by provider
llm-models models-dev --count # Count onlyllm-models cache # Show cache info
llm-models cache clear # Clear cacheAPI responses are cached for 1 hour in ~/.cache/llm-models/. Use --no-cache to bypass.
llm-models config # Show current config
llm-models config init # Create default config file
llm-models config path # Show config file path
llm-models config profiles # List workload profilesConfig file locations: .llm-modelsrc (project) or ~/.config/llm-models/config.json (global).
# Bash
llm-models completion bash >> ~/.bashrc
# Zsh
llm-models completion zsh >> ~/.zshrc
# Fish
llm-models completion fish > ~/.config/fish/completions/llm-models.fishStart an MCP (Model Context Protocol) server for AI agent integration:
llm-models mcpExposes 8 tools: find_models, compare_models, estimate_cost, cheapest_models, get_provider, list_providers, get_stats, recommend_models.
| Option | Description |
|---|---|
--no-cache |
Disable cache for this request |
--quiet |
Suppress informational messages |
--verbose |
Show debug info (timing, cache status) |
-V, --version |
Show version |
-h, --help |
Show help |
| Option | Description |
|---|---|
-p, --provider <id> |
Filter by provider |
-C, --capability <cap> |
Filter: reasoning, tool_call, structured_output, open_weights, attachment |
-m, --modality <mod> |
Filter by input modality (image, audio, video, pdf) |
--max-cost <n> |
Max input cost ($/million tokens) |
--max-cost-output <n> |
Max output cost ($/million tokens) |
--min-context <n> |
Min context window |
-s, --search <term> |
Search by name or ID |
--status <status> |
Filter: active, beta, deprecated |
-f, --family <name> |
Filter by model family (e.g. gpt, claude, gemini) |
--sort <field> |
Sort: cost_input, cost_output, context_length, release_date, name, knowledge_cutoff, value |
--desc |
Sort descending |
-n, --limit <n> |
Max results (default: 20) |
-c, --count |
Show model count only |
--ids-only |
Output model IDs only, one per line |
--json |
Raw JSON output |
--format <fmt> |
Output format: table, json, csv, markdown |
| Option | Description |
|---|---|
-i, --input <n> |
Input tokens (supports K/M suffix) |
-o, --output <n> |
Output tokens (supports K/M suffix) |
--daily <n> |
Number of daily requests for cost projection |
--monthly <n> |
Number of monthly requests for cost projection |
-P, --profile <name> |
Use workload profile (chatbot, code-gen, rag, summarization, translation) |
import {
fetchUnifiedModels,
findModels,
compareModels,
getProvider,
listProviders,
estimateCost,
cheapestModels,
getStats,
recommendModels,
listUseCases,
diffModels,
query,
setCacheEnabled,
clearCache,
} from "llm-models";
// Get all models (unified from both APIs)
const models = await fetchUnifiedModels();
// Find cheapest reasoning models
const cheap = await findModels({
filter: { capability: "reasoning" },
sort: "cost_input",
limit: 5,
});
// Fluent query builder
const results = await query()
.provider("openai")
.capability("reasoning")
.maxCost(5)
.sortBy("cost_input")
.limit(10)
.execute();
// Compare models side by side
const comparison = await compareModels(["openai/gpt-4o", "anthropic/claude-sonnet-4"]);
// Get provider info (SDK, env vars, docs, models)
const openai = await getProvider("openai");
console.log(openai.npm); // "@ai-sdk/openai"
// Estimate costs
const costs = await estimateCost({
modelIds: ["openai/gpt-4o"],
inputTokens: 1_000_000,
outputTokens: 100_000,
});
// Get recommendations for a use case
const recs = await recommendModels("code-gen", { maxCost: 5, limit: 5 });
// Get market statistics
const stats = await getStats();
// Detect changes since last fetch
const diff = await diffModels();
// Cache control
setCacheEnabled(false); // Disable caching
clearCache(); // Clear cache directoryimport { fetchOpenRouterModels, fetchModelsDevModels } from "llm-models";
const openrouter = await fetchOpenRouterModels();
const modelsDev = await fetchModelsDevModels();All schemas are exported for custom validation:
import { UnifiedModelSchema, OpenRouterResponseSchema, ModelsDevResponseSchema } from "llm-models";All types are inferred from Zod schemas:
import type {
UnifiedModel,
NormalizedCost,
Capabilities,
ProviderInfo,
ModelFilter,
ModelSortField,
CostEstimate,
ModelComparison,
} from "llm-models";Use llm-models in CI/CD to check model costs and status:
- uses: maxgfr/llm-models@v1
with:
command: cost
args: "openai/gpt-4o anthropic/claude-sonnet-4 -i 1M -o 100K"
max-budget: "10"
fail-on-deprecated: "true"- OpenRouter: https://openrouter.ai/api/v1/models
- models.dev: https://models.dev/api.json
MIT