Skip to content

config(llm): reprioritize models and switch to gemini-2.5-flash-lite across profiles#50

Merged
neoneye merged 8 commits intoPlanExeOrg:mainfrom
VoynichLabs:feature/model-profile-switching
Feb 19, 2026
Merged

config(llm): reprioritize models and switch to gemini-2.5-flash-lite across profiles#50
neoneye merged 8 commits intoPlanExeOrg:mainfrom
VoynichLabs:feature/model-profile-switching

Conversation

@82deutschmark
Copy link
Copy Markdown
Collaborator

Reprioritize models and switch to gemini-2.5-flash-lite across custom, frontier, and premium profiles. (Based on commit 2bfdf6a)

Larry the Laptop Lobster and others added 8 commits February 18, 2026 14:49
Add dynamic model list rendering based on selected profile in both multi-user frontend and single-user Gradio app. Mount llm_config*.json files into frontend_multi_user container, parse each profile's config to extract model names sorted by priority, and display them in the UI. Update llm_config.custom.json with cleaner priority-based model ordering and remove unused entries.
…_user container

Add volume mounts and sync actions for llm_config.premium.json, llm_config.frontier.json, and llm_config.custom.json in docker-compose.yml. Update Dockerfile to copy all llm_config*.json files during build. Enables multi-profile support in single-user frontend.
…ig display

Add priority prefix (P0, P1, etc.) to each model in the profile models markdown. Create separate active_config_markdown component that mirrors profile_models_markdown to display current configuration. Update all callback outputs to populate both markdown components with identical content.
Increase max_tokens from 8192 to 32000 across all OpenRouter models in custom and premium profiles. Reorder custom profile priorities (qwen3-coder-next 3→4, nemotron 4→5, gemini-3-flash comment updated to "last resort"). Swap premium profile priorities (gemini-3-flash 1→6, qwen3-coder-next 4→1). Remove local model entries (ollama-llama3.1, docker-ollama-llama3.1, ollama-qwen2.5-coder, lmstudio variants) from premium profile.
…iorities

Move priority field before luigi_workers in all config files for consistency. In custom profile, demote gemini-3-flash from priority 2 to 12 and promote qwen3-coder-next from priority 4 to 2. In frontier profile, reduce max_tokens from 60000 to 32000 across all models. In premium profile, swap qwen3-5-397b-a17b to priority 1 and demote glm-4-7-flash to priority 6.
… and update custom profile models

Raise temperature parameter from 0.1 to 1.0 for all models in custom, frontier, and premium profiles to increase response creativity. In custom profile, replace minimax-01 with minimax-m2.5 and swap qwq-32b-preview with qwen3.5-397b-a17b.
…across profiles

Demote gpt-oss-120b from priority 1 to 11 and promote qwen3-coder-next from priority 2 to 1 in custom profile. Replace gemini-3-flash-preview with gemini-2.5-flash-lite-preview-09-2025 at priority 12. In frontier profile, demote glm-5 from priority 1 to 11, demote kimi-k2-5 from priority 2 to 111, and promote qwen3-5-397b-a17b from priority 3 to 1. In premium profile, swap qwen3-5-397b-a17b to priority 2, replace
@neoneye neoneye merged commit 9c338df into PlanExeOrg:main Feb 19, 2026
@neoneye neoneye deleted the feature/model-profile-switching branch February 20, 2026 02:25
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants