This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Description
Cortex version
1.0.1-203
Describe the Bug
Mac: Concurrent chats for the same model are queued up rather than parallel
- Models tested: tinyllama, llama3.2
- I expect to open 2 CLI windows / Postman window and have concurrent chats
- Works well if separate models (eg tinyllama chat & llama3.2 chat)
May be related to n_parallel parameter in model.yaml
Windows, Ubuntu: Working as expected
Steps to Reproduce
No response
Screenshots / Logs
No response
What is your OS?
What engine are you running?