Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

bug: Model failed to get model status with status code 409 but model runs successfully #1504

@gabrielle-ong

Description

@gabrielle-ong

Cortex version

v172

Describe the Bug

cortex run returns model failed 409 status code.
But model starts successfully, can successfully chat (CLI, API postman).
This is confusing to users as they may think there is an error
Seems like Louis also got a 409 error in #1475

> cortex run model
Starting server ...
Host: 127.0.0.1 Port: 39281
Server started
API Documentation available at: http://127.0.0.1:39281
Error: Model failed to get model status with status code: 409
tinyllama:1b-gguf model started successfully. Use `cortex-nightly chat tinyllama:1b-gguf` for interactive chat shell

Steps to Reproduce

No response

Screenshots / Logs

image

What is your OS?

  • MacOS
  • Windows
  • Linux

What engine are you running?

  • cortex.llamacpp (default)
  • cortex.tensorrt-llm (Nvidia GPUs)
  • cortex.onnx (NPUs, DirectML)

Metadata

Metadata

Assignees

No one assigned

    Labels

    category: model runningInference ux, handling context/parameters, runtimetype: bugSomething isn't working

    Type

    No type

    Projects

    Milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions