Skip to content

Model deepseek-r1-distill-qwen-14b does not work on NVidia RTX A6000 48GB #4710

Closed as not planned
@huksley

Description

@huksley

LocalAI version:

LocalAI Version d9204ea (d9204ea)

Environment, CPU architecture, OS, and Version:

x86, Ubuntu 24.04, CUDA 12.6

Describe the bug

Using docker compose and downloading model. gpt4 and gp4o works.
Downloaded model successfully and when I go to Chat => Select model deepseek-r1-distill-qwen-14b and write to chat,
After sometime, no response generated

To Reproduce

  1. install with docker compose
  2. download model
  3. go to Chat => Select model deepseek-r1-distill-qwen-14b
  4. write to chat
  5. no response, no loading progress indicator

Expected behavior

chat works

Logs

Additional context

Running it using DollarDeploy and this docker compose setup: https://github.com/dollardeploy/templates/tree/main/local-ai-nvidia-cuda-12

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions