Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

serving llama3 does not work #3876

Closed
lambdaofgod opened this issue Apr 24, 2024 · 5 comments
Closed

serving llama3 does not work #3876

lambdaofgod opened this issue Apr 24, 2024 · 5 comments
Labels
bug Something isn't working

Comments

@lambdaofgod
Copy link

lambdaofgod commented Apr 24, 2024

What is the issue?

I am able to run llama 3 (ollama run llama3) but when I try to run the server I get

{"error":"model 'llama3' not found, try pulling it first"}

This is in spite of ollama list detecting the model.

Specifically I ran

curl $LLAMA_URL -d '{
  "model": "llama3",      
  "messages": [
    { "role": "user", "content": "why is the sky blue?" }
  ]
}'

OS

Linux

GPU

No response

CPU

No response

Ollama version

0.1.32

@lambdaofgod lambdaofgod added the bug Something isn't working label Apr 24, 2024
@EverThingy
Copy link

EverThingy commented Apr 25, 2024

Can you try adding a tag to the model parameter (by default it's latest)

curl $LLAMA_URL -d '{
  "model": "llama3:latest",      
  "messages": [
    { "role": "user", "content": "why is the sky blue?" }
  ]
}'

@lambdaofgod
Copy link
Author

That's weird, that was the first thing I tried and it didn't work. Must have been fixed by an update.
Thanks!

@sridvijay
Copy link

sridvijay commented Apr 26, 2024

I'm still running into the same issue. I've tried llama3 as well as llama3:latest and still get the same error:

model 'llama3' not found, try pulling it first.

When I run ollama list, I do see the model:

ollama list
NAME            ID              SIZE    MODIFIED    
llama3:latest   a6990ed6be41    4.7 GB  5 hours ago

I've tried reinstalling ollama as well, but no luck either :/

This is running on Ubuntu 20.04 running ollama 0.1.32

EDIT:

I solved it by running curl -X POST http://localhost:[port]/api/pull -d '{"model":"llama3"}'.

Not sure what I got wrong here, as I thought doing ollama pull llama3 in the command line would result in the same thing, but I guess the models are stored separately if you want to serve? Not sure, but it works!

@lambdaofgod
Copy link
Author

@sridvijay this basically solves it, thanks! I've updated FAQ accordingly in #3936

@m-aliabbas
Copy link

same error. not solved

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

4 participants