-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
first install, first try : error raised #54
Comments
Remove ellama-provider from your configuration or pull "mistral:7b-instruct-v0.2-q6/K" |
Ok. Just fyi I used the exact install use package from the GitHub readme.
(use-package ellama
:init
(setopt ellama-language "German")
(require 'llm-ollama)
(setopt ellama-provider
(make-llm-ollama
:chat-model "mistral:7b-instruct-v0.2-q6/K"
:embedding-model "mistral:7b-instruct-v0.2-q6/K"))
;; Predefined llm providers for interactive switching.
;; You shouldn't add ollama providers here - it can be selected
interactively
;; without it. It is just example.
(setopt ellama-providers
'(("zephyr" . (make-llm-ollama
:chat-model "zephyr:7b-beta-q6_K"
:embedding-model "zephyr:7b-beta-q6_K"))
("mistral" . (make-llm-ollama
:chat-model
"mistral:7b-instruct-v0.2-q6_K"
:embedding-model
"mistral:7b-instruct-v0.2-q6_K"))
("mixtral" . (make-llm-ollama
:chat-model
"mixtral:8x7b-instruct-v0.1-q3/K/M-4k"
:embedding-model
"mixtral:8x7b-instruct-v0.1-q3/K/M-4k")))))
…On Mon, 22 Jan 2024, 19:54 Sergey Kostyaev, ***@***.***> wrote:
Remove ellama-provider from your configuration or pull
"mistral:7b-instruct-v0.2-q6/K"
—
Reply to this email directly, view it on GitHub
<#54 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AACYTC2KCIICELQ4GT2FAKLYP2YW5AVCNFSM6AAAAABCFWDKICVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSMBUGYYTAMJQG4>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
I think you read it before copy pasting 😃 |
I have added more comments about it into readme. You are not the first person with this exact problem. |
Sergey Kostyaev ***@***.***> writes:
Remove ellama-provider from your configuration or pull "mistral:7b-instruct-v0.2-q6/K"
I'm sorry but I dont really understand what you mean. I did read the
readme, but I dont understand. By "pull mistral" you mean to delete it
from the ellama-providers? Not pull from github?
|
I mean call in your terminal: ollama pull mistral:7b-instruct-v0.2-q6_K # here was also typo |
Or remove all ellama configuration from your config file and restart emacs - all should works fine by default if you have ollama installed and zephyr pulled. |
FIY I just deleted all lines with provider in and it worked. (use-package ellama |
Even that lines can be removed 🙂 |
FYI I did the ollama pull you suggested
ollama pull mistral:7b-instruct-v0.2-q6_K
and it still "didnt work" - I checked the "mistral:7b-instruct-v0.2-q6_K"
in the config and all seemed to look ok.
I dont know enough about this and dont want to waste your time, but I'll
come back to it and provide an update if I can.
Thank you for your help.
…On Mon, 22 Jan 2024 at 20:40, Sergey Kostyaev ***@***.***> wrote:
FIY I just deleted all lines with provider in and it worked.
(use-package ellama :init (setopt ellama-language "German") (require
'llm-ollama))
Even that lines can be removed 🙂
—
Reply to this email directly, view it on GitHub
<#54 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AACYTC6PEZVA54PQ43LV263YP26CJAVCNFSM6AAAAABCFWDKICVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSMBUGY4DEMZSHE>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
To work without configuration it should be: ollama pull zephyr in your terminal. If you use: ollama pull mistral:7b-instruct-v0.2-q6_K You need this ellama configuration: (use-package ellama
:init
(require 'llm-ollama)
(setopt ellama-provider
(make-llm-ollama
;; this model should be pulled to use it
;; value should be the same as you print in terminal during pull
:chat-model "mistral:7b-instruct-v0.2-q6_K"
:embedding-model "mistral:7b-instruct-v0.2-q6_K"))) Check updated readme. |
Close this issue if your setup works, or feel free to go for advice here. |
Closing this due to inactivity. @rileyrg feel free to reopen it. |
I installed ollama and had that pull zephyr.
When I "C-c e a i" to ask Ella something:- I type in a question, hit enter and ....
Debugger entered--Lisp error: (wrong-number-of-arguments (2 . 2) 1)
#f(compiled-function (_ msg) #<bytecode -0x39ad22ed7423a08>)("Unknown error calling ollama")
llm-request-callback-in-buffer(#<buffer ellama when is christmas? (mistral:7b-instruct-v0.2-q6/K).org> #f(compiled-function (_ msg) #<bytecode -0x39ad22ed7423a08>) "Unknown error calling ollama")
#f(compiled-function (_ ) #<bytecode 0x1a7d3104803e9095>)(404 ((error . "model 'mistral:7b-instruct-v0.2-q6/K' not found, t...")))
#f(compiled-function ( on-success on-error) #<bytecode -0x46530f480170ce3>)((:error (error http 404)) nil #f(compiled-function (_ _) #<bytecode 0x1a7d3104803e9095>))
url-http-activate-callback()
url-http-content-length-after-change-function(125 207 82)
url-http-wait-for-headers-change-function(1 212 211)
url-http-generic-filter(# "HTTP/1.1 404 Not Found\15\nContent-Type: application/...")
I havent tried to debug or anything, just in case this is obvious to someone here. Using consult btw.
The text was updated successfully, but these errors were encountered: