You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
i got this error
"E:\phi3-mini-128k-gguf\model>ollama create phi-3-mini-128k -f Modelfile
transferring model data
creating model layer
Error: invalid file magic"
here is the modelfile used
FROM ./Phi-3-mini-128k-instruct.IQ4_XS.gguf
Hi @atb29 sorry about this issue, this error is due to the fact you are trying to load phi with IQ quantization, which isn't supported yet.
It looks like there are some other open issues to get Ollama to support IQ quantization, so I'm going to resolve this issue for now to keep things organized around #3622. Please follow that issue to keep up to date with when we get the IQ support in.
What is the issue?
i got this error
"E:\phi3-mini-128k-gguf\model>ollama create phi-3-mini-128k -f Modelfile
transferring model data
creating model layer
Error: invalid file magic"
here is the modelfile used
FROM ./Phi-3-mini-128k-instruct.IQ4_XS.gguf
PARAMETER num_ctx 65536
PARAMETER num_keep 4
PARAMETER stop <|user|>
PARAMETER stop <|assistant|>
PARAMETER stop <|system|>
PARAMETER stop <|end|>
PARAMETER stop <|endoftext|>
TEMPLATE """
{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|end|>
{{ end }}<|assistant|>
{{ .Response }}<|end|>
"""
OS
Windows
GPU
No response
CPU
No response
Ollama version
0.1.31
The text was updated successfully, but these errors were encountered: