issues Search Results · repo:Mozilla-Ocho/llamafile language:C++
Filter by
474 results
(71 ms)474 results
inMozilla-Ocho/llamafile (press backspace or delete to remove)Contact Details
tridu33@qq.com
What happened?
download and compile in Ubuntu20.04, aarch64
#!/bin/bash
/usr/bin/g++ -Dcpuinfer_ext_EXPORTS -march=armv8-a \
-I/home/tridu33/ktransformers/ktransformers_ext/../../third_party ...
bug
critical severity
Tridu33
- 7
- Opened 5 days ago
- #706
Contact Details
No response
What happened?
A couple of weeks ago I was looking for a similar option as llamafile server for whisperfile but I couldn t find it in
the whisperfile documentation. Then ...
bug
low severity
alonsosilvaallende
- 1
- Opened 7 days ago
- #704
Contact Details
No response
What happened?
When I run the whisperfile server:
./whisperfile-0.9.0 -m ggml-tiny.en.bin --gpu nvidia
and then on another terminal:
curl 127.0.0.1:8080/inference -H ...
bug
high severity
alonsosilvaallende
- 1
- Opened 8 days ago
- #703
What happened?
./Llama-3.2-1B-Instruct.Q6_K.llamafile --server --embedding --nobrowser
error: unknown argument: --nobrowser
Without passing --nobrowser
/Llama-3.2-1B-Instruct.Q6_K.llamafile --server ...
bug
critical severity
harikt
- 1
- Opened 8 days ago
- #702
Contact Details
nathan@mozilla.ai
What happened?
I m using LiteLLM to inference Llamafile in Lumigator
(https://github.com/mozilla-ai/lumigator/blob/main/lumigator/jobs/inference/model_clients.py#L65) ...
bug
medium severity
njbrake
- Opened 9 days ago
- #701
Hi guys,
I m trying to run this
$ llamafile -m qwen2.5-coder-7b-instruct-q8_0.gguf --gpu NVIDIA
█...╝
launching server...
error: Uncaught SIGSEGV (SEGV_MAPERR) at 0x328 on legion pid 249503 tid 249511 ...
betim
- 1
- Opened 11 days ago
- #700
Contact Details
tridu33@qq.com
What happened?
i download GGUF files DeepSeek-R1-Q4_K_M from https://www.modelscope.cn/models/unsloth/DeepSeek-R1-GGUF/files , merge
them into one single file:
/home/tridu33/workspace/llama.cpp4Ascend/build/bin/llama-gguf-split ...
bug
critical severity
Tridu33
- 4
- Opened 12 days ago
- #699
Contact Details
stonez56@gmail.com
What happened?
I would like to include my own system prompt file when start the Llamafile. I have tried: -spf FNAME
--system-prompt-file FNAME both won t work.
How ...
bug
low severity
stonez56
- Opened 19 days ago
- #698
Contact Details
No response
What happened?
I can t use the --lora parameter. I expected it to work, but maybe I’m specifying the parameters wrong. Has anyone
experienced this issue?
Version
llamafile ...
bug
high severity
michaelact
- 1
- Opened 19 days ago
- #697
Contact Details
No response
What happened?
Segmentation Fault during CUDA Initialization with GPU Offloading Enabled
Description:
When running the binary with GPU offloading enabled (e.g., using -ngl ...
bug
critical severity
FordUniver
- 5
- Opened 20 days ago
- #696

Learn how you can use GitHub Issues to plan and track your work.
Save views for sprints, backlogs, teams, or releases. Rank, sort, and filter issues to suit the occasion. The possibilities are endless.Learn more about GitHub IssuesProTip!
Restrict your search to the title by using the in:title qualifier.
Learn how you can use GitHub Issues to plan and track your work.
Save views for sprints, backlogs, teams, or releases. Rank, sort, and filter issues to suit the occasion. The possibilities are endless.Learn more about GitHub IssuesProTip!
Press the /
key to activate the search input again and adjust your query.