Skip to content

Bug: Failed to load model #8516

@Fulgurance

Description

@Fulgurance

What happened?

Hi guys. I have got a problem after I compile Llama on my machine. It built properly, but when I try to run it, it is looking for a file don't even exist (a model).

Is it normal ?

Name and Version

version: 0 (unknown)
built with cc (Gentoo Hardened 14.1.1_p20240622 p2) 14.1.1 20240622 for x86_64-pc-linux-gnu

What operating system are you seeing the problem on?

Linux

Relevant log output

zohran@alienware-m17-r3 ~/Downloads/llama.cpp-b3400 $ ./examples/chat.sh
Log start
main: build = 0 (unknown)
main: built with cc (Gentoo Hardened 14.1.1_p20240622 p2) 14.1.1 20240622 for x86_64-pc-linux-gnu
main: seed  = 1721142929
llama_model_load: error loading model: llama_model_loader: failed to load model from ./models/llama-7b/ggml-model-q4_0.gguf

llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model './models/llama-7b/ggml-model-q4_0.gguf'
main: error: unable to load model

zohran@alienware-m17-r3 ~/Downloads/llama.cpp-b3400 $ ls 
AUTHORS                        llama-convert-llama2c-to-ggml  llama-simple
build                          llama-cvector-generator        llama-speculative
ci                             llama-embedding                llama-tokenize
cmake                          llama-eval-callback            llama-train-text-from-scratch
CMakeLists.txt                 llama-export-lora              llama-vdot
CMakePresets.json              llama-finetune                 main
common                         llama-gbnf-validator           main.log
CONTRIBUTING.md                llama-gguf                     Makefile
convert_hf_to_gguf.py          llama-gguf-hash                media
convert_hf_to_gguf_update.py   llama-gguf-split               models
convert_llama_ggml_to_gguf.py  llama-gritlm                   mypy.ini
convert_lora_to_gguf.py        llama-imatrix                  Package.swift
docs                           llama-infill                   pocs
examples                       llama-llava-cli                poetry.lock
flake.lock                     llama-lookahead                prompts
flake.nix                      llama-lookup                   pyproject.toml
ggml                           llama-lookup-create            pyrightconfig.json
gguf-py                        llama-lookup-merge             README.md
grammars                       llama-lookup-stats             requirements
include                        llama-parallel                 requirements.txt
libllava.a                     llama-passkey                  scripts
LICENSE                        llama-perplexity               SECURITY.md
llama-baby-llama               llama-q8dot                    server
llama-batched                  llama-quantize                 spm-headers
llama-batched-bench            llama-quantize-stats           src
llama-bench                    llama-retrieval                tests
llama-benchmark-matmult        llama-save-load-state
llama-cli                      llama-server

Metadata

Metadata

Assignees

No one assigned

    Labels

    bug-unconfirmedcritical severityUsed to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions