Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ci: [SYCL] Use main GPU and enable sysman #12547

Merged
merged 1 commit into from
Mar 24, 2025
Merged

ci: [SYCL] Use main GPU and enable sysman #12547

merged 1 commit into from
Mar 24, 2025

Conversation

qnixsynapse
Copy link
Collaborator

Set environmental variables for using the main GPU for now and enable SYSMAN for correct GPU memory reporting.

@qnixsynapse qnixsynapse requested a review from ggerganov as a code owner March 24, 2025 13:34
@github-actions github-actions bot added the devops improvements to build systems and github actions label Mar 24, 2025
@qnixsynapse qnixsynapse changed the title ci: Use main GPU and enable sysman ci: [SYCL] Use main GPU and enable sysman Mar 24, 2025
@ggerganov
Copy link
Member

If you add the keyword ggml-ci somewhere in the commit message, it will trigger the CI on this branch, so we can see if it works before merging.

@qnixsynapse
Copy link
Collaborator Author

Ah. It is broken!

@ggerganov
Copy link
Member

It could be related to the shapes of the specific model. Here is how to get and convert the model:

llama.cpp/ci/run.sh

Lines 411 to 421 in 48d7021

# pythia_1.4b
function gg_run_pythia_1_4b {
cd ${SRC}
gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/config.json
gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/tokenizer.json
gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/tokenizer_config.json
gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/raw/main/special_tokens_map.json
gg_wget models-mnt/pythia/1.4B/ https://huggingface.co/EleutherAI/pythia-1.4b/resolve/main/pytorch_model.bin

llama.cpp/ci/run.sh

Lines 435 to 437 in 48d7021

python3 ../convert_hf_to_gguf.py ${path_models} --outfile ${path_models}/ggml-model-f16.gguf

@qnixsynapse
Copy link
Collaborator Author

qnixsynapse commented Mar 24, 2025

@Alcpz @ggerganov Confirmed Q4_0 broken by 08d5986

To quickly test, run llama-cli with this specific model with GGML_SYCL_DISABLE_OPT=1.

@Alcpz
Copy link
Collaborator

Alcpz commented Mar 24, 2025

Thanks for pinpointing this. We may want to disable the GGML_SYCL_DISABLE_OPT path by default until it is fixed.
@qnixsynapse It's probably worth it to start a discussion in the SYCL discussion thread so everyone using the backend is aware.

@qnixsynapse
Copy link
Collaborator Author

We may want to disable the GGML_SYCL_DISABLE_OPT path by default until it is fixed.

Probably in a separate PR. This PR should be merged IMO(It enables the necessary flags).

I will enable SYCL CI on whisper.cpp and ggml repos tomorrow.

@ggerganov ggerganov merged commit c95fa36 into master Mar 24, 2025
10 checks passed
@qnixsynapse qnixsynapse deleted the sycl/ci branch March 25, 2025 02:14
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
devops improvements to build systems and github actions
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants