Skip to content

Conversation

ai-fonsi
Copy link
Contributor

@ai-fonsi ai-fonsi commented Sep 28, 2025

"Integrated" CUDA devices seem to be bugged and produce incorrect output in specific cases. Since disabling the integrated flag seems to neither affect performance nor memory usage on Jetson, I propose disabling the option until the underlying issue is fixed.

Fixes #15034 and probably also #15923.

@ai-fonsi ai-fonsi requested a review from slaren as a code owner September 28, 2025 15:07
@github-actions github-actions bot added Nvidia GPU Issues specific to Nvidia GPUs ggml changes relating to the ggml tensor library for machine learning labels Sep 28, 2025
@slaren
Copy link
Member

slaren commented Sep 28, 2025

This is probably the same synchronization issue with the scheduler that @ggerganov found when making the Metal backend async. To confirm this, can you verify if it works (without this change) by launching with the env variable CUDA_LAUNCH_BLOCKING=1 (effectively disabling async compute)?

@ai-fonsi
Copy link
Contributor Author

I tried starting llama-server with CUDA_LAUNCH_BLOCKING=1, it didn't fix the issue.

@slaren slaren merged commit 9d08828 into ggml-org:master Oct 8, 2025
64 of 67 checks passed
anyshu pushed a commit to anyshu/llama.cpp that referenced this pull request Oct 10, 2025
* master: (113 commits)
  webui: updated the chat service to only include max_tokens in the req… (ggml-org#16489)
  cpu : optimize the ggml NORM operation (ggml-org#15953)
  server : host-memory prompt caching (ggml-org#16391)
  No markdown in cot (ggml-org#16483)
  model-conversion : add support for SentenceTransformers (ggml-org#16387)
  ci: add ARM64 Kleidiai build and test support (ggml-org#16462)
  CANN: Improve ACL graph matching (ggml-org#16166)
  kleidiai: kernel interface refactoring (ggml-org#16460)
  [SYCL] refactor soft_max, add soft_max_back (ggml-org#16472)
  model: EmbeddingGemma Adding Support for SentenceTransformers Dense Modules (ggml-org#16367)
  refactor: centralize CoT parsing in backend for streaming mode (ggml-org#16394)
  Disable CUDA host buffers on integrated GPUs (ggml-org#16308)
  server : fix cancel pending task (ggml-org#16467)
  metal : mark FA blocks (ggml-org#16372)
  server : improve context checkpoint logic (ggml-org#16440)
  ggml webgpu: profiling, CI updates, reworking of command submission (ggml-org#16452)
  llama : support LiquidAI LFM2-MoE hybrid model (ggml-org#16464)
  server : add `/v1/health` endpoint (ggml-org#16461)
  webui : added download action (ggml-org#13552) (ggml-org#16282)
  presets : fix pooling param for embedding models (ggml-org#16455)
  ...
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ggml changes relating to the ggml tensor library for machine learning Nvidia GPU Issues specific to Nvidia GPUs
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Eval bug: Broken/no Gemma 3n output on CUDA (Nvidia Jetson Orin Nano)
2 participants