Skip to content

Sync master with upstream release b7951#415

Merged
jan-service-account merged 11 commits intodevfrom
update-dev-from-master-2026-02-06-00-44
Feb 6, 2026
Merged

Sync master with upstream release b7951#415
jan-service-account merged 11 commits intodevfrom
update-dev-from-master-2026-02-06-00-44

Conversation

@jan-service-account
Copy link
Copy Markdown

Updates dev branch with latest release (b7951) from ggml-org/llama.cpp

CISC and others added 11 commits February 5, 2026 02:27
* add missing llama_add_compile_flags

* disable all warnings for ssl, crypto and fipsmodule
* vulkan: fix GPU deduplication logic.

As reported in ggml-org#19221, the
(same uuid, same driver) logic is problematic for windows+intel igpu.

Let's just avoid filtering for MoltenVK which is apple-specific, and
keep the logic the  same as before 88d23ad - just dedup based on UUID.

Verified that MacOS + 4xVega still reports 4 GPUs with this version.

* vulkan: only skip dedup when both drivers are moltenVk
Signed-off-by: Adrien Gallouët <angt@huggingface.co>
* bench : update script

* benches : update numbers
…org#19281)

Write out a 2-bit code per block and avoid loading the mask when it
matches these two common cases.

Apply this optimization when the mask is relatively large (i.e. prompt
processing).
@jan-service-account jan-service-account merged commit 9b083a5 into dev Feb 6, 2026
3 checks passed
@jan-service-account jan-service-account deleted the update-dev-from-master-2026-02-06-00-44 branch February 6, 2026 00:45
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants