Skip to content

Sync master with upstream release b8992#502

Merged
jan-service-account merged 11 commits intodevfrom
update-dev-from-master-2026-05-01-01-09
May 2, 2026
Merged

Sync master with upstream release b8992#502
jan-service-account merged 11 commits intodevfrom
update-dev-from-master-2026-05-01-01-09

Conversation

@jan-service-account
Copy link
Copy Markdown

Updates dev branch with latest release (b8992) from ggml-org/llama.cpp

petersid2022 and others added 11 commits April 30, 2026 08:18
* port ggml-org#22358 PR to examples/speculative/speculative.cpp
* use vocab_[tgt,dft] instead of ctx_[tgt,dft] when logging on draft
  model / target model vocabulary mismatch

Co-authored-by: Petros Sideris <petros.sideris@nokia.com>
* spec : fix draft model checkpoints

* cont : clean-up

* cont : gate the ngram-mod reset warning behind verbose flag
…22513)

* scripts : add wc2wt.sh - create worktree from current HEAD

Add a script to create a git worktree on a new branch from the current
HEAD. Similar to pr2wt.sh but for local development branches instead of
PRs.

Usage:
  ./scripts/wc2wt.sh gg/new-feature
  ./scripts/wc2wt.sh gg/new-feature "bash -l"

Assisted-by: llama.cpp:local pi

* cont : no need to try to delete the branch
Signed-off-by: Adrien Gallouët <angt@huggingface.co>
* bump ty to 0.0.33

* update typings
* vulkan: add get/set_tensor_2d functions

* fix backend interface comments

* Update ggml/src/ggml-metal/ggml-metal.cpp

Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com>
Signed-off-by: Adrien Gallouët <angt@huggingface.co>
* Update llama-mmap to work with 32-bit wasm and >2GB models

* Update to gguf.cpp style
@jan-service-account jan-service-account merged commit 5cbfb18 into dev May 2, 2026
16 of 17 checks passed
@jan-service-account jan-service-account deleted the update-dev-from-master-2026-05-01-01-09 branch May 2, 2026 01:30
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

10 participants