Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

⬆️ Update ggerganov/llama.cpp #1750

Merged
merged 1 commit into from
Feb 23, 2024

Conversation

localai-bot
Copy link
Contributor

Bump of ggerganov/llama.cpp version

Signed-off-by: GitHub <noreply@github.com>
Copy link

netlify bot commented Feb 23, 2024

Deploy Preview for localai canceled.

Name Link
🔨 Latest commit 7378af0
🔍 Latest deploy log https://app.netlify.com/sites/localai/deploys/65d8fac2dc23b500083a1423

@mudler mudler merged commit ff88c39 into mudler:master Feb 23, 2024
19 checks passed
truecharts-admin added a commit to truecharts/charts that referenced this pull request Feb 24, 2024
….0 by renovate (#18546)

This PR contains the following updates:

| Package | Update | Change |
|---|---|---|
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
minor | `v2.8.2-cublas-cuda11-ffmpeg-core` ->
`v2.9.0-cublas-cuda11-ffmpeg-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
minor | `v2.8.2-cublas-cuda11-core` -> `v2.9.0-cublas-cuda11-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
minor | `v2.8.2-cublas-cuda12-ffmpeg-core` ->
`v2.9.0-cublas-cuda12-ffmpeg-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
minor | `v2.8.2-cublas-cuda12-core` -> `v2.9.0-cublas-cuda12-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
minor | `v2.8.2-ffmpeg-core` -> `v2.9.0-ffmpeg-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
minor | `v2.8.2` -> `v2.9.0` |

---

> [!WARNING]
> Some dependencies could not be looked up. Check the Dependency
Dashboard for more information.

---

### Release Notes

<details>
<summary>mudler/LocalAI (docker.io/localai/localai)</summary>

### [`v2.9.0`](https://togithub.com/mudler/LocalAI/releases/tag/v2.9.0)

[Compare
Source](https://togithub.com/mudler/LocalAI/compare/v2.8.2...v2.9.0)

This release brings many enhancements, fixes, and a special thanks to
the community for the amazing work and contributions!

We now have sycl images for Intel GPUs, ROCm images for AMD GPUs,and
much more:

- You can find the AMD GPU images tags between the container images
available - look for `hipblas`. For example,
[master-hipblas-ffmpeg-core](https://quay.io/repository/go-skynet/local-ai/tag/master-hipblas-ffmpeg-core).
Thanks to [@&#8203;fenfir](https://togithub.com/fenfir) for this nice
contribution!
- Intel GPU images are tagged with `sycl`. You can find images with two
flavors, sycl-f16 and sycl-f32 respectively. For example,
[master-sycl-f16](https://quay.io/repository/go-skynet/local-ai/tag/master-sycl-f16-core).
Work is in progress to support also diffusers and transformers on Intel
GPUs.
- Thanks to [@&#8203;christ66](https://togithub.com/christ66) first
efforts in supporting the Assistant API were made, and we are planning
to support the Assistant API! Stay tuned for more!
- Now LocalAI supports the Tools API endpoint - it also supports the
(now deprecated) functions API call as usual. We now also have support
for SSE with function calling. See
[mudler/LocalAI#1726
for more
- Support for Gemma models - did you hear? Google released OSS models
and LocalAI supports it already!
- Thanks to [@&#8203;dave-gray101](https://togithub.com/dave-gray101) in
[mudler/LocalAI#1728
to put efforts in refactoring parts of the code - we are going to
support soon more ways to interface with LocalAI, and not only restful
api!

##### Support the project

First off, a massive thank you to each and every one of you who've
chipped in to squash bugs and suggest cool new features for LocalAI.
Your help, kind words, and brilliant ideas are truly appreciated - more
than words can say!

And to those of you who've been heros, giving up your own time to help
out fellow users on Discord and in our repo, you're absolutely amazing.
We couldn't have asked for a better community.

Just so you know, LocalAI doesn't have the luxury of big corporate
sponsors behind it. It's all us, folks. So, if you've found value in
what we're building together and want to keep the momentum going,
consider showing your support. A little shoutout on your favorite social
platforms using [@&#8203;LocalAI_OSS](https://twitter.com/LocalAI_API)
and [@&#8203;mudler_it](https://twitter.com/mudler_it) or joining our
sponsorship program can make a big difference.

Also, if you haven't yet joined our Discord, come on over! Here's the
link: https://discord.gg/uJAeKSAGDy

Every bit of support, every mention, and every star adds up and helps us
keep this ship sailing. Let's keep making LocalAI awesome together!

Thanks a ton, and here's to more exciting times ahead with LocalAI! 🚀

##### What's Changed

##### Bug fixes 🐛

- Add TTS dependency for cuda based builds fixes
[#&#8203;1727](https://togithub.com/mudler/LocalAI/issues/1727) by
[@&#8203;blob42](https://togithub.com/blob42) in
[mudler/LocalAI#1730

##### Exciting New Features 🎉

- Build docker container for ROCm by
[@&#8203;fenfir](https://togithub.com/fenfir) in
[mudler/LocalAI#1595
- feat(tools): support Tool calls in the API by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#1715
- Initial implementation of upload files api. by
[@&#8203;christ66](https://togithub.com/christ66) in
[mudler/LocalAI#1703
- feat(tools): Parallel function calling by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#1726
- refactor: move part of api packages to core by
[@&#8203;dave-gray101](https://togithub.com/dave-gray101) in
[mudler/LocalAI#1728
- deps(llama.cpp): update, support Gemma models by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#1734

##### 👒 Dependencies

- deps(llama.cpp): update by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#1714
- ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#1740

##### Other Changes

- ⬆️ Update docs version mudler/LocalAI by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#1718
- ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#1705
- Update README.md by
[@&#8203;lunamidori5](https://togithub.com/lunamidori5) in
[mudler/LocalAI#1739
- ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#1750

##### New Contributors

- [@&#8203;fenfir](https://togithub.com/fenfir) made their first
contribution in
[mudler/LocalAI#1595
- [@&#8203;christ66](https://togithub.com/christ66) made their first
contribution in
[mudler/LocalAI#1703
- [@&#8203;blob42](https://togithub.com/blob42) made their first
contribution in
[mudler/LocalAI#1730

**Full Changelog**:
mudler/LocalAI@v2.8.2...v2.9.0

</details>

---

### Configuration

📅 **Schedule**: Branch creation - "before 10pm on monday" in timezone
Europe/Amsterdam, Automerge - At any time (no schedule defined).

🚦 **Automerge**: Enabled.

♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the
rebase/retry checkbox.

🔕 **Ignore**: Close this PR and you won't be reminded about these
updates again.

---

- [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check
this box

---

This PR has been generated by [Renovate
Bot](https://togithub.com/renovatebot/renovate).

<!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzNy4yMTMuMCIsInVwZGF0ZWRJblZlciI6IjM3LjIxMy4wIiwidGFyZ2V0QnJhbmNoIjoibWFzdGVyIn0=-->
@olariuromeo
Copy link

olariuromeo commented Feb 25, 2024

Building images with this change get error, instead, the old version works perfectly. With the current version it is not possible to prepare the backend or generate images:

make clean
make GO_TAGS=stablediffusion,tts BUILD_TYPE=cuBLAS BUILD_GRPC_FOR_BACKEND_LLAMA=true build
mkdir -p backend-assets
LLAMA_VERSION=201294ae177b308fb3a99dc504dd6d27e8afa907 make -C backend/cpp/llama llama.cpp	
make[1]: Entering directory '/home/user/apps/localai/backend/cpp/llama'
git clone --recurse-submodules https://github.com/ggerganov/llama.cpp llama.cpp
Cloning into 'llama.cpp'...
remote: Enumerating objects: 19344, done.
remote: Counting objects: 100% (3/3), done.
remote: Compressing objects: 100% (3/3), done.
remote: Total 19344 (delta 0), reused 1 (delta 0), pack-reused 19341
Receiving objects: 100% (19344/19344), 22.71 MiB | 11.65 MiB/s, done.
Resolving deltas: 100% (13524/13524), done.
Submodule 'kompute' (https://github.com/nomic-ai/kompute.git) registered for path 'kompute'
Cloning into '/home/user/apps/localai/backend/cpp/llama/llama.cpp/kompute'...
remote: Enumerating objects: 9078, done.        
remote: Counting objects: 100% (213/213), done.        
remote: Compressing objects: 100% (125/125), done.        
remote: Total 9078 (delta 94), reused 167 (delta 78), pack-reused 8865        
Receiving objects: 100% (9078/9078), 17.58 MiB | 10.31 MiB/s, done.
Resolving deltas: 100% (5701/5701), done.
remote: Enumerating objects: 1, done.
remote: Counting objects: 100% (1/1), done.
remote: Total 1 (delta 0), reused 1 (delta 0), pack-reused 0
Unpacking objects: 100% (1/1), 213 bytes | 213.00 KiB/s, done.
From https://github.com/nomic-ai/kompute
 * branch            4565194ed7c32d1d2efa32ceab4d3c6cae006306 -> FETCH_HEAD
Submodule path 'kompute': checked out '4565194ed7c32d1d2efa32ceab4d3c6cae006306'
if [ -z "201294ae177b308fb3a99dc504dd6d27e8afa907" ]; then \
	exit 1; \
fi
cd llama.cpp && git checkout -b build 201294ae177b308fb3a99dc504dd6d27e8afa907 && git submodule update --init --recursive --depth 1
Switched to a new branch 'build'
make[1]: Leaving directory '/home/user/apps/localai/backend/cpp/llama'
git clone --recurse-submodules https://github.com/go-skynet/go-llama.cpp sources/go-llama
Cloning into 'sources/go-llama'...
remote: Enumerating objects: 769, done.
remote: Counting objects: 100% (412/412), done.
remote: Compressing objects: 100% (166/166), done.
remote: Total 769 (delta 287), reused 317 (delta 234), pack-reused 357
Receiving objects: 100% (769/769), 237.45 KiB | 2.05 MiB/s, done.
Resolving deltas: 100% (445/445), done.
Submodule 'llama.cpp' (https://github.com/ggerganov/llama.cpp) registered for path 'llama.cpp'
Cloning into '/home/user/apps/localai/sources/go-llama/llama.cpp'...
remote: Enumerating objects: 19344, done.        
remote: Counting objects: 100% (3/3), done.        
remote: Compressing objects: 100% (3/3), done.        
remote: Total 19344 (delta 0), reused 1 (delta 0), pack-reused 19341        
Receiving objects: 100% (19344/19344), 22.71 MiB | 10.49 MiB/s, done.
Resolving deltas: 100% (13524/13524), done.
Submodule path 'llama.cpp': checked out 'ac43576124a75c2de6e333ac31a3444ff9eb9458'
cd sources/go-llama && git checkout -b build aeba71ee842819da681ea537e78846dc75949ac0 && git submodule update --init --recursive --depth 1
Switched to a new branch 'build'
git clone --recurse-submodules https://github.com/go-skynet/go-llama.cpp sources/go-llama-ggml
Cloning into 'sources/go-llama-ggml'...
remote: Enumerating objects: 769, done.
remote: Counting objects: 100% (437/437), done.
remote: Compressing objects: 100% (172/172), done.
remote: Total 769 (delta 311), reused 336 (delta 253), pack-reused 332
Receiving objects: 100% (769/769), 234.34 KiB | 1.71 MiB/s, done.
Resolving deltas: 100% (448/448), done.
Submodule 'llama.cpp' (https://github.com/ggerganov/llama.cpp) registered for path 'llama.cpp'
Cloning into '/home/user/apps/localai/sources/go-llama-ggml/llama.cpp'...
remote: Enumerating objects: 19344, done.        
remote: Counting objects: 100% (3/3), done.        
remote: Compressing objects: 100% (3/3), done.        
remote: Total 19344 (delta 0), reused 1 (delta 0), pack-reused 19341        
Receiving objects: 100% (19344/19344), 22.49 MiB | 11.36 MiB/s, done.
Resolving deltas: 100% (13526/13526), done.
Submodule path 'llama.cpp': checked out 'ac43576124a75c2de6e333ac31a3444ff9eb9458'
cd sources/go-llama-ggml && git checkout -b build fd43d66f46ee3b5345fb8a74a252d86ccd34a409 && git submodule update --init --recursive --depth 1
fatal: reference is not a tree: fd43d66f46ee3b5345fb8a74a252d86ccd34a409
make: *** [Makefile:241: sources/go-llama-ggml] Error 128

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants