Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: Support WasmEdge and its ggml plugin on ARM NPU #3411

Open
alabulei1 opened this issue May 17, 2024 · 2 comments
Open

feat: Support WasmEdge and its ggml plugin on ARM NPU #3411

alabulei1 opened this issue May 17, 2024 · 2 comments
Labels
c-WASI-NN enhancement New feature or request help wanted Extra attention is needed

Comments

@alabulei1
Copy link
Contributor

Summary

One of the advantages of using WasmEdge as the LLM inference runtime is that WasmEdge is portable across different CPUs and GPUs. So it's important to support more chips for WasmEdge.

ARM NPU chip is a popular AI processor that WasmEdge should support.

Details

Support running LLM inference with WasmEdge on ARM NPU

Appendix

No response

@alabulei1 alabulei1 added enhancement New feature or request help wanted Extra attention is needed c-WASI-NN labels May 17, 2024
@Wck-iipi
Copy link
Contributor

I would like to work on this issue. If you can guide me with references I would be grateful thanks.

@hangedfish
Copy link
Collaborator

I think we can start with Rockchip RK3588 SOC, which is a popular chip recently. It supports 32GB of memory, which is enough for LLM. There are also a large number of SBC (Single-board computers) products that can be tested, such as Radxa ROCK 5B/5C and Orange Pi 5 Plus.

https://github.com/airockchip/rknn-toolkit2

Good luck

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
c-WASI-NN enhancement New feature or request help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

3 participants