Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: LLM inference libraries support plan #3124

Open
hydai opened this issue Dec 29, 2023 · 1 comment
Open

feat: LLM inference libraries support plan #3124

hydai opened this issue Dec 29, 2023 · 1 comment
Labels
enhancement New feature or request

Comments

@hydai
Copy link
Member

hydai commented Dec 29, 2023

Summary

There are various LLM inference libraries. WasmEdge already integrated llama.cpp, but we want to bring more to the community.

Details

Already supported:

  1. PyTorch
  2. TFLite
  3. OpenVINO
  4. llama.cpp

The support priority list:

  • Tier 1:
    • burn-rs
  • Tier 2:
    • Intel extension for transformers
    • whisper.cpp
    • RWKV
  • Tier 3:
    • vllm
    • CTranslate2
    • candle
    • mlx

Please feel free to add any comments and suggestions. We would like to hear the voice of the community.
Also, if you are interested in contributing the new LLM inference libraries, please tell and show to us :-)

Happy new year!

Appendix

No response

@hydai hydai added the enhancement New feature or request label Dec 29, 2023
@katopz
Copy link

katopz commented Dec 30, 2023

Cool, i'm now playing with

  • For RAG: llm-chain // just support Qdrant
  • For mac: mlx // still in heavy development
  • For fun: candle // infer tps seem to be a lot slower than llama.cpp atm but it will keep up somehow i hope.
  • For co-pilot: Tabby // Support repos lookup but the output not satisfy me yet

@alabulei1 alabulei1 pinned this issue Jan 2, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants