Skip to content

Annoucing easyquant for speeding up LLM inference via quantization

Latest
Compare
Choose a tag to compare
@peterjc123 peterjc123 released this 31 May 09:41
· 54 commits to main since this release
841294e

With the help of quantization, we could achieve LLM inference efficiently with lower resource usage. Please install the package below and try out the examples here. We look forward to your feedback.