Run 70B+ LLMs on a single 4GB GPU — no quantization required.
-
Updated
Feb 28, 2026 - Python
Run 70B+ LLMs on a single 4GB GPU — no quantization required.
Add a description, image, and links to the airllm topic page so that developers can more easily learn about it.
To associate your repository with the airllm topic, visit your repo's landing page and select "manage topics."