Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.
-
Updated
Apr 13, 2026 - Python
Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.
Local AI workstation — discover, run, chat, benchmark, and generate images from open-weight models. DFlash/DDTree speculative decoding, five cache compression strategies (RotorQuant, TriAttention, TurboQuant, ChaosEngine), MLX + llama.cpp + vLLM backends.
Add a description, image, and links to the dflash topic page so that developers can more easily learn about it.
To associate your repository with the dflash topic, visit your repo's landing page and select "manage topics."