SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
-
Updated
Apr 30, 2024 - Python
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
Must read research papers and links to tools and datasets that are related to using machine learning for compilers and systems optimisation
Kernel Tuner
Machine Learning Framework for Operating Systems - Brings ML to Linux kernel
Stretching GPU performance for GEMMs and tensor contractions.
CLTune: An automatic OpenCL & CUDA kernel tuner
Phoebe
Benchmark scripts for TVM
ebpf profiler for jvm
Collective Knowledge crowd-tuning extension to let users crowdsource their experiments (using portable Collective Knowledge workflows) such as performance benchmarking, auto tuning and machine learning across diverse platforms with Linux, Windows, MacOS and Android provided by volunteers. Demo of DNN crowd-benchmarking and crowd-tuning:
A Generic Distributed Auto-Tuning Infrastructure
A GPU benchmark suite for autotuners
Backoff uses an exponential backoff algorithm to backoff between retries with optional auto-tuning functionality.
Autotuner for Spark applications
A pattern-based algorithmic auto-tuner for graph processing on GPUs
MarGotAspect - An AspectC++ code generator for the mARGOt framework
Add a description, image, and links to the auto-tuning topic page so that developers can more easily learn about it.
To associate your repository with the auto-tuning topic, visit your repo's landing page and select "manage topics."