Skip to content

Commit

Permalink
First Model Compression Cards
Browse files Browse the repository at this point in the history
  • Loading branch information
nassimberrada authored and nassimberrada committed Nov 7, 2023
1 parent f618d6e commit 72e404d
Show file tree
Hide file tree
Showing 8 changed files with 179 additions and 34 deletions.
27 changes: 27 additions & 0 deletions compression/aimet.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
aimet:

name: "AIMET"

image_url: https://www.pillar.vc/wp-content/uploads/2021/11/Dark-blue-1600x522.png

tags:
- Pruning
- Tensorization
- Quantization
- PyTorch
- TensorFlow
- ONNX
- Open-Source

url: https://github.com/quic/aimet

description: "AIMET is a library that provides advanced model quantization and compression techniques for
trained neural network models. It provides features that have been proven to improve run-time performance of deep learning
neural network models with lower compute and memory requirements and minimal impact to task accuracy."

features:
- "Batch-Norm Folding"
- "Cross-Layer Equalization"
- "AdaRound"
- "Channel Pruning"
- "Singular Vector Decomposition"
23 changes: 23 additions & 0 deletions compression/bigdl.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,23 @@
bigdl:

name: "BigDL"

# URL of relevant image (e.g. logo)
image_url: https://www.intel.com/content/dam/www/central-libraries/us/en/images/2022-09/bigdl-logo-dark-rwd.png

tags:
- Intel
- PyTorch
- Open-Source

url: https://github.com/intel/neural-compressor

description: "BigDL is a set of toolkits developed by Intel to seamlessly scales data analytics & AI applications to be run easily on both laptops to cloud infrastructures.
The project introduces the following set of libraries:
(LLM)[https://github.com/intel-analytics/BigDL/tree/main/python/llm]: Low-bit (INT3/INT4/INT5/INT8) large language model library for Intel CPU/GPU
(Orca)[https://bigdl.readthedocs.io/en/latest/doc/Orca/index.html]: Distributed Big Data & AI (TF & PyTorch) Pipeline on Spark and Ray
(Nano)[https://bigdl.readthedocs.io/en/latest/doc/Nano/index.html]: Transparent Acceleration of Tensorflow & PyTorch Programs on Intel CPU/GPU
(DLlib)[https://bigdl.readthedocs.io/en/latest/doc/DLlib/index.html]: “Equivalent of Spark MLlib” for Deep Learning
(Chronos)[https://bigdl.readthedocs.io/en/latest/doc/Chronos/index.html]: Scalable Time Series Analysis using AutoML
(Friesian)[https://bigdl.readthedocs.io/en/latest/doc/Friesian/index.html]: End-to-End Recommendation Systems
(PPML)[https://bigdl.readthedocs.io/en/latest/doc/PPML/index.html]: Secure Big Data and AI (with SGX Hardware Security)"
34 changes: 0 additions & 34 deletions compression/compression_sample.yaml

This file was deleted.

22 changes: 22 additions & 0 deletions compression/modelcompressiontoolkit.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,22 @@
modelcompressiontoolkit.yaml:

name: "Model Compression Toolkit"

image_url: https://avatars.githubusercontent.com/u/8435219?s=48&v=4

tags:
- Quantization
- SONY
- Open-Source

url: https://github.com/sony/model_optimization

description: "SONY’s Model Compression Toolkit is a model compression tool that focuses on quantization
and comes with a suite of features that make it easier to optimise neural networks for efficient deployment,
including synthetic image data generation and visualisation tools."

features:
- "Synthetic Image Data Generation"
- "Power-of-Two Quantization"
- "Symmetric Quantization"
- "Enhanced Post Training Quantization"
29 changes: 29 additions & 0 deletions compression/neuralcompressor.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,29 @@
neuralcompressor:

name: "Neural Compressor"

# URL of relevant image (e.g. logo)
image_url: https://avatars.githubusercontent.com/u/17888862?s=48&v=4

tags:
- Pruning
- Distillation
- Quantization
- PyTorch
- TensorFlow
- ONNX
- MxNet
- Open-Source

url: https://github.com/intel/neural-compressor

description: "Intel® Neural Compressor aims to provide popular model compression techniques such
as quantization, pruning (sparsity), distillation, and neural architecture search on mainstream frameworks
such as TensorFlow, PyTorch, ONNX Runtime, and MXNet, as well as Intel extensions such as Intel Extension
for TensorFlow and Intel Extension for PyTorch."

features:
- "Unified scikit-learn-like API"
- "Accuracy Aware Tuning"
- "AWQ"
- "GPTQ"
25 changes: 25 additions & 0 deletions compression/neuralnetworkintelligence.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
neuralnetworkintelligence:

name: "Neural Network Intelligence"

image_url: https://github.com/microsoft/nni/raw/master/docs/img/nni_logo.png

tags:
- Pruning
- Quantization
- Microsoft
- Open-Source

url: https://github.com/quic/aimet

description: "NNI automates feature engineering, neural architecture search, hyperparameter tuning,
and model compression for deep learning."

features:
- "DoReFa Quantization"
- "BNN Quantization"
- "LSQ Quantization"
- "Ln Norm Pruning"
- "Slim Pruning"
- "FPGM Pruning"
- "AGP Pruning"
27 changes: 27 additions & 0 deletions compression/sparseml.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
sparseml:

name: "SparseML"

image_url: https://www.pillar.vc/wp-content/uploads/2021/11/Dark-blue-1600x522.png

tags:
- Pruning
- Distillation
- Quantization
- PyTorch
- TensorFlow
- Keras
- Open-Source

url: https://docs.neuralmagic.com/sparseml/

description: "SparseML is an open source library developed and maintained by (Neural Magic)[https://neuralmagic.com/] for applying compression recipes to neural networks.
Currently, it supports pruning, quantization and knowledge distillation for compressing Vision, NLP, and now, large language models as well.
SparseML also provides pre-compressed and pre-quantized models in their SparseZoo."

features:
- "Gradual Magnitude Pruning (GMP)"
- "Alternating Compressed/DeCompressed Pruning (AC/DC)"
- "Optimal BERT Surgeon (oBERT)"
- "Pre-Compressed models via SparseZoo"
- "Knowledge Distillation Recipes"
26 changes: 26 additions & 0 deletions compression/tensorly.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
tensorly:

name: "TensorLy"

image_url: http://tensorly.org/stable/_static/tensorly-logo.png

tags:
- Tensorization
- PyTorch
- NumPy
- TensorFlow
- MxNet
- CuPy
- Jax
- Open-Source

url: https://github.com/intel/neural-compressor

description: "TensorLy is a Python library that aims at making tensor learning simple and accessible. It allows simple performing of tensor decomposition,
tensor learning and tensor algebra. Its backend system allows to seamlessly perform computation with NumPy,
PyTorch, JAX, MXNet, TensorFlow or CuPy, and run methods at scale on CPU or GPU."

features:
- "Tucker Decomposition"
- "Canonical Polyadic Decomposition"
- "Tensor Train Decomposition"

0 comments on commit 72e404d

Please sign in to comment.