There are thousands of publications on making AI more energy and time efficient. We subsume these proposals as sustainability measures for now.
From an AI security perspective: Can a bad actor nullify these optimizations for fun (and profit)? We call this an energy-latency attack. This is a novel angle on AI security which will become very relevant very soon. Sponge Examples: Energy-Latency Attacks on Neural Networks will give you a good idea on energy-latency attacks.
The objective of this repo is to match sustainability measures to existing attacks. This way, we can identify trends and possibly blank spots.
For now, this is a collection of papers and articles with personal annotations. It can already help you find sources on (the robustness of) sustainable AI.
Feel free to star and fork. Contact me if you are curious (See my Github bio).
No sense in listing them. There are good repos for that.
- Awesome-Mixture-of-Experts-Papers - Collection of MoE papers
- A collection of AWESOME things about mixture-of-experts
| Publication | Venue/Journal | Notes |
|---|---|---|
| Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning | Prequisite for understanding PEFT | |
| LoRA: Low-rank adaptation of large language models | ||
| Few-shot parameter-efficient finetuning is better and cheaper than in-context learning | ||
| ComPEFT: Compression for Communicating Parameter Efficient Updates via Sparsification and Quantization | Author: Collin Raffel |
Too much to list them all for now. refer to attacks on them.
| Publication | Venue/Journal | Notes |
|---|---|---|
| Dynamic Neural Networks: A Survey | Good overview of all techniques in Figure 1 |
| Publication | Venue/Journal | Notes |
|---|---|---|
| Dynamic Channel Pruning: Feature Boosting and Suppression | ICLR 2019 | |
| Channel gating neural networks | NeurIPS 2019 |
| Publication | Venue/Journal | Notes |
|---|---|---|
| Speculative Decoding with Big Little Decoder | NeurIPS Poster 2023 | |
| Big Little Transformer Decoder |
| Publication | Venue/Journal | Notes |
|---|---|---|
| on-the-fly, on-chip latency predictors for Edge TPUs |
| Publication | Venue/Journal | Notes |
|---|---|---|
| Sponge Examples: Energy-Latency Attacks on Neural Networks | Euro S&P 2021 | First paper to define the goal of energy-latency attacks very clearly |
| Energy-Latency Attacks via Sponge Poisoning |
| Publication | Venue/Journal | Notes |
|---|---|---|
| Sponge Examples: Energy-Latency Attacks on Neural Networks | Euro S&P 2021 | First paper to define the goal of energy-latency attacks very clearly |
| Bad Characters: Imperceptible NLP Attacks | S&P 2022 | E-L attacks are hard to do imperceptibly |
| NMTSloth: understanding and testing efficiency degradation of neural machine translation systems |