This repository contains the code used for training/fine-tuning the models introduced in the paper "MiniALBERT: Model Distillation via Parameter-Efficient Recursive Transformers".
-
Updated
Mar 26, 2024 - Python
This repository contains the code used for training/fine-tuning the models introduced in the paper "MiniALBERT: Model Distillation via Parameter-Efficient Recursive Transformers".
Add a description, image, and links to the minialbert topic page so that developers can more easily learn about it.
To associate your repository with the minialbert topic, visit your repo's landing page and select "manage topics."