- [2024/04/30] "VMT-Adapter: Parameter-Efficient Transfer Learning for Multi-Task Dense Understanding" code will be released.
- [2024/04/30] "MmAP: Multi-modal Alignment Prompt for Cross-domain Multi-task Learning" code will be released.
- ✅ [2024/03/01] "Visual PEFT Library/Benchmark" repo is created.
If you find our survey and repository useful for your research, please cite it below:
@article{xin2024parameter,
title={Parameter-Efficient Fine-Tuning for Pre-Trained Vision Models: A Survey},
author={Xin, Yi and Luo, Siqi and Zhou, Haodi and Du, Junlong and Liu, Xiaohong and Fan, Yue and Li, Qing and Du, Yuntao},
journal={arXiv preprint arXiv:2402.02242},
year={2024}
}
@inproceedings{xin2024vmt,
title={VMT-Adapter: Parameter-Efficient Transfer Learning for Multi-Task Dense Scene Understanding},
author={Xin, Yi and Du, Junlong and Wang, Qiang and Lin, Zhiwen and Yan, Ke},
booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
volume={38},
number={14},
pages={16085--16093},
year={2024}
}
@inproceedings{xin2024mmap,
title={Mmap: Multi-modal alignment prompt for cross-domain multi-task learning},
author={Xin, Yi and Du, Junlong and Wang, Qiang and Yan, Ke and Ding, Shouhong},
booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
volume={38},
number={14},
pages={16076--16084},
year={2024}
}