SPA: Towards A Computational Friendly Cloud-Base and On-Devices Collaboration Seq2seq Personalized Generation
This is the official repository of the paper: SPA: Towards A Computational Friendly Cloud-Base and On-Devices Collaboration Seq2seq Personalized Generation.
- Thoroughly prepare your dataset, and adjust it to the prescribed format:
{
"instruction": ...,
"input": ...,
"output": ...,
}
- Clone this repo.
git clone git@github.com:OceannTwT/SPA.git
-
Replace the code on
LlamaForCausalLM
withmodel/modeling_SPA.py
. -
Add your dataset in
data/dataset_info.json
. -
Tune and get the SPA model!
bash train.sh
- Execute the model inference by running the
llama_SPA_predict.py
script, ensuring to modify the directory in your additional parameters if necessary.
- This repository is dedicated to on-device personalized Language Models (LLMs), with the potential to significantly enhance the speed and reliability of on-device LLMs.
- We express our gratitude for the valuable contributions made by all co-authors and the dedicated efforts of the Siri-China teams involved in this project.
[24/08/20] 🔥 We are happy to announce that SPA has been accepted to PRICAI 2024 main conference(oral)!
If you use the SPA for your work, please cite:
@misc{liu2024spa,
title={SPA: Towards A Computational Friendly Cloud-Base and On-Devices Collaboration Seq2seq Personalized Generation},
author={Yanming Liu and Xinyue Peng and Jiannan Cao and Le Dai and Xingzu Liu and Weihao Liu and Mingbang Wang},
year={2024},
eprint={2403.07088},
archivePrefix={arXiv},
primaryClass={cs.CL}
}