Skip to content

ttw1018/MoPE-DST

Repository files navigation

MoPE: Mixture of Prefix Experts for Zero-Shot Dialogue State Tracking

Abstract

Zero-shot dialogue state tracking (DST) transfers knowledge to unseen domains, reducing the cost of annotating new datasets. Previous zero-shot DST models mainly suffer from domain transferring and partial prediction problems. To address these challenges, we propose Mixture of Prefix Experts (MoPE) to establish connections between similar slots in different domains, which strengthens the model transfer performance in unseen domains. Empirical results demonstrate that MoPE-DST achieves the joint goal accuracy of 57.13% on MultiWOZ2.1 and 55.40% on SGD.

Model Architecture

Illustration of our proposed method, including (a) Slot clustering, (b) Deep Prefix Prompt Tuning, and (c) Multiple Prefix Prompt Generation. Slot clustering is used to categorize all slots into distinct clusters and establishes connections between slots in different domains. Deep Prefix Prompt Tuning is our method to strengthen the LLM's conditional generation. Multiple Prefix Prompt Generation shows the complete pipeline of solving DST task.

Depenedency

Download packages

python >= 3.10

pip install -r requirements.txt

Download pretrained model

chatglm-6b download link

Download & process source data

bash download_process_data.sh

Train & Evaluate

bash run.sh

note:

--chatglm_path should be changed to your personal pretrained model save path.

--result_csv_dir is the directory of the saved results. We provide two display forms for the results, you can refer to the csv file to check the main results. Further more, we provide a sqlite3 db to record all results during the whole train and evaluation, such as the performance on the validation data and so on.

--checkpoint all tuned parameters are stored in save/${dataset}/${exclude_domain}/${train_id}/${feature}/best, so you can get the whole information of the checkpoint in the directory mentioned above.

Other notes

we find that there exists some variances during the train. Therefore, there may be some fluctuations in the experimental results.

Citation

If you think that our work is helpful to you, don't forget to cite us.

@article{tang2024mope,
  title={MoPE: Mixture of Prefix Experts for Zero-Shot Dialogue State Tracking},
  author={Tang, Tianwen and Zhu, Tong and Liu, Haodong and Bai, Yin and Cheng, Jia and Chen, Wenliang},
  journal={arXiv preprint arXiv:2404.08559},
  year={2024}
}

About

The code for "MoPE: Mixture of Prefix Experts for Zero-Shot Dialogue State Tracking"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published