Our cleaned datasets can be downloaded at:
The training script is train.py. For example, to train the GPT-2 model on the cleaned DailyDialog dataset:
python train.py \
--train-path=path-to-the-training-csv-file \
--eval-path=path-to-the-validation-csv-file \
--num-epochs=50 \
--model-str=gpt2 \
LSTM and Transformer are trained using the Fairseq framework.
The training script will automatically generate a timestamped logging directory to store the checkpoints as well as log files. The validation performance can be monitored during training through tensorboard:
tensorboard --logdir=path-to-the-timestamped-logging-folder
If the performance is still increasing at the end of training, you can resume with the following command:
python train.py \
--train-path=path-to-the-training-csv-file \
--eval-path=path-to-the-validation-csv-file \
--num-epochs=100 \
--model-str=gpt2 \
--resume-path=path-to-the-timestamped-logging-folder
After the performance has peaked, you can evaluate the model using eval.py:
python eval.py --ckpt=path-to-the-best-validation-checkpoint --eval-path=path-to-the-test-csv-file