New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About Fine-tuning for Text Summarization #27
Comments
@deep-cv-nlp I have some suggestions:
Some other tips:
Besides, our pre-trained model is still under training. I will update it when having a better result. (But anyway, your Rouge-1 F1 seems worse, I can achieve nearly 37.8 RG-1 F1 when my RG-2 F1 at 18.5). |
@StillKeepTry Thank you for your response.
Thanks. |
@deep-cv-nlp just use : |
Do you happen to know, what format is the data used for fine-tuning the model? |
@deep-cv-nlp @hivestrung We have released a pre-trained model (base setting) for summarization task |
A copy of cnndm fine-tuned model can be download from here. |
hi, Could you share the hyper-parameters for training on the Gigawords? i cannot reproduce the results reported in the paper, around 2 points off. Thanks! |
A copy of gigaword data can be obtained from here. So the command can be:
Different from cnndm, gigaword do not need to set
A fine-tuned model can be downloaded from this link. You can get 38.80/19.92/36.11 from the provided checkpoint. |
hi, thanks for sharing the command for reproducing the Gigaword dataset. Now I can reach what was reported in the paper, however, I cannot reproduce the score of 38.80/19.92/36.11 in your recently released trained model. The only difference of settings is that I only used one gpu for training, so I set the update-freq to 4 instead of 1. I am wondering whether this could lead to the performance discrepancy. Thanks! |
Hi,
Thank you for the great work. Recently I tried fine-tuning based on the pre-trained model (https://modelrelease.blob.core.windows.net/mass/mass_summarization_1024.pth).
I followed the instructions (https://github.com/microsoft/MASS#fine-tuning-2) in the readme file and ran this command on a single GPU machine. After the command finished, I tested the output by running
python translate_ensemble.py --exp_name giga_test --src_lang ar --tgt_lang ti --beam 5 --batch_size 1 --model_path ./dumped/mass_summarization/bvk6g6f9xl/checkpoint.pth --output_path ./dumped/mass_summarization/bvk6g6f9xl/output.txt.beam5 < ./data/processed/giga/test.ar-ti.ar
. Then I processed the output to remove the BPE mark@@
and tested the ROUGE scores. The ROUGE scores are ROUGE-1 F1=37.2 and ROUGE-2 F1=18.8.I think I have missed something important here. Could you please instruct me how to correctly fine-tune the model?
The text was updated successfully, but these errors were encountered: