New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Get worse result on test set while training with the proposed curriculum learning algorithm #10
Comments
Hi, Thanks for your interest and reporting this issue. |
Thanks for your reply! |
Hi, There is no special training skill. bash run_seq2seq_with_pretrain.bash -d 0 -k 3 -f tree -m t5-base --label_smoothing 0 -l 1e-4 --lr_scheduler linear --warmup_steps 2000 -b 16 -i dyiepp_ace2005_subtype
bash run_seq2seq_with_pretrain.bash -d 0 -k 3 -f tree -m t5-large --label_smoothing 0.2 -l 5e-5 --lr_scheduler linear --warmup_steps 2000 -b 8 -i dyiepp_ace2005_subtype
bash run_seq2seq_with_pretrain.bash -d 0 -k 3 -f tree -m t5-large --label_smoothing 0.2 -l 5e-5 --lr_scheduler linear --warmup_steps 2000 -b 8 -i one_ie_ace2005_subtype
bash run_seq2seq_with_pretrain.bash -d 0 -k 3 -f tree -m t5-large --label_smoothing 0.2 -l 5e-5 --lr_scheduler linear --warmup_steps 2000 -b 8 -i one_ie_ere_en_subtype |
Hi, I trained with exactly the same arguments except for '-k 3', what is this? |
Sorry about that. K is the number of running in my raw code. K=1 means seed=421 |
got it. Thank you! |
Hi, I evaluate t5-base, training with the proposed curriculum learning algorithm, on the dataset namely ACE05-EN+ and get worse test result:
test_trigger-F1 = 65.9436
test_trigger-P = 61.0442
test_trigger-R = 71.6981
test_role-F1 = 49.6
test_role-P = 45.8693
test_role-R = 53.9913
While evaluating the model trained without curriculum learning algorithm, I get:
test_trigger-F1 = 68.8863
test_trigger-P = 67.1141
test_trigger-R = 70.7547
test_role-F1 = 49.0647
test_role-P = 48.6448
test_role-R = 49.492
The performance drops when training with the curriculum learning algorithm. The test-trigger-P drops much if +CL.
Is there anything wrong?
Here is my training args:
epoch: 5+30, batch_size=32, metric_for_best_model=eval_role_F1, label_smoothing=0.2, model: t5-base, dataset: ACE05-EN+
Looking forward for your reply, thank you!
The text was updated successfully, but these errors were encountered: