Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

运用prefix_projection 方法训练test acc不变一直是62.1 #13

Closed
yh351016 opened this issue Dec 7, 2021 · 5 comments
Closed

运用prefix_projection 方法训练test acc不变一直是62.1 #13

yh351016 opened this issue Dec 7, 2021 · 5 comments

Comments

@yh351016
Copy link

yh351016 commented Dec 7, 2021

12/07/2021 19:30:19 - INFO - training.trainer_base - ***** Epoch 12: Best results ***** 12/07/2021 19:30:19 - INFO - training.trainer_base - best_epoch = 0
12/07/2021 19:30:19 - INFO - training.trainer_base - best_eval_accuracy = 0.6217125382262997
12/07/2021 19:30:19 - INFO - training.trainer_base - epoch = 12.0
OrderedDict([('best_epoch', 0), ('best_eval_accuracy', 0.6217125382262997), ('epoch', 13.0)])
{'loss': 0.7488, 'learning_rate': 0.006054054054054054, 'epoch': 13.51}

@Xiao9905
Copy link
Member

Xiao9905 commented Dec 7, 2021

Hi, prefix projection needs a careful hyper-parameter tuning, and in practice we find it does not work for every tasks.

@yh351016
Copy link
Author

yh351016 commented Dec 9, 2021

BoolQ and RTE result on the table of github is not same with paper report

@Xiao9905
Copy link
Member

Xiao9905 commented Dec 9, 2021

As we stated in the README.md, the hyperparameters provided in the repo is reimplemented with 3090 GPUs, rather than A100s (on which our paper results come from).
Since experiments reported in our paper are all conducted on NVIDIA DGX-A100 servers (which might be difficult to acquire), we reimplement P-tuning v2's results on BERT-large/RoBERTa-large with:

@yh351016
Copy link
Author

yh351016 commented Dec 9, 2021

As we stated in the README.md, the hyperparameters provided in the repo is reimplemented with 3090 GPUs, rather than A100s (on which our paper results come from). Since experiments reported in our paper are all conducted on NVIDIA DGX-A100 servers (which might be difficult to acquire), we reimplement P-tuning v2's results on BERT-large/RoBERTa-large with:

thanks,I think it will only affect the speed. Does it have anything to do with the result?

@Xiao9905
Copy link
Member

Xiao9905 commented Dec 9, 2021

As is in the README.md
We notice that the best hyper-parameters can be sensitive to your server environment and package version.

If you do not have the exact same environment, we highly recommend you to run hyper-parameter search in your environment
based on our example hyper-parameter search script in search_script and result collection scripts search.py.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants