Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

coco-LM post-pretrain? #1

Closed
paulpaul91 opened this issue Dec 9, 2021 · 2 comments
Closed

coco-LM post-pretrain? #1

paulpaul91 opened this issue Dec 9, 2021 · 2 comments

Comments

@paulpaul91
Copy link

We want to use coco-LM to do post-ptrtain(like MLM), but the main transformer �didn‘t seen MLM task,we don't know if it‘s feasible

@yumeng5
Copy link
Collaborator

yumeng5 commented Dec 9, 2021

Hi @paulpaul91,

I believe the best way of doing continued pretraining of COCO-LM is to use the original COCO-LM pretraining tasks (CLM + SCL) to avoid the discrepancy of objectives in pretraining/continued pretraining.

However, if you have to continue pretraining COCO-LM with different objectives like MLM, it probably is still feasible: As shown in Section 5.3 of our paper, although COCO-LM is not trained with MLM, it performs well in MLM-based prompt-based fine-tuning.

Best,
Yu

@paulpaul91
Copy link
Author

thanks

@yumeng5 yumeng5 closed this as completed Dec 16, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants