Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fine-tuning after post-trained #5

Closed
MrGGLS opened this issue Dec 4, 2022 · 2 comments
Closed

fine-tuning after post-trained #5

MrGGLS opened this issue Dec 4, 2022 · 2 comments

Comments

@MrGGLS
Copy link

MrGGLS commented Dec 4, 2022

Hi : )

I'm trying to introduce CL into CGEC (Chinese GEC) task. You used post-trained model (trained on non-native learner data) and then fine-tune (trained on native leaner data) with two strategies (NLL & CL) if I remember correctly...

I did the almost same steps use NLL strategy, but the fine-tuned model got a lower score than post-trained model in test-set (their $F_{0.5}$ scores were 5 and 25 respectively)

I think the reason might be the different data distribution, and I wanna know how you can make NLL better than DI (in paper)

I hope I made myself clear.

thanks

@MichaelCaohn
Copy link
Collaborator

Hi,

Thank you for the question. I think you might need to re-tune the hyper-parameters to make NLL better than DI. Also, as far as I know, the Chinese GEC needs to go through some segmentation process, which may also affect the performance of NLL.

Ideally, NLL should be better than DI.

Hope this addresses your concern

@MrGGLS
Copy link
Author

MrGGLS commented Jul 28, 2023 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants