You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, thanks for your great work.
According to the paper description, the predecessor module weights are frozen after fine-tuned on the task data( including embedding & output classifier).
The code, however, if my understanding is correct, the fine-tuned predecessor weights are not frozen, instead, the loss can BP to the corresponding parameters.
So, which pattern is supposed to be correct? Thanks in advance.
The text was updated successfully, but these errors were encountered:
Hi, thanks for your great work.
According to the paper description, the predecessor module weights are frozen after fine-tuned on the task data( including embedding & output classifier).
The code, however, if my understanding is correct, the fine-tuned predecessor weights are not frozen, instead, the loss can BP to the corresponding parameters.
So, which pattern is supposed to be correct? Thanks in advance.
The text was updated successfully, but these errors were encountered: