Skip to content

Conversation

mattdangerw
Copy link
Member

@mattdangerw mattdangerw commented Aug 19, 2024

Just noticed while porting #1767 that the default learning rate for our classifier does not work for albert pretrained checkpoints. Let's lower it for this model

Fixes #831

Just noticed while porting keras-team#1767
that the default learning rate for our classifier does not work
for albert pretrained checkpoints. Let's lower it for this model
@mattdangerw mattdangerw force-pushed the albert-learning-rate branch from 558ac3f to 964bce1 Compare August 19, 2024 23:58
@mattdangerw mattdangerw merged commit 4931eae into keras-team:master Aug 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Albert fine tuning does not always converge
2 participants