Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Lower the default learning rate for albert #1786

Merged
merged 1 commit into from
Aug 20, 2024

Conversation

mattdangerw
Copy link
Member

@mattdangerw mattdangerw commented Aug 19, 2024

Just noticed while porting #1767 that the default learning rate for our classifier does not work for albert pretrained checkpoints. Let's lower it for this model

Fixes #831

Just noticed while porting keras-team#1767
that the default learning rate for our classifier does not work
for albert pretrained checkpoints. Let's lower it for this model
@mattdangerw mattdangerw force-pushed the albert-learning-rate branch from 558ac3f to 964bce1 Compare August 19, 2024 23:58
@mattdangerw mattdangerw merged commit 4931eae into keras-team:master Aug 20, 2024
8 of 11 checks passed
pkgoogle pushed a commit to pkgoogle/keras-hub that referenced this pull request Aug 22, 2024
Just noticed while porting keras-team#1767
that the default learning rate for our classifier does not work
for albert pretrained checkpoints. Let's lower it for this model
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Albert fine tuning does not always converge
2 participants