-
Notifications
You must be signed in to change notification settings - Fork 44
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
specAugment policy and schedules #48
Comments
We already have variations of that, where we also play around with scheduling of SpecAugment. Note that the Yes sure, you can play around with learning rate warm-up as well. My experience however is that increasing usually will not help. Reducing the LR decay helps when you want to increase your overall training time, i.e. train more epochs. And training longer usually helps. When you look at this original SpecAugment paper, you will see that they effectively train much longer than we do. |
Thanks for your answer, Albert! I am sorry for a possibly naive question, but in the config example you mention above, the My understanding is: |
Yes sure. |
Hi,
I wanted to run an experiment with LD augmentation policy (as described in the Google Brain paper ) along with D learning schedule.
I was wondering what would be the right way of doing something like with base2.conv2l.specaug.curric3.config.
I was thinking of doing:
random_mask
two more times.10
to20
or40
newbob_learning_rate_decay
from0.9
to0.95
Would it be a reasonable thing to do?
Thanks
The text was updated successfully, but these errors were encountered: