Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ADE20k Learning Rate #52

Closed
cjrd opened this issue Mar 9, 2022 · 4 comments
Closed

ADE20k Learning Rate #52

cjrd opened this issue Mar 9, 2022 · 4 comments

Comments

@cjrd
Copy link

cjrd commented Mar 9, 2022

Section A.4 of the paper has "We search for the optimal lr for each entry in Table 5 " when referring to segmentation on ADE20k. Could you share these learning rates? My reproduced baselines are about 2 points too low.

Thanks!

@implus
Copy link

implus commented Mar 10, 2022

see our reproduction here: #2 (comment)

@endernewton
Copy link
Contributor

Closing with the reproduced results above. Please check lr, layer-wise rate decay, and drop path rate similar to what we have done in detection.

@youngwanLEE
Copy link

@endernewton Hi,

Are LR, layer-wise rate decay, and drop-path hyper-parameters of UperNet the same as those of detection?

Do you have a plan to release the segmentation code and weights?

@hellojialee
Copy link

Dose layer-wise rate decay really make differences obviously?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants