Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training code #3

Closed
miguelvr opened this issue May 7, 2019 · 2 comments
Closed

Training code #3

miguelvr opened this issue May 7, 2019 · 2 comments

Comments

@miguelvr
Copy link

miguelvr commented May 7, 2019

Are you planning on releasing the training code?

Also, did you try to implement the ResNet BasicBlock with OctConv?

I'm trying to do it in my own implementation, but it is tricky due to the lack of down sampling on the first layer.

@d-li14
Copy link
Owner

d-li14 commented May 7, 2019

An experimental training script is provided here. As it is a trivial implementation resembling the official tutorial, I didn't include it in this repo.

As for hyperparameters, I strictly follow the ResNet paper other than decaying the LR as a cosine function shape during the total 120 epochs.

Transferring to ResNet built with basic block follows the same principle by setting alpha_in to zero in the first stage. Temporarily, it is not the focus of the original paper as well as my reproduction plan.

@miguelvr miguelvr closed this as completed May 7, 2019
@gasvn
Copy link

gasvn commented Jun 2, 2019

Thanks for your excellent work. I am wondering if you could share the "cosine function shape"- adjust_learning_rate function of your training code. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants