Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about fine-tuning on Middlebury2014 #82

Open
kongdebug opened this issue Mar 23, 2023 · 5 comments
Open

Question about fine-tuning on Middlebury2014 #82

kongdebug opened this issue Mar 23, 2023 · 5 comments

Comments

@kongdebug
Copy link

Hi @lahavlipson,
Thank you for your great work!

I am now using the weights trained on the scenefrlow dataset to fine-tune on the middlebury dataset. After I finish the fine-tuning, the results on the D1 of the full middlebury dataset are even worse than before. Is this normal?

The raftstereo-sceneflow.pth result is consistent with Table 1 of the paper:
image

However, the result after fine tuning on the Middlebury2014 dataset are relatively poor:
image

@lahavlipson
Copy link
Collaborator

I've found that the performance is better and more stable if the learning rate is small, e.g. --lr 0.00002, similar to what we use for KITTI; I've updated the command in the README.

@kongdebug
Copy link
Author

I've found that the performance is better and more stable if the learning rate is small, e.g. --lr 0.00002, similar to what we use for KITTI; I've updated the command in the README.

Thank you for your reply and look forward to the updated README.

@kongdebug
Copy link
Author

I've found that the performance is better and more stable if the learning rate is small, e.g. --lr 0.00002, similar to what we use for KITTI; I've updated the command in the README.

In addition, how much learning rate do you use to fine-tune the KITTI 2015 dataset?
In section 4.2 of the paper, it is mentioned that the minimum learning rate used for fine-tuning the KITTI 2015 dataset is 1e-5. What is the maximum learning rate? I hope you can tell me, thank you!

@lahavlipson
Copy link
Collaborator

On KITTI, we use --lr 0.00001

@kongdebug
Copy link
Author

On KITTI, we use --lr 0.00001

Thank you. I used --lr 0.0002 and submitted the results to KITTI website for testing. The metrics of D1-all are consistent with those of RAFT-Stereo on the list. However, fine-tuning on the middlebury dataset with --lr 0.00002 did not get the same precision as the Middlebury.pth weights you supplied.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants