New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Trained your network locally, but my eval. result is not as good as yours #12
Comments
How long did you train it for? |
Hi, I ran 20 epochs, and tried training locally twice, but got almost same results. -- Training messages When I ran inference with your weight or your ".npy", I got almost the same results as your paper. CJ |
This is consistent with my observation too. The system could start overfitting on KITTI if trained for too long. You could use the validation set to decide the stopping point. |
I have the same problem. According to the paper (section4), I chose the check point in 150930 steps to test the result (eigen test split). But the performance is really bad, almost double the errors. Do you just resize the image to 128x416 first, then feed it to the network? Or you perform some pre-processing? |
The best model I had was trained on Cityscapes for ~100K steps and then fine-tuned on KITTI for ~50K steps. This should have been made more clear in the paper. |
I faced similar problem too. And there are a more serious problem about the explainability mask. while the model trained with mask (weight is 0.2 from the paper) have a result, The both two model are selected around the 200k steps, model_191178. |
I have gotten reports that if you use a Tensorflow version later than 1.0, the results could get somewhat worse (not sure why). Also, I have done some tweaks to the code that improve the results for the non-mask model (see Notes section in README), which might have made the explainability weight non-optimal any more. |
Hi,
Thanks for sharing your wonderful works.
I followed your readme file, and it seems that everything goes well until when I see different evaluation result compared with yours (Ours(K) in table 1 of your paper) when I do evaluate based on locally trained weight. During training, I used the same parameters as suggested on this webpage.
The followings metrics which I got:
abs_rel, sq_rel, rms, log_rms, d1_all, a1, a2, a3
0.2621, 3.6171, 8.2036, 0.3806, 0.0000, 0.6577, 0.8520, 0.9258
I checked and reviewed procedure, but I cannot find any hint for the above different evaluation result.
Let me know what do you think about it.
Regards,
CJ
The text was updated successfully, but these errors were encountered: