-
Notifications
You must be signed in to change notification settings - Fork 64
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
关于模型泛化性的问题 #3
Comments
Thanks for your attention to our work. I believe the generalization problem lies in two parts. The first is the scene generalization. Previous methods, such as DIW, MiDaS, DPT, and our DiverseDepth/LeReS, mainly focus on this part. They merge large-scale diverse data in training. As they employ a ranking loss or scale-shift invariant loss, the camera variants issues are decoupled. Thus they can produce a strong and robust depth model, i.e. generalize to diverse scenes. If you only mix indoor datasets, the model cannot work on in-the-wild scenes. The second is the camera problem, which is related to the metric problem. If you do not need the absolute depth, you can ignore this and follow previous method to train a robust relative depth model. However, if your evaluation focuses on metric, you can follow our method to preprocess data. This can help the model to converge and achieve both strong generalization and metric recovery ability. Also note that only indoor data cannot ensure strong generalization. Hope this can help you. |
非常感谢您的回复,我还有几个疑问。
|
我的理解哈: |
|
在训练中,大部分数据也并不1000。比如taskonomy,大部分在500-700左右,测试的NYU等数据集也并不在500左右。所以并不是平均值在1000。关于这部分的ablation,还可以继续深入探索一下。 |
大佬您好,我看您这篇论文中用了大量的RGBD数据集,模型的泛化性比较好是不是主要原因是混合了大量数据集呢?还是说这种强泛化性也得益于将不同相机都规范化的做法?我之前做的是针对于室内场景的单模态RGB深度估计,我混合了几个公开RGBD数据集,还有我自己采集的一些实验场景的RGBD数据(普通混合,并没有用过这种相机规范化做法)。但是,泛化性依旧不是很好,只要物体换个场景,我这边预测的深度误差就非常大了。不知道使用您这种相机规范化方法,只是混合室内场景的数据,泛化性能不能有所提高?
The text was updated successfully, but these errors were encountered: