New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
tips for training real 360 inward-facing scene #41
Comments
That actually would be my first suggestion -- we got the best results for 360 scenes when capturing them on the ground, not on a table. Having them on a table where the background changes more between views hurt performance in the examples we tried. Another problem we had was keeping exposure constant across all captures. This is important for good performance and is trickier with 360 because it's hard to get a 360 capture where lighting doesn't vary dramatically between directions (since usually one direction faces a window and one does not). One of the example views you posted has a huge blown out specular reflection on the table which is probably causing big problems. 360 captures also require a lot more inputs than forward facing since they effectively cover a huge amount of disparity. You probably want at least 80-100 inputs. |
Hi. I am just starting to experiment with 360 scans using this network as well (great work btw). Looking at your results it seems that the depth estimate on the statuette is quite consistent while it falls apart for background pixels. |
Uniform background(floor) might also be a problem, as you said. But actually what I found is that even a little arbitrary background (like I mentioned in the previous posts) makes training difficult. After deleting those images, the result looks better, even still with some incorrect depth prediction due to the uniform floor. So what I found good practices:
I will close this issue by posting the final result... still experimenting. |
@kwea123 Hello, I'm also trying to train NeRF with 360 captures scene (the AIST++ dataset
|
i request a feature which takes in the ordinary positions format like T_cam_w, which is common |
I tried to train on my own 360 inward-facing scenes, however, there is a huge portion of noise in the output: It doesn't disappear no matter how many steps I train.
I follow the suggestion in the readme and add
--no_ndc --spherify --lindisp
, the other settings are same asfern
.I suspect that it is due to the fact that input images has many arbitrary backgrounds, which deteriorates the training? For example some of the training images:
These arbitrary backgrounds are inevitable unless I have a infinite ground or an infinitely large table...
What's your opinion on this problem? Is it really due to the background or any other reason?
The text was updated successfully, but these errors were encountered: