Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

about prepocess #56

Open
UestcJay opened this issue Feb 2, 2023 · 10 comments
Open

about prepocess #56

UestcJay opened this issue Feb 2, 2023 · 10 comments

Comments

@UestcJay
Copy link

UestcJay commented Feb 2, 2023

Hi,

Thanks for the great work !
because 384 is the size for Omnidata-model, the dtu image size is 1200x1600, if I want to use monocular cues with original size, can I first resize the 1200x1600 -> 1152x1536, then get the monocular cues and upsamle them to 1200x1600?
looking forward to your reply!

@niujinshuchong
Copy link
Member

Hi, we also simply resize the monocular outputs to 1200x1200 with padding for dtu images with 1200x1600. You could check it here: https://github.com/autonomousvision/monosdf/blob/main/preprocess/paded_dtu.py.
Other way to get high-resolution monocular priors can be found here https://github.com/autonomousvision/monosdf#high-resolution-cues.

@UestcJay
Copy link
Author

UestcJay commented Feb 3, 2023

I still have some problems, is my method more convenient than the way in thepaded_dtu.py, because there is no need to modify the parameters of the camera?

@niujinshuchong
Copy link
Member

You could just try it out.

@UestcJay
Copy link
Author

How many experiments are averaged for the CD value on the DTU dataset reported in the paper?

@niujinshuchong
Copy link
Member

It's average over 15 scenes.

@Wuuu3511
Copy link

Hi,

Thanks for the great work ! because 384 is the size for Omnidata-model, the dtu image size is 1200x1600, if I want to use monocular cues with original size, can I first resize the 1200x1600 -> 1152x1536, then get the monocular cues and upsamle them to 1200x1600? looking forward to your reply!

Helllo! I'd like to ask a question.
Omnidata-model is trained with img_size 384. Can it support input at any image resolution such as 1152*1536?
Thank you!

@UestcJay
Copy link
Author

yes, as long as the length and width are multiples of 384.

@Wuuu3511
Copy link

yes, as long as the length and width are multiples of 384.

Thank you very much for your reply!
I try to use images 512 * 640 as input , Omnidata-model can also return a depth map which is 512*640. Picture of this size is not a multiple of 384. Does this result in a larger depth error?

@liuxiaozhu01
Copy link

Hello! I've got question here.
I am wondering whether the resolution of rgb images, depth and normal cues will impact on the reconstruction result. If it will, and why?
Thank you for your reply! My experiment result really confused me.

@niujinshuchong
Copy link
Member

Hi, omnidata is not trained on large resolution images. So it's not clear whether it can generalise in this case and the reconstruction results might vary scene by scene.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants