New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How about the result on real world datasets? #5
Comments
I think you'll have to be a bit more specific about the problem. Do you have a problem with COLMAP or does my method give you bad results when you use it with the COLMAP poses? If it's the latter, make sure the cameras are in the correct coordinate system and that you are correctly loading the depth maps. You will have to either save the data in the described format or write your own dataset loader. Since you have depth, you can also try aligning the images with BundleFusion. |
The poses are estimated by colmap. The depth image are normalized to 0-1 for training. I wondered whether I need to reconstruct my scene my rgbd reconstruction and get the sc_factor and translation correct for the network? |
The depth images need to be in metric space. |
@endlesswho is the problem solved? could you please post your result here? |
I'm so sad the problem still remains. The depth images in metric space, but the output pose of colmap is a scaled value. I think a rgbd reconstruction method would work! |
You can use some flavor of KinectFusion to to obtain camera poses. If you want to use the COLMAP poses with your depth sensor's measurements, you will need to scale the translation vectors of your camera poses. |
@rancheng My problem was solved with a rgbd reconstruction with icp matching and get the trajectory. However, the reconstruction results with @dazinovic 's method seems no so good. I also run the result with breakfast_room. With a disturb of trajectory, the result was shown bellow: |
Reasonable! I'll have a try and pose my new results. |
My camea extrinsics are in wrong coordinate system. I transform my coordinate system to OpenGL convention, the results are better. However, what if I don't know the bounding of the scene, any suggestions to solve this question? |
You can approximate it with your camera positions. |
My result is all right with the help of your advices. Thanks for your kindly reply. |
Hello @endlesswho @dazinovic, can you briefly describe what changes you made to get it to work with real world datasets? Is it something as follows?:
Also a few more questions @endlesswho:
|
I generated poses using BundleFusion (although, you can also use Colmap for this) and then applied the transformed as described in the linked issue. The depth maps are not normalized. The values need to be in meters. Scannet depth maps are in millimeters, so you simply need to divide by 1000. The method will work with other scales too, but the depth maps need to be consistent with the camera poses. |
Hello, I try to reproducing Neural-RGBD with the data used by manhattan SDF and NICE-SLAM (e.g. Replica). Do you have any ideas? Thanks! @dazinovic @endlesswho |
I have encounter the same issue with @junshengzhou , could you reply to us? @dazinovic @endlesswho |
@dazinovic How about the result on real-world datasets. I collect the datasets with my own RGBD camera and estimate the pose with colmap. But the results is a mass. Any advices?
The text was updated successfully, but these errors were encountered: