Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[question] Importing Real Laser Depth Data #764

Closed
flolu opened this issue Jan 20, 2020 · 8 comments
Closed

[question] Importing Real Laser Depth Data #764

flolu opened this issue Jan 20, 2020 · 8 comments

Comments

@flolu
Copy link

flolu commented Jan 20, 2020

Describe the problem
I was wondering if it is possible to feed Meshroom with laser data? I've seen this video where they were using Laser in combination with photo scanning, and they've got really good results.
But I also don't know which hardware is required for capturing real depth information with the help of a laser or something similar.

Is there already such a feature in Meshroom?
Which Hardware would I need?

Video
https://youtu.be/M6Uf8kjfbSA?t=177

Additional context
Maybe someone could point me in the right direction.

@natowi
Copy link
Member

natowi commented Jan 20, 2020

No, at the moment there is no such feature in Meshroom. (But there is a draft for this based on PCL)

Cloud Compare (or Meshlab) could be used to align the point cloud ply created in the meshing node with laser scan. In Meshlab you could then create the mesh from the point cloud and do the texturing back in Meshroom. https://github.com/alicevision/meshroom/wiki/Texturing-after-external-re-topology

Laser scanners are expensive. There are 2D laser scanners starting at 80$ like a Slamtec RPLIDAR A1, but 3d scanners start at 2500$.

Depth cameras like the Kinect would be cheaper, but Meshroom does not yet support them #255

Another way to improve accuracy would be using projected light patterns, but this is not suitable for outdoor usage.

@flolu flolu changed the title [question] [question] Importing Real Laser Depth Data Jan 20, 2020
@fabiencastan
Copy link
Member

The draft for 3D-3D alignment is here if someone is interested to contribute on this topic. But it's far from being ready for end-users.

@natowi
Copy link
Member

natowi commented Feb 24, 2020

This issue is solved and closed due to inactivity. Feel free to re-open if new information is available.

@remmel
Copy link
Contributor

remmel commented Feb 18, 2021

If for each laser depth data you have an image, you could try to improve each exr files generated by Meshroom with you depth data..
As I have the same needs (Use the Kinect v1 depth data or use the tof images from my Honor View 20) I'll try to read and improve exr files and I'll also try to do it on OpenMVS
Main avantage of mixing tof images with photogrammetry depth maps is that tof could be more precise and consume less time, but photogrammetry depth maps "see" dark surfaces

@mellinger
Copy link

@remmel Were you able to try this with OpenMVS? I couldn't figure out how to manually import those depth maps.

@remmel
Copy link
Contributor

remmel commented Mar 21, 2021

@mellinger I have not done anything with OpenMVS yet

@bednarhonza
Copy link

Hello everyone, I have noticed that the PR #425 has been merged. Is it already in some working phase, or is still a development needed to use it?

Thanks for the great work!

@natowi
Copy link
Member

natowi commented Jun 7, 2023

@bednarhonza alicevision/AliceVision#1441 it is on the milestone list for an upcoming release https://github.com/alicevision/AliceVision/milestone/35

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

6 participants