New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Regarding information required to be able to train on custom dataset #85
Comments
Yes, you need rgb images, depth maps, extrinsic and intrinsic parameters to train LoFTR. |
Thanks for your response. Once the d2net processed npz file is created is there any specific way to be able to get the npz files that were used for training? |
Also wanted to ask regarding what the 'central_matches' value is in pair_infos |
just wondering, is it possible to train the network with image pairs plus ground truth (matching keypoints), like the output from the network. |
Possibly not. Only if you have ground truth point on the 1/8 grids. And please open another issue to discuss this if you have more questions. |
We didn't use this. |
I'm sorry that I don't have the code now. I think you should set-up train/val/test splits and reorder the data dictionary. Have you solved the problem yet? |
Yes, thanks! |
What exactly would I need to provide to be able to train LoFTR on a custom dataset? It was pointed out in some of the previous issues that all one would need would be depth map, extrinsics and intrinsics for every image and the ground truth is generated on the fly.
The npz files seem to have fields - 'image_paths', 'depth_paths', 'intrinsics', 'poses', 'pair_infos'. Assuming 'poses' is just the extrinsics (R | t), how exactly is 'pair_infos' generated and how are the covisibility scores calculated?
The text was updated successfully, but these errors were encountered: