Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any pre-processed data for training and test #2

Closed
yangengt123 opened this issue May 23, 2019 · 16 comments
Closed

Any pre-processed data for training and test #2

yangengt123 opened this issue May 23, 2019 · 16 comments

Comments

@yangengt123
Copy link

Hi,
Thanks for sharing the code of your great work.

I am trying to run the training code, but I found the whole ScanNet dataset is rather than large. May I ask if it is necessary to download all the .sens from the dataset? Is it possible to get access to your per-processed data, or to train the network with part of this dataset?

BTW, a small contribution, if any one uses Cuda 9 + RTX 2080 GPU, the arch should be sm_70 for the nvcc compilement. A link for other gpu types: https://arnon.dk/matching-sm-architectures-arch-and-gencode-for-various-nvidia-cards/

@art-programmer
Copy link
Contributor

We are processing that. I will keep you updated. And thank you for the pointer!

@yangengt123
Copy link
Author

yangengt123 commented May 29, 2019

Hi, I was trying to run the training code, and I found in

with open(self.dataFolder + '/scannetv2-labels.combined.tsv') as info_file:

it asked a .tsv file named scannetv2-labels.combined.tsv is needed, and in
if not os.path.exists(scenePath + '/' + scene_id + '.txt') or not os.path.exists(scenePath + '/annotation/planes.npy'):

a .txt file named with the scene_id seems required to be under each scene's directory.

May I ask where I can get access to these files?

@art-programmer
Copy link
Contributor

They are from the original ScanNet dataset. You need to download the data and uncompress the .sens file (maybe delete the .sens file after uncompression to save space).

@yangengt123
Copy link
Author

yangengt123 commented Jun 7, 2019

Thanks.
Just for someone else who is not familiar with ScanNet dataset, to get thees files, please use the argument --type .sens .txt and --label_map during download.

By the way, how can I run the training with multiple gpu? I cannot find any line contain the function torch.nn.DataParallel(model).cuda(), which I usually use for multi-gpu training. I tried to uncommment

# Uncomment to train on 8 GPUs (default is 1)

but it seems does not work as well.

@dreamPoet
Copy link

Sorry but is there any way we can run ScanNet original sens reader.py for all sens files scattered in different scene files? It looks like I cannot move all .sens file into one directory as it is important to map to corresponding annotations in later merging operation.

@yangengt123
Copy link
Author

yangengt123 commented Jun 10, 2019

Hi @dreamPoet , I just simply set the --filename as the path/to/Scannet_data/scans/, and it seems work well for me.

@dreamPoet
Copy link

Hi @yangengt123, do you simply use --filename path/to/scans/ ? I get the error Is a directory: '../../ScanNet/scans/'. If I use ../../ScanNet/scans/*/*.sens, it gives error unrecognized arguments: scans/scene0000_01/scene0000_01.sens...

@yangengt123
Copy link
Author

yangengt123 commented Jun 12, 2019

Yeah, the only difference is I used absolute path, something like /hdd/Scannet_data/scans/, and I did not meet any trouble to parse it. But I think it is always possible to look into the SensorData.py to check which path the load function is using.

@dreamPoet
Copy link

Thanks.
Just for someone else who is not familiar with ScanNet dataset, to get thees files, please use the argument --type .sens .txt and --label_map during download.

By the way, how can I run the training with multiple gpu? I cannot find any line contain the function torch.nn.DataParallel(model).cuda(), which I usually use for multi-gpu training. I tried to uncommment

# Uncomment to train on 8 GPUs (default is 1)

but it seems does not work as well.

Thank you, but have you solve the problem of multi-GPU training now? I meet the same problem.

@dreamPoet
Copy link

By the way I use srun with gpu:8 and -n8 as bash command.

@dreamPoet
Copy link

And there are three parameters related to batch size, one in config, one in option, and one for dataset loader...Frankly, I am a bit of confused...

@wullish
Copy link

wullish commented Oct 1, 2019

@dreamPoet did you solve the .sens files extraction problem using reader.py? I'm in the exactly same problem. Any idea?

Best.

@wullish
Copy link

wullish commented Oct 2, 2019

Well, I could solve it using python 2.7 but calling one .sens file per once. And dont forget about the reader.py options arguments such as --export_depth_images, --export_color_images, etc.

@yangengt123
Copy link
Author

That is right, I use python2.7. Maybe this is the reason I did not meet any problem.

By the way, may I ask if you can still download the data from ScanNet recently? I tried to download the data to a new server but it seems my download script cannot work.

@wullish
Copy link

wullish commented Oct 8, 2019

Ah thanks.
Yes, I've downloaded the whole data last week. Maybe they've changed the urls. I think you can contact scannet@googlegroups.com and ask them about the script.

@yangengt123
Copy link
Author

yeah, maybe I should. Thanks.
I will close this issue now.
Please let me know if any pre-processed data become available.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants