Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any data on expected training time? Also, Are there plans for realtime visualized output? #10

Closed
ExponentialML opened this issue Feb 16, 2020 · 5 comments

Comments

@ExponentialML
Copy link

Hello. First, I would like to thank you for such great, fully fleshed out project.

In regards to training, is there any data on how long it should generally take per GPU? If we're talking about a GPU with 8GB of memory, how long would it take to finish training on a small to medium sized data set?

This one kind of ties into the training time, but are there any plans for realtime visualization of training data? If the training does take long periods of time, it would be beneficial to see the results updated at a faster rate.

Any input is appreciated. Thanks!

@AliaksandrSiarohin
Copy link
Owner

Hi, thank you for your interest,
I usually train on 2 Titan X with 12GB of memory. It take 2 days for dataset like TaiChi and VoxCeleb. For simple dataset like Nemo, training can be done in 24 hours. With 1 gpu number should be multiplied by 2.

The problem with visualization is that it take a lot of time to compute. If you need more frequent visualizations you can change the parameter num_repeats


. Basically the number of passes over all the videos is num_epochs * num_repeats. You can decrease number of repeats and increase number of epochs. Because visualization is drawn after each epoch, you will get more frequent visualizations at a cost of some efficiency.

Regards, AS

@ExponentialML
Copy link
Author

Hi, thank you for your interest,
I usually train on 2 Titan X with 12GB of memory. It take 2 days for dataset like TaiChi and VoxCeleb. For simple dataset like Nemo, training can be done in 24 hours. With 1 gpu number should be multiplied by 2.

The problem with visualization is that it take a lot of time to compute. If you need more frequent visualizations you can change the parameter num_repeats

. Basically the number of passes over all the videos is num_epochs * num_repeats. You can decrease number of repeats and increase number of epochs. Because visualization is drawn after each epoch, you will get more frequent visualizations at a cost of some efficiency.

Regards, AS

Excellent. You rock!

@kartikJ-9
Copy link

Can i train this on colab. How many videos I should keep for train and test? I want the output quality to be 256/512 resolution. I can see that the fashion dataset had 500 videos for train and 100 for test. My dataset is also similar to fashion with a person in the frame. So will I be able to train the model in colab if I use 500,100 for train,test? Please help me with this. Thanks

@AliaksandrSiarohin
Copy link
Owner

Most likely not possible, colab gpus is rather slow. And it will shut down and kick you out after 24 hours.

@MentalGear
Copy link

@kartikJ-9 Have you made progress with training a higher res model on colab or elsewhere? Would appreciate if you could share what you tried.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants