Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training on NTU 60 XSUB #10

Closed
erinchen824 opened this issue May 10, 2022 · 3 comments
Closed

Training on NTU 60 XSUB #10

erinchen824 opened this issue May 10, 2022 · 3 comments

Comments

@erinchen824
Copy link

erinchen824 commented May 10, 2022

Thanks for your awesome work and generous code sharing.
Recently, I have tried to reproduce the results on NTU 60 XSUB benchmark with one 3090 GPU.
However, it seems the training is really slow and it would take more than a week to train 20 epochs following the original setting.
I have tried to enlarge the batch size, and it still quite slow.
So, what is the reasonable training time for NTU 60 dataset?
Is there anything I have missed?

Looking forward to your reply.

Thanks

@erinchen824
Copy link
Author

erinchen824 commented May 10, 2022

Thanks for your awesome work and generous code sharing. Recently, I have tried to reproduce the results on NTU 60 XSUB benchmark with one 3090 GPU. However, it seems the training is really slow and it would take more than a week to train 20 epochs following the original setting. I have tried to enlarge the batch size, and it still quite slow. So, what is the reasonable training time for NTU 60 dataset? Is there anything I have missed?

Looking forward to your reply.

Thanks

I modify the dataloader to only sample one sequence for each video, and here are part of the training logs.

截屏_-_5月10日_下午9_02

@hehefan
Copy link
Owner

hehefan commented May 13, 2022

Hi,

I think the training time is reasonable. One reason is that we used all frames with all combinations, which is a common problem with video processing. Selecting a fixed frame combination for each video can significantly reduce time but may slightly affect the accuracy. Another reason is that we did not implement PSTConv in a parallel way like the standard convolutions in torch.nn. Point tubes are sequentially constructed over time.

Best

@erinchen824
Copy link
Author

Hi,

I think the training time is reasonable. One reason is that we used all frames with all combinations, which is a common problem with video processing. Selecting a fixed frame combination for each video can significantly reduce time but may slightly affect the accuracy. Another reason is that we did not implement PSTConv in a parallel way like the standard convolutions in torch.nn. Point tubes are sequentially constructed over time.

Best

Thanks for your quick reply, that resolves my question.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants