Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The accuracy is so high, so will this paper be the end of NTU-60? Or will there be any potential for promotion? #11

Closed
desertfex opened this issue Jun 28, 2020 · 11 comments

Comments

@desertfex
Copy link

desertfex commented Jun 28, 2020

What's the possible directions for the following research? Thank you.

@ZZYuting
Copy link

I think the directions are how to light the model. It seems cost much time.

@desertfex
Copy link
Author

@ZZYuting The number of layers has dropped to 3 comparing with the ST-GCN with 10 layers. I have no idea how to light the model without loss of accuracy.

@ZZYuting
Copy link

@ZZYuting The number of layers has dropped to 3 comparing with the ST-GCN with 10 layers. I have no idea how to light the model without loss of accuracy.
Although it has dropped to 3 compared with the st-gcn, the flops i caculate the model still up to 20G+flops, you can read some paper in cvpr 2020

@kenziyuliu
Copy link
Owner

kenziyuliu commented Jul 3, 2020

My personal opinions:

  • It seems that the performance on NTU RGB+D 60 dataset is getting saturated (e.g. suggested by this preprint, section 4.1) even with model ensembles, which are used by most work on this topic now; I would suggest working on NTU RGB+D 120 and Kinetics Skeleton 400 for future work.
  • One future direction is to look at how to make MS-G3D more efficient (as @ZZYuting has pointed out). Some components that slow down training include the spatial-temporal window graph/feature construction and the use of graph residual masks. Please also check our supplementary material where we discuss some future directions.

@desertfex
Copy link
Author

@kenziyuliu Thank you very much indeed.

@kenziyuliu
Copy link
Owner

I'll close this issue now but please feel free to add more comments to this thread.

@Lyndsey-Xing
Copy link

@ZZYuting The number of layers has dropped to 3 comparing with the ST-GCN with 10 layers. I have no idea how to light the model without loss of accuracy.
Although it has dropped to 3 compared with the st-gcn, the flops i caculate the model still up to 20G+flops, you can read some paper in cvpr 2020

Hello, how do you calculate flops in this paper? Is there any way to reduce flops?
Thanks.

@ZZYuting
Copy link

@ZZYuting The number of layers has dropped to 3 comparing with the ST-GCN with 10 layers. I have no idea how to light the model without loss of accuracy.
Although it has dropped to 3 compared with the st-gcn, the flops i caculate the model still up to 20G+flops, you can read some paper in cvpr 2020

Hello, how do you calculate flops in this paper? Is there any way to reduce flops?
Thanks.
I use the ptflops to calculate flops coarsely. In fact, you can calculate flops on your hand which is more precise. I think it's a good way to learn how to reduce flops if you know which parts is complex.

@desertfex
Copy link
Author

@ZZYuting Could you please provide the codes calculating the flops? I know its may be only two or three codes, but I still get "error" from the compiler. Thank you.

@ZZYuting
Copy link

ZZYuting commented Jul 16, 2020

from ptflops import get_model_complexity_info
flops, params = get_model_complexity_info(model, (3, 10, 25, 2), as_strings=True, print_per_layer_stat=True)
print("%s |%s |%s" % ('MSG3D', flops, params))

@desertfex
Copy link
Author

@ZZYuting Thank you very much.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants