Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

customized inputs #29

Closed
Mingyuan1997 opened this issue Apr 19, 2022 · 6 comments
Closed

customized inputs #29

Mingyuan1997 opened this issue Apr 19, 2022 · 6 comments

Comments

@Mingyuan1997
Copy link

Is there any suggestion on how to run the model on customized input videos?

Thank you!

@yrcong
Copy link
Owner

yrcong commented Apr 20, 2022

Hi,
i think it is not difficult to use the model to infer the customized videos.
Line 167-178 in https://github.com/yrcong/STTran/blob/main/dataloader/action_genome.py may help you:)
best

@xiaodanhu
Copy link

If we test on custom video, it seems that the information like attention_relationship and bboxes are required?

@Mingyuan1997
Copy link
Author

I think so.

@yrcong
Copy link
Owner

yrcong commented Apr 25, 2022

If we test on custom video, it seems that the information like attention_relationship and bboxes are required?

Why? The attention relationships should be predicted and the bboxes should be inferred by the object detector.

@xiaodanhu
Copy link

If we test on custom video, it seems that the information like attention_relationship and bboxes are required?

Why? The attention relationships should be predicted and the bboxes should be inferred by the object detector.

Thanks for replying!
When I checked the dataloader in action_genome.py, the person_bbox and object_bbox were loaded from person_bbox.pkl and object_bbox_and_relationship.pkl, respectively. So I was guessing, if loading the customized video, we also need this information so that we can load it properly. Can you indicate where the model automatically predicts the relationship and bboxes? Thank you very much!

@yrcong
Copy link
Owner

yrcong commented Apr 29, 2022

If you just want to test on your customized video dataset, only the video frames (self.video_list in the class AG) are necessary (for the setting SGDET).
Sometimes person_bbox and object_bbox are used in the test because there are other two settings PredCLS/SGCLS which are widely used in the image scene graph generation.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants