-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
data processing #7
Comments
Thank you for your interest. Because I didn't have time to clean the feature extraction code so I can't publish it. |
Thank you for your answer.I want to implement your work on my own dataset, but it seems to be very difficult.I really need detailed information on data processing. |
For the videos in Activitynet, I simply rescale them to 1600 frames and extract features with a window size of 16 frames so that each video will be represented by a sequence of 100 features. The above processing in both videos follows the success experiments of BMN, G-TAD. But I observed that more recent methods in temporal action detection can now work on original size videos in both datasets. |
Thank you for your answer. Is the sliding window non overlapping for handling thumos14? |
Sorry, I think my last comment gets misrendered. |
May I ask what network is used to extract environmental features and participant features, what is the format of the extracted data, and how is it converted into the current data input format of the aoe-net network? Could you please show me some details of data processing?
The text was updated successfully, but these errors were encountered: