-
Notifications
You must be signed in to change notification settings - Fork 246
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Keypoints to check for New training ? #10
Comments
@angenieux Hi, the size of the sliding window is defined in this script src/mylib/feature_proc.py. Its constructor is shown below. You can see the parameter FEATURE_T_LEN=5, which is the window size:
This class is used for extracting features in both training and inference. So this FEATURE_T_LEN is the only thing you need to change. |
HI ! I would like to evaluate the prediction results regarding the number of frames used for the temporal windows. Does the FEATURE_T_LEN values must be the same for training AND inference ? I have trained my data with FEATURE_T_LEN=5. Traceback (most recent call last):python3 run_detector.py --source webcam File "run_detector.py", line 339, in Thanks for your help |
@angenieux Hi, the training and inference must use the same length of feature vector. So the FEATURE_T_LEN must be the same. |
Hi One last question. Can you give me any advice regarding the FEATURE_T_LEN value to use for training and inference ? best regards |
@angenieux Hi. First of all, since your inference speed is 5 fps, you need to lower down the video input to 5 fps (no matter how fast the original video streaming is). You can do this either by triggering the video frame at 5 fps, or manually sampling the video frame at 5 fps. As for the FEATURE_T_LEN, since your shortest action is only 0.5s, which corresponding to only 0.5s*5fps=2.5 frame. So you may use 2 frame for FEATURE_T_LEN. |
@angenieux Hi, I just refactored the code to make it more readable, as well as making the API easier to use. Just for your information. Thanks for the support! The FEATURE_T_LEN is now renamed as |
Hi, thanks for this huge work!!
My intend is to test your solution on a JETSON NANO (inference only)-->
Your pre-trained model runs at 3.4 FPS with realtime webcam video input (224x224, mobilenet_v2) on my device = Great
I've prepared my own training set of images based on video, but the movie rate is 25FPS (not 10 FPS).
As you used a 5 Frames sliding windows, what are the keypoints to check in the code :
-For training ?
-For Inference on nano?
Thanks
The text was updated successfully, but these errors were encountered: