Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Post processing results worse than original #36

Open
siddtmb opened this issue Mar 12, 2024 · 0 comments
Open

Post processing results worse than original #36

siddtmb opened this issue Mar 12, 2024 · 0 comments

Comments

@siddtmb
Copy link

siddtmb commented Mar 12, 2024

I used ultralytics yolov8 to label some data using the predict function on a video. I set the confidence threshold to 0.01 while predicting. I was only tracking a single class so I set the scores = [majority_class_score, (1 - majority_class_score)/5, (1 - majority_class_score)/5, (1 - majority_class_score)/5, (1 - majority_class_score)/5] (essentially simulating the main class plus 4 dummy classes with lower probabilities).

The unfortunate part is that when I run REPP on these detections, I actually get worse detections. More flickering. It flickers or drops the objects even when the original predict function (without any other tracking or post processing) produces a label with a confidence score of 0.8 to 0.95 in the exact same frame.

Config file:

{
	"min_tubelet_score": 5e-3,
	"add_unmatched": false,
	"distance_func": "logreg",
	"clf_thr": 0.5,
	"clf_mode": "dot",
	"recoordinate": true,
	"recoordinate_std": 0.6,
	"min_pred_score":  5e-3,
	"appearance_matching": false
}

I tried a bunch of hyperparameters, including clf_thr: 0.00001 and different re-coordinate values, but perhaps I am doing something wrong. Why is it deteriorating results over even the base predictor? Please help I would be eternally grateful to you as this is very much needed for our project. We can even pay you some amount for advice on getting this set up if this works well and you are willing to help us.

I even tried different ways of defining the score such as only including the confidence for the main class, or increasing to 19 dummy classes with low probabilities + the main class, etc.

I believe that your work is very great and it is me who is doing something wrong in using it. Please let me know if you can help point out some fundamental flaw in my approach.

Please help with this. Thank you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant