Cosine similarity between CLIP-Reid features of this repo and original repo #1288
Closed
1 task done
Labels
question
Further information is requested
Search before asking
Question
Hi! Thanks for sharing this awesome work. I wanted to improve the features of CLIP-Reid model. The reason is that it is trained on Market1501 dataset which has similar lighting conditions for different instances of the same person. But when we try to reid a person from different lighting conditions (as the camera setup is some other area), sometimes it fails. So, I wanted to fine-tune the CLIP model for our dataset. But before doing so, just for a sanity check, I checked features from the CLIP model used in your repo and features from the CLIP original repo. There was a huge difference in terms of cosine similarity (calculated between different IDs). The features from this repo were much better (there was a considerable gap between the same IDs and different IDs) as compared to the features from the original repo (the gap was very low). So, my question is, did you retrain or fintune the original CLIP model or directly use it as it is (I supposed that you used it as it is)? Also, do you have any suggestions or comments for the real-world problem that we face (different lighting conditions for the same person id) while doing reid on people?
The text was updated successfully, but these errors were encountered: