-
Notifications
You must be signed in to change notification settings - Fork 100
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Could you tell me how to train on oxford5k? #16
Comments
"Training" there means k-means vocabulary learning, not the training HardNet itself. |
Sorry, I just understand your method. Firstly you use HardNet model trained on UBS dataset to extract 128 features. Then, K-means method is used to learn vocabulary on Oxford5k dataset. At the end, you use the BOW model test on Paris6k. But I still confused how you deal with "Junk" images? Do you treat them as null images? |
@filipradenovic was doing this experiment. But if I remember correctly, they were ignored - as suggested in original paper protocol. Btw, there is new protocol here https://github.com/filipradenovic/revisitop |
@shanYanGuan junk images are always ignored, ie treated as null images. However, in the new protocols described in https://github.com/filipradenovic/revisitop, null set is different depending on the protocol difficulty. More details can be found in the paper. |
Thank you bro! @ducha-aiki @filipradenovic |
Well... first, datasets are quite small. Second, we want to see how method generalize. |
ok, I see. Hardnet is to extract an image descriptor. This descriptor is designed to deal with matching problem. So I thought it should contain the key point informations. And Patch Datasets is more about viewpoint or illumination changes on the same objects. But image retrieval is to find the same class object not the same objects. So why hardnet can do well in Oxford5k? |
the key point is I think patch match task is different from image retrieval, for the reason that patch match is to fine the same object with different viewpoint or illumination, but image retrieval is to find the same image with same semantics So the descriptor for patch match can't adopt to image retrieval. |
You are wrong here. Image retrieval, at least in classical computer vision sense is about finding SAME object under different conditions, don`t mix it with "similarity search". E.g. see here https://www.kaggle.com/c/landmark-retrieval-challenge In this competition, Kagglers are given query images and, for each query, are expected to retrieve all database images containing the same landmarks (if any)." |
@shanYanGuan but you are right, that actually Oxford5k task is "particular object retrieval" or "particular instance retrieval" to be precise. It is just community used to call it "image retrieval". |
You mean image retrieval is to retrieve all database images containing the same content? But image retrieval based on Hash method seems different. For example, cifair10 is to retrieve the same class image. But image content in same class is inconsistent. |
cifar10... sorry for wrong split. E.g. below images is airplane, but is different airplane. |
CIFAR is classification problem, not retrieval. Or you can call it "category retrieval". And Oxford5k is "instance retrieval". |
So Nework for match cannot be used to category retrieval. Am I right? |
@shanYanGuan yes, you are right. |
Sounds strange. And if match network just can get similarity about same object not same category, what meaning it has? |
@shanYanGuan when you are looking for YOUR phone or dog, do you want to find your phone or dog, or any kind would suffice? |
OK, but why match network can't get high semantics, such as content that image have? If they can get some high semantics clues, they should be able to classify if they are same category. |
I don`t know this for sure. I guess, that networks do only things, which we train them for. If we train them for semantics, they do semantics. If we train for correspondences - they do correspondences, etc. These are just different tasks. Why trucks cannot compete with F1 cars? |
ok. I can get you. Maybe I should do some experiments to solve this problem. Thanks for your instructive answer! |
From your paper, it could be find that u evaluate paris dataset with model trained on Oxford5k. But i don't know what protocol you set on Oxford5k and Paris.
Have you cropped Oxford5k' query images? And Junk images are splited to no-match images?
The text was updated successfully, but these errors were encountered: