New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Why the argumentation transform of the orginal image contains colorjitter #16
Comments
EDIT The queue contains the key features for calculating the contrastive loss, not for the style vectors. The first element of a batch is the resized, randomly flipped, and normalized image. |
Got it. So you mean using 1) coloring jittering and 2) transformed samples to construct negative samples queue helps to learn the representation? |
You are right that we construct the positive and negative samples by using color jittering and transformation. I got your question related to color jittering. I adopted color jittering as a way to generate hard negative samples for contrastive learning and did not take into account the style vector view. However, I think that randomness of the parameters makes the color jittering less harmful to the style space in terms of the generator. (The color jittering operation applied with randomly selected parameters in (0.4, 0.4, 0.125) so that the effect of color jittering might not be harmful to the style vector extraction.) |
Got it. Thanks! |
Yes, I think so. But I also think that part might be compensated with random augmentation without the transformations not related to the color. |
Thanks for your reply :-)! |
Hi, thanks for sharing your code.
I have several questions about your design choice and looking forward to your reply.
I find the augmentation operation contains
transforms.ColorJitter(0.4, 0.4, 0.4, 0.125)
operation.Since the style information always includes the color, why you involve the ColorJitter operation and regard this transformation sample as the positive sample of the original image?
Will that influence the final results?
You use
x_k = data[1]
in
def initialize_queue(model_k, device, train_loader, feat_size=128)
This means you use a transformed image to extract style vector rather than the original image, why?
Looking forward to your reply, thanks!
The text was updated successfully, but these errors were encountered: