Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How was the fine-tunimg done exactly #32

Closed
travellingsasa opened this issue Apr 29, 2024 · 5 comments
Closed

How was the fine-tunimg done exactly #32

travellingsasa opened this issue Apr 29, 2024 · 5 comments

Comments

@travellingsasa
Copy link

Hey there,

I am wondering how you did the fine-tuning here. You do not describe it in the paper.

Did you

  1. Continue training starting from a pre trained model like open clip
  2. Added a classification head and froze all other layer
  3. Added a classification head and updated image and text encoders

I don't think you did 2 or 3 since you used full sentences as captions.

How did you do it?

All the best

@vinid
Copy link
Collaborator

vinid commented Apr 29, 2024

Hi!!

It's contrastive fine-tuning, we use the same task CLIP was trained on. All unfrozen.

Let me know if you need more details!

@travellingsasa
Copy link
Author

So when you say "same task CLIP was trained on" do I correctly assume you continued training without adding a classifier?

@vinid
Copy link
Collaborator

vinid commented Apr 30, 2024

Yup, we keep the same contrastive pre-training objective

@travellingsasa
Copy link
Author

Thank you for the clarification and the super quick reply :)

@vinid
Copy link
Collaborator

vinid commented Apr 30, 2024 via email

@vinid vinid closed this as completed May 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants