Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why clip_guided works better than text2im, inconsistent with the paper's claim? #19

Open
a43992899 opened this issue Feb 2, 2022 · 2 comments

Comments

@a43992899
Copy link

I have tested the glide model for a few days (I tried many kinds of prompts), and my result is that clip_guided works better than classifier-free text2im.

clip_guided can correctly follow the intention of my prompt, like "a boat on the top of the mountain", or "Pablo Picasso: Into the wind", and text2im failed to do that.

However the paper claims that classifier-free text2im > clip_guided. I wonder why? Is there anything wrong with the released model?

@thunanguyen
Copy link

The glide-text2sim model that is better than clip_guided model is the one trained with a large private dataset and it is not released. Whereas, this released model is trained on a filtered and less diverse dataset.

@Arcitec
Copy link

Arcitec commented Feb 19, 2022

Exactly. A full explanation of the differences is here: #21 (comment)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants