Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bulldozing the latent space on the way to your token #15

Open
JoePenna opened this issue Sep 12, 2022 · 1 comment
Open

Bulldozing the latent space on the way to your token #15

JoePenna opened this issue Sep 12, 2022 · 1 comment

Comments

@JoePenna
Copy link

JoePenna commented Sep 12, 2022

It seems like the entire latent space is shifted towards what you're training. And also, the longer you train, the more is affected.

However, @nikopueringer was figuring out what's missing in @XavierXiao's code from Google's implementation -- regularization on the go.

To be reductive, it's:

  • Generate an image from the original ckpt
  • Move the new ckpt toward your class / face
  • Generate the same image from step 1 from the new ckpt
  • If it's too far, rewind... try again...

As a test, I trained my face on the class word "brazilian". At 9K steps, here are some unrelated prompts (euler, seed 1, cfg 15):

photo of an apple:

man:

brazilian:

annakendrick:

kit harington:

photo of a horse:

Some of the issues above might be ameliorated by removing "photo of" from the personalized.py file? Or with more regularization images, perhaps as many images as there are steps? Or perhaps a much much narrower class?

@JoePenna
Copy link
Author

A find:

By removing the "photo of" from the training:

We're able to salvage some of the latent space.

Here's "photograph of an apple"
photograph-of-an-apple-0023

As opposed to the same settings, but trained with the unaltered personalized.py file:
photograph-of-an-apple-0025

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant