Replies: 6 comments 23 replies
-
Someone's already got it working, phew! |
Beta Was this translation helpful? Give feedback.
-
Was Google's code the basis of the instruct_to_pix extension in A1111, or is it different/better? |
Beta Was this translation helpful? Give feedback.
-
Another great repo implementing this https://github.com/cloneofsimo/inversion_edits original ddim inversion is almost perfect plus blue boots - looks a bit strong / too high cfg here The repo references this paper which is really good - injecting at different layers has interesting results! So, new feature idea - a way to add emphasis to different prompts at different layers? |
Beta Was this translation helpful? Give feedback.
-
I'm able to duplicate his results with a quick-hack A1111 Extension, where I couldn't even get Google's implementation to compile, but he's got something wrong with his null-text inversion (middle image) and I think that's affecting the output. I tried with lower CFG scales, so it's not the CFG scale. Notice how his inversion image is missing part of the legs. I assume this then affects the re-prompting, which puts the "boots" all over the horse. The results are even worse if using a different image ... If anyone knows what he did wrong, please chime in!! :D But here are his results duplicated in A1111. |
Beta Was this translation helpful? Give feedback.
-
quick one - is anyone working on porting null textual inversion to |
Beta Was this translation helpful? Give feedback.
-
@Luke2642 Thanks for your implementation! Just one note. For the invert function, you/anyone might want to add a grad scaler! I might do a pr if I get the time |
Beta Was this translation helpful? Give feedback.
-
Google's null text inversion produces a near perfect textual inversion and allows prompt to prompt editing on any model!
I've been trying to get it working on google colab, I'm sort of half way there. I've sorted the xformers requirements, and reduced it down to fp16 so it'll fit in the T4 16GB memory, and I'm getting images from the VAE, but I've got some sort of float/half precision problem that means it's only generating black images from the latent. If any pytorch people have any advice I'd really appreciate it!
https://github.com/Luke2642/prompt-to-prompt-colab/blob/main/null_text_w_ptp_colab_fp16.ipynb
You'll have to paste in your own huggingface token to download the sd1.4 model. Once I've got it working in colab I can think about implementing it as an extension or feature.
Could anyone verify it actually works on colab with fp32 with colab pro? The original site/paper is https://null-text-inversion.github.io/
Beta Was this translation helpful? Give feedback.
All reactions