OLDStyleCLIPS: Empowering Image Editing with Textual Guidance using StyleGAN Latent Space Optimization, CLIP, and Image Segmentation
Khalil Haroun Achache, Farah Briki, Haitao Zhou
This project aims at performing text based editing on StyleGAN2 images. It was done in the context of the course of Computational Photography (CS-413) at EPFL. This project is based on the paper StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery by Or Patashnik, Zongze Wu, Eli Shechtman, Daniel Cohen-Or, Dani Lischinski, and Daniel Cohen and LELSD: OPTIMIZING LATENT SPACE DIRECTIONS FOR GAN-BASED LOCAL IMAGE EDITING by Ehsan Pajouheshgar, Tong Zhang, and Sabine Süsstrunk.
notebooks/: contains the notebooks used for the projectnotebooks/playground.ipynb: the notebook to run on colab to perform image editingnotebooks/inversion_playground.ipynb: the notebook to run on colab to perform image inversion, i.e. find the latent code of an image
models/: contains the wrappers for the models used in the projectutils.py: contains the utility functions used in the projectlicenses/: contains the licenses of the models used in the projectimg/: contains the images for this READMEoptimization/: contains the code for running the main method of the projectpretrained/: contains the weights of some of the pretrained models used in the project (that are uploadable to github, the others are hosted on google drive)criteria/: contains wrappers for the different losses used for optimization