Skip to content
#

vqgan-clip

Here are 31 public repositories matching this topic...

Translation of speech to image directly without text is an interesting and useful topic due to the potential application in computer-aided design, human to computer interaction, creation of an art form, etc. So we have focused on developing Deep learning and GANs based model which will take speech as an input from the user, analyze the emotions …

  • Updated Aug 25, 2024
  • Jupyter Notebook

VQGAN and CLIP are actually two separate machine learning algorithms that can be used together to generate images based on a text prompt. VQGAN is a generative adversarial neural network that is good at generating images that look similar to others (but not from a prompt), and CLIP is another neural network that is able to determine how well a c…

  • Updated Mar 15, 2022
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the vqgan-clip topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the vqgan-clip topic, visit your repo's landing page and select "manage topics."

Learn more