TheSequence

TheSequence

Share this post

TheSequence
TheSequence
๐ŸŒ… Edge#229: VQGAN + CLIP
Copy link
Facebook
Email
Notes
More

๐ŸŒ… Edge#229: VQGAN + CLIP

+the original VQGAN+CLIP paper; +VQGAN+CLIP implementations

Sep 27, 2022
โˆ™ Paid
46

Share this post

TheSequence
TheSequence
๐ŸŒ… Edge#229: VQGAN + CLIP
Copy link
Facebook
Email
Notes
More
Share

In this issue:

  • we introduce VQGAN + CLIP architecture;

  • we discuss the original VQGAN+CLIP paper;ย 

  • we explore the VQGAN+CLIP implementations.ย 

Enjoy the learning!ย ย 


๐Ÿ’ก ML Concept of the Day: VQGAN + CLIPย 

In Edge#219, we introduced OpenAIโ€™s CLIP as a method that was able to represent both text and images in the same feature space. OpenAI published CLIP and DALL-E together but open-sourced only the former. Almost immediately after CLIP was made available, AI researchers started playing with the idea of combining it with generative adversarial networks (GANs). That evolution produced one of the most popular text-to-image synthesis in the current ecosystem: VQGAN+CLIP.ย ย 

The idea of combining CLIP and GANs is very intuitive. By

This post is for paid subscribers

Already a paid subscriber? Sign in
ยฉ 2025 Jesus Rodriguez
Privacy โˆ™ Terms โˆ™ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More