TheSequence

TheSequence

Share this post

TheSequence
TheSequence
📝➡️📺 Edge#234: Inside Meta AI’s Make-A-Video

📝➡️📺 Edge#234: Inside Meta AI’s Make-A-Video

The new model builds on the principles of text-to-image methods to produce visually astonishing videos

Oct 13, 2022
∙ Paid
24

Share this post

TheSequence
TheSequence
📝➡️📺 Edge#234: Inside Meta AI’s Make-A-Video
Share

On Thursdays, we dive deep into one of the freshest research papers or technology frameworks that is worth your attention. Our goal is to keep you up to date with new developments in AI to complement the concepts we debate in other editions of our newsletter.

💥 What’s New in AI: Inside Meta AI’s Make-A-Video – The New Super Model that can Generate Videos from Textual Inputs

Image credit: Meta AI

Text-to-Video (T2V) is considered the next frontier for generative artificial intelligence (AI) models. While the text-to-image (T2I) space is experiencing a revolution with models like DALL-E, Stable Diffusion, and Midjouney, T2V still remains a monumental challenge. Recently, researchers from Meta AI unveiled Make-A-Video, a T2V model able to create realistic short video clips from textual inputs.

The T2V space is perfectly positioned to benefit from the advancements in T2I architectures, but there are still monumental challenges ahead. For starters,

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Jesus Rodriguez
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share