Beyond Uncanny Valley: Breaking Down Sora
a16z PodcastIn early 2024, the notion of high fidelity, believable AI-generated video seemed a distant future to many. Yet, a mere few weeks into the year, OpenAI unveiled Sora, its new state of the art text-to-video model producing videos of up to 60 seconds. The output shattered expectations – even for other builders and researchers within generative AI – sparking widespread speculation and awe.How does Sora achieve such realism? And are explicit 3D modeling techniques or game engines at play?In this episode of the a16z Podcast, a16z General Partner Anjney Midha connects with Stefano Ermon, Professor of Computer Science at Stanford and key figure at the lab behind the diffusion models now used in Sora, ChatGPT, and Midjourney. Together, they delve into the challenges of video generation, the cutting
0
(-)
Rate this episode:
Episode mentions
You can listen to this podcast, but not everything mentioned in it has been defined yet. Upvote it if you want it to be added next - the most voted podcasts will be added as soon as possible
Transcript
This episode has not been transcribed yet. Upvote it if you want it to be added next - the most voted podcasts will be added as soon as possible.