Today we’re releasing Sora 2, our flagship video and audio generation model.
The original Sora model from February 2024 was in many ways the GPT‑1 moment for video—the first time video generation started to seem like it was working, and simple behaviors like object permanence emerged from scaling up pre-training compute. Since then, the Sora team has been focused on training models with more advanced world simulation capabilities. We believe such systems will be critical for training AI models that deeply understand the physical world. A major milestone for this is mastering pre-training and post-training on large-scale video data, which are in their infancy compared to language.
With Sora 2, we are jumping st…
Today we’re releasing Sora 2, our flagship video and audio generation model.
The original Sora model from February 2024 was in many ways the GPT‑1 moment for video—the first time video generation started to seem like it was working, and simple behaviors like object permanence emerged from scaling up pre-training compute. Since then, the Sora team has been focused on training models with more advanced world simulation capabilities. We believe such systems will be critical for training AI models that deeply understand the physical world. A major milestone for this is mastering pre-training and post-training on large-scale video data, which are in their infancy compared to language.
With Sora 2, we are jumping straight to what we think may be the GPT‑3.5 moment for video. Sora 2 can do things that are exceptionally difficult—and in some instances outright impossible—for prior video generation models: Olympic gymnastics routines, backflips on a paddleboard that accurately model the dynamics of buoyancy and rigidity, and triple axels while a cat holds on for dear life.
Prompt: a guy does a backflip
Prior video models are overoptimistic—they will morph objects and deform reality to successfully execute upon a text prompt. For example, if a basketball player misses a shot, the ball may spontaneously teleport to the hoop. In Sora 2, if a basketball player misses a shot, it will rebound off the backboard. Interestingly, “mistakes” the model makes frequently appear to be mistakes of the internal agent that Sora 2 is implicitly modeling; though still imperfect, it is better about obeying the laws of physics compared to prior systems. This is an extremely important capability for any useful world simulator—you must be able to model failure, not just success.
The model is also a big leap forward in controllability, able to follow intricate instructions spanning multiple shots while accurately persisting world state. It excels at realistic, cinematic, and anime styles.
Prompt: intense anime battle between a boy with a sword made of blue fire and an evil demon demon
As a general purpose video-audio generation system, it is capable of creating sophisticated background soundscapes, speech, and sound effects with a high degree of realism.
Prompt: Two mountain explorers in bright technical shells, ice crusted faces, eyes narrowed with urgency shout in the snow, one at a time
You can also directly inject elements of the real world into Sora 2. For example, by observing a video of one of our teammates, the model can insert them into any Sora-generated environment with an accurate portrayal of appearance and voice. This capability is very general, and works for any human, animal or object.
Prompt: Bigfoot is really kind to him, a little too kind, like oddly kind. Bigfoot wants to hang out but he he wants to hang too much
The model is far from perfect and makes plenty of mistakes, but it is validation that further scaling up neural networks on video data will bring us closer to simulating reality.
Deployment of Sora 2
On the road to general-purpose simulation and AI systems that can function in the physical world, we think people can have a lot of fun with the models we’re building along the way.
We first started playing with this “upload yourself” feature several months ago on the Sora team, and we all had a blast with it. It kind of felt like a natural evolution of communication—from text messages to emojis to voice notes to this.
So today, we’re launching a** new social iOS app just called “Sora,”** powered by Sora 2. Inside the app, you can create, remix each other’s generations, discover new videos in a customizable Sora feed, and bring yourself or your friends in via cameos. With cameos, you can drop yourself straight into any Sora scene with remarkable fidelity after a short one-time video-and-audio recording in the app to verify your identity and capture your likeness.
Last week, we launched the app internally to all of OpenAI. We’ve already heard from our colleagues that they’re making new friends at the company because of the feature. We think a social app built around this “cameos” feature is the best way to experience the magic of Sora 2.
Launching responsibly
Concerns about doomscrolling, addiction, isolation, and RL-sloptimized feeds are top of mind—here is what we are doing about it.
We are giving users the tools and optionality to be in control of what they see on the feed. Using OpenAI’s existing large language models, we have developed a new class of recommender algorithms that can be instructed through natural language. We also have built-in mechanisms to periodically poll users on their wellbeing and proactively give them the option to adjust their feed.
By default, we show you content heavily biased towards people you follow or interact with, and prioritize videos that the model thinks you’re most likely to use as inspiration for your own creations. We are not optimizing for time spent in feed, and we explicitly designed the app to maximize creation*,*not consumption. You can find more details in our Feed Philosophy
This app is made to be used with your friends. Overwhelming feedback from testers is that cameos are what make this feel different and fun to use—you have to try it to really get it, but it is a new and unique way to communicate with people. We’re rolling this out as an invite-based app to make sure you come in with your friends. At a time when all major platforms are moving away from the social graph, we think cameos will reinforce community.
Protecting the wellbeing of teensis important to us. We are putting in default limits on how many generations teens can see per day in the feed, and we’re also rolling out with stricter permissions on cameos for this group. In addition to our automated safety stacks, we are scaling up teams of human moderators to quickly review cases of bullying if they arise. We are launching with Sora parental controls via ChatGPT so parents can override infinite scroll limits, turn off algorithm personalization, as well as manage direct message settings.
With cameos, you are in control of your likeness end-to-end with Sora. Only you decide who can use your cameo, and you can revoke access or remove any video that includes it at any time. Videos containing cameos of you, including drafts created by other people, are viewable by you at any time.
There are a lot of safety topics we’ve tackled with this app—consent around use of likeness, provenance, preventing the generation of harmful content, and much more. See our Sora 2 Safety doc for more details.
A lot of problems with other apps stem from the monetization model incentivizing decisions that are at odds with user wellbeing. Transparently, our only current plan is to eventually give users the option to pay some amount to generate an extra video if there’s too much demand relative to available compute. As the app evolves, we will openly communicate any changes in our approach here, while continuing to keep user wellbeing as our main goal.
We’re at the beginning of this journey, but with all of the powerful ways to create and remix content with Sora 2, we see this as the beginning of a completely new era for co-creative experiences. We’re optimistic that this will be a healthier platform for entertainment and creativity compared to what is available right now. We hope you have a good time :)
Sora 2 availability and what’s next
The Sora iOS app(opens in a new window) is available to download now. You can sign up in-app for a push notification when access opens for your account. We’re starting the initial rollout in the U.S. and Canada today with the intent to quickly expand to additional countries. After you’ve received an invite, you’ll also be able to access Sora 2 through sora.com(opens in a new window). Sora 2 will initially be available for free, with generous limits to start so people can freely explore its capabilities, though these are still subject to compute constraints. ChatGPT Pro users will also be able to use our experimental, higher quality Sora 2 Pro model on sora.com(opens in a new window) (and soon in the Sora app as well). We also plan to release Sora 2 in the API. Sora 1 Turbo will remain available, and everything you’ve created will continue to live in your sora.com(opens in a new window) library.
Video models are getting very good, very quickly. General-purpose world simulators and robotic agents will fundamentally reshape society and accelerate the arc of human progress. Sora 2 represents significant progress towards that goal. In keeping with OpenAI’s mission, it is important that humanity benefits from these models as they are developed. We think Sora is going to bring a lot of joy, creativity and connection to the world.
— Written by the Sora Team
Sora 2
ResearchHarold Li, Dmytro Okhonko, Avi Verma, Eric Zhang, Ricky Wang, Troy Luhman, Eric Luhman, Bram Wallace, Eric Mintun, Michael Chang, Gabriel Petersson, Jure Zbontar, Daniel Geng, Will DePue, Alex Zhao, Cheng Lu, Yufei Guo, Pritam Damania, Larry Kai, Farzad Khorasani, Kenji Hata, James Betker, Vladimir Chalyshev, Connor Holmes, Aditya Ramesh, Bill Peebles
ProductAndrew Kondrich, Andrew Sima, Andrew Thieck, Andrey Malevich, Antonio Di Francesco, Bin Wen, Bing Liang, Boyang Niu, Cheng Su, Cristina Scheau, Daniel Latta-Lin, David Schnurr, Dhruba Borthakur, Duc Tran, Gilman Tolle, Greg Hochmuth, Joe Taylor, Joey Flynn, Joey Pereira, Julius Hochmuth, Key Shin, Liam Esparraguera, Liang Wu, Liang Xiong, Mengchao Zhong, Michelle Hwang, Mick Jermsurawong, Mike Starr, Omar Elfanek, Patrick Hum, Pavel Komlev, Rajeev Nayak, Raunak Daga, Rohan Sahai, Sergii Rudenko, Shuyi Chen, Tarek Younes, Thomas Bredillet, Thomas Bredillet, Thomas Dimson, Victoria Huang, Vladimir Chalyshev, Welton Wang, Wesam Manassra, Xiaolong Wang, Yizhe Yu, Yun Jiang, Zhigang Wang
ContributorsAarash Heydari, Chad Nelson, Daniel Fradin, David Duxin, Hessam Bagherinezhad, Jasmyn Samaroo, Jay Wang, Jess Manzano, Kendra Rimbach, Nikki Sommer, Sergei Vorobev, Shirong Wu, Soham Govande, Souki Mansoor, Tifa Chen, Tomer Kaftan, Tyce Walters, Varun Shetty
Leadership
Bill Peebles
Sora
Connor Holmes
Systems
Rohan Sahai
Product
Thomas Dimson
Product
Natalie Summers
Chief of Staff
Aditya Ramesh
Organization
Special ThanksAdam Majmudar, Adele Li, Aravind Suresh, Arun Vijayvergiya, Ashkay Pall, Ben Leimberger, Brad Lightcap, Charlotte Cole, Chris Hallacy, Chris Koch, Christine McLeavey, Christopher Lehane, Dane Stuckey, Eric Wallace, Fidji Simo, Gabriel Goh, Gary Briggs, Geoff Salmon, Giancarlo Lionetti, Greg Brockman, Hannah Wong, Ian Sohl, Jakub Pachocki, Jamie Kiros, Jason Kwon, Jeffrey Han, Joanne Jang, Johannes Heidecke, Josh Achiam, Kate Rouch, Kevin Weil, Lauren Itow, Li Jing, Mark Chen, Mark Gewurz, Matt Knight, Matthew Isono, Max Burkhardt, Mayank Gupta, Mia Glaese, Nick Turley, Patrick Geonetta, Peter Welinder, Philip Bogdanov, Prafulla Dhariwal, Robert Xiong, Ryan O’Rourke, Sam Altman, Sarah Friar, Sarah Russell, Sarah Warkov, Specer Papay, Srinivas Narayanan, Sulman Choudhry, Szymon Sidor, Tejal Patwardhan, Vikki Lampton, Vlad Fomenko, Wojciech Zaremba, Young Cha, Yuchen Zhang
Safety, Integrity, Product Policy, i2, User OpsAdam Wells, Aleah Houze, Annie Cheng, Artyi Xu, Carolina Paz, Claudia Fischer, Garrett Harkins, Gilman Tolle, Jackie Hehir, Jake Brill, Jesika Haria, Kate Birks, Kelly Stirman, Lauren Jonas, Mentong Zhang, Pedram Keyani, Pedro Aguilar, Ryan Rinaldi, Sam Toizer, Sarah Ryan, Savannah Heon, Shalli Jain, Shauna O’Brien, Tim Boll, Zoe Stoll
LegalTyce Walters, Ali Buttars, Brian McKnight, Gideon Myles, Tom Rubin, Dani Westbrook, Charles Proctor
CommunicationsAlex Baker-Whitcomb, Anna McKean, Ashley Tyra, Bailey Richardson, Gaby Raila, Julie Steele, Leah Anise, Niko Felix
Marketing, Design, & CreativeAdam Brandon, Adrian Gunadi, Alexandr Khomyakov, Anne Oburgh, Antonia Richmond, Ben King, Cary Hudson, Chloe Bowers, Chris Hutchinson, Ciaran Rogers, Dalhae Lee, Dana Palmie, Daniel Stuhlpfarrer, Daniel Zhang, Elisha Greenwell Dargan, Ian Silber, Indgila Sama Ali, Jeffrey Sabin-Matsumoto, Josh Cleveland, Kaitlin Giannetti, Kenneth Kuh, Kim Baschet, Malisa Kuch, Melia Tandiono, Michaela McCrink, Minnia Feng, Nick Ciffone, Paymon Parsia, Phillip Kim, Phillip Kim, Raegan Allsbrook, Roy Chen, Shannon Jager, Thomas Degry, Xingle Huang, Yara Khakbaz, Zach Stubenvoll
Global AffairsClaudia Fischer Debbie Mesloh
Strategic FinanceChengpeng Mou Caroline Zhao
APIAdam Wells, Alina Wu, Amelia Liu, Andi Liu, Ankit Gohel, Annie Cheng, Artyi Xu, Brian Ratajczak, Chad Nelson, Erika Kettleson, Filippo Raso, Gilman Tolle, Jackie Hehir, Jeff Harris, Jen Robinson, Joanne Shin, Jono Oko, Katia Gil Guzman, Kelly Stirman, Leher Pathak, Manoli Liodakis, Miqdad Jaffer, Olivia Morgan, Robin Koenig, Rohan Sahai, Ruth Costigan, Sarah Ryan, Savannah Heon, Shaokyi Amdo, Shaili Jain, Tabarak Khan, Tonia Osadebe, Tyce Walters, Wei Sun, Wendy Jiao, Woo Kim, Yi Ma
Built by OpenAI in San Francisco, California Published September 30, MMXXV