OpenAI - for all its superlatives - acknowledges the model isn’t perfect. And these videos maintain reasonable coherence in the sense that they don’t always succumb to what I like to call “AI weirdness,” like objects moving in physically impossible directions.Ĭheck out this tour of an art gallery, all generated by Sora (ignore the graininess - compression from my video-GIF conversion tool): But the cherry-picked samples from the model do look rather impressive, at least compared to the other text-to-video technologies we’ve seen.įor starters, Sora can generate videos in a range of styles (e.g., photorealistic, animated, black and white) up to a minute long - far longer than most text-to-video models. Now, there’s a lot of bombast in OpenAI’s demo page for Sora - the above statement being an example. “ The model understands not only what the user has asked for in the prompt, but also how those things exist in the physical world.” “Sora has a deep understanding of language, enabling it to accurately interpret prompts and generate compelling characters that express vibrant emotions,” OpenAI writes in a blog post. Sora can also “extend” existing video clips - doing its best to fill in the missing details. Given a brief - or detailed - description or a still image, Sora can generate 1080p movie-like scenes with multiple characters, different types of motion and background details, OpenAI claims. OpenAI today unveiled Sora, a generative AI model that creates video from text. OpenAI, following in the footsteps of startups like Runway and tech giants like Google and Meta, is getting into video generation.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |