Home MANAGEMENT OpenAI’s New Breakthrough Technology: Transforms your descriptive sentences into lifelike videos effortlessly

OpenAI’s New Breakthrough Technology: Transforms your descriptive sentences into lifelike videos effortlessly

OpenAI: Creating video from text - Sora is an AI model that can create realistic and imaginative scenes from text instructions.

We’re teaching AI to understand and simulate the physical world in motion, with the goal of training models that help people solve problems that require real-world interaction. Introducing Sora, our text-to-video model. Sora can generate videos up to a minute long while maintaining visual quality and adherence to the user’s prompt. – OpenAI

On Thursday, OpenAI made headlines by unveiling its latest innovation: Sora, a revolutionary video-generation AI that goes beyond text and images. With Sora, users can now input descriptions of scenes and watch as the AI transforms them into stunning, high-definition video clips.

This groundbreaking technology introduces new challenges for platforms grappling with misinformation, particularly in the lead-up to crucial global elections scheduled for this year.

Following its meteoric rise with ChatGPT, OpenAI is now venturing into the video domain with its cutting-edge artificial intelligence.

Prompt: Several giant wooly mammoths approach treading through a snowy meadow, their long wooly fur lightly blows in the wind as they walk, snow covered trees and dramatic snow capped mountains in the distance, mid afternoon light with wispy clouds and a sun high in the distance creates a warm glow, the low camera view is stunning capturing the large furry mammal with beautiful photography, depth of field.

The company introduced Sora, its new generative AI model

Sora works similarly to OpenAI’s image-generation AI tool, DALL-E. A user types out a desired scene and Sora will return a high-definition video clip.

Sora can also generate video clips inspired by still images, and extend existing videos or fill in missing frames.

Sora is currently limited to generating videos that are a minute long or less. OpenAI, backed by Microsoft, has made multimodality — the combining of text, image and video generation — a goal in its effort to offer a broader suite of AI models.

Sora is able to generate complex scenes with multiple characters, specific types of motion, and accurate details of the subject and background. The model understands not only what the user has asked for in the prompt, but also how those things exist in the physical world – OpenAI

Prompt: A cat waking up its sleeping owner demanding breakfast. The owner tries to ignore the cat, but the cat tries new tactics and finally the owner pulls out a secret stash of treats from under the pillow to hold the cat off a little longer.

Sora has thus far only been available to a small group of safety testers, or “red teamers,” who test the model for vulnerabilities in areas such as misinformation and bias. The company hasn’t released any public demonstrations beyond 10 sample clips available on its website, and it said its accompanying technical paper will be released later on Thursday.

Sora is a diffusion model, which generates a video by starting off with one that looks like static noise and gradually transforms it by removing the noise over many steps. -OpenAI

“Sora serves as a foundation for models that can understand and simulate the real world,” – OpenAI

Click Here for details:

Exit mobile version