Telegram (AI) YouTube Facebook X
Ру
OpenAI Unveils Sora: Enthusiasm and Challenges

OpenAI Unveils Sora: Enthusiasm and Challenges

On February 15, OpenAI introduced Sora, a new generative AI model that transforms text into video. The tool has generated excitement on social media, though it requires significant refinement before a full launch.

Sora can generate videos up to 60 seconds long with resolutions up to 1080p based on simple text prompts. These videos can include multiple characters, specific types of motion, and precise details of objects and backgrounds.

The tool is based on research from GPT and DALL-E 3. It operates using a diffusion model—transforming an initial image into statistical noise and then gradually removing this noise.

Developers acknowledge that Sora still has several shortcomings. It struggles to accurately model the physics of complex scenes, often confusing cause and effect.

“For example, a person might bite a cookie, but the bite mark might not appear,” OpenAI explains.

The tool also has issues with spatial details. The resulting video may not align with specified directions, mistaking right for left.

Currently, Sora is available to a “red team” of testers, as well as select designers, artists, and filmmakers.

Social Media Reaction

The tool has captivated social media, trending on X with over 173,000 posts.

To demonstrate the model’s capabilities, OpenAI CEO Sam Altman has begun accepting user requests for video generation. At the time of writing, he has shared a total of nine videos created by Sora.

AI experts have noted that Sora’s capabilities are “speechless.”

According to Nvidia senior research scientist Jim Fan, Sora is much more than just another “creative toy” like DALL-E 3. He described it as a “data-driven physics engine,” as the AI model not only generates abstract video but also intuitively creates the physics of objects within the scene.

Alongside this, some users have expressed concerns that tools like Sora could exacerbate the deepfake problem.

One user suggested that major social networks should consider built-in protection against realistic fakes, highlighting the threat of substituting video evidence of crimes.

Another user called for De-AI reverse engineering technology to prevent misinterpretation and misuse of content.

Previously, OpenAI began testing a “memory” feature for the ChatGPT chatbot, which retains information discussed in conversations to enhance user interaction.

Meanwhile, the company is actively countering the illegal use of its products.

Подписывайтесь на ForkLog в социальных сетях

Telegram (основной канал) Facebook X
Нашли ошибку в тексте? Выделите ее и нажмите CTRL+ENTER

Рассылки ForkLog: держите руку на пульсе биткоин-индустрии!

We use cookies to improve the quality of our service.

By using this website, you agree to the Privacy policy.

OK