简体中文
优投实业(深圳)有限公司
以智慧赋能,点亮未来生活的每一刻
新闻中心
Decoding Sora: "World Model" reveals more clues
来源: | 作者:佚名 | 发布时间: 2024-02-26 | 23 次浏览 | 分享到:
After the release of ChatGPT more than a year ago sparked a global AI frenzy, global AI leader OpenAI has thrown a new "* bomb" into the industry - the AI cultural video model Sora, just a few hours after Google released the multimodal model Gemini1.5.

After the release of ChatGPT more than a year ago sparked a global AI frenzy, global AI leader OpenAI has thrown a new "* bomb" into the industry - the AI cultural video model Sora, just a few hours after Google released the multimodal model Gemini1.5.

In the demonstration of Gemini1.5, machines can "read" various human actions: when the tester takes out a piece of paper, Gemini1.5 immediately answers "You took out a piece of paper"; Sora can generate a one minute high-definition video based on short text prompts, which includes delicate and complex scenes, vivid character expressions, and complex camera movements. Two recently released AI models indicate that machines have taken a step closer to understanding human language and text, "said Fang Shuai, an analyst at the Shanghai Institute of Artificial Intelligence

OpenAI stated in a detailed technical report on Sora that it has a strong ability to understand natural language. OpenAI sees it as the foundation for models that can understand and simulate the real world, and believes its ability is an important milestone in achieving AGI (General Artificial Intelligence).

How can AI "Wensheng Video" change the future? As industry experts have pointed out, Sora's impact can be described as a breakthrough of an order of magnitude in Wensheng Video Technology, which can be directly applied to industries such as short videos and advertising; On a larger scale, it provides the possibility of understanding, reconstructing, and simulating this world.