Telegram (AI) YouTube Facebook X
Ру

Chinese Startup DeepSeek Unveils AI Model Surpassing Meta and OpenAI

Chinese AI startup DeepSeek has introduced its own large language model, which has outperformed competitors from Meta and OpenAI in tests.

DeepSeek V3 boasts 671 billion parameters. In comparison, Llama 3.1 405B has 405 billion. This figure reflects the AI’s ability to adapt to more complex applications and provide more accurate responses.  

Chinese Startup DeepSeek Unveils AI Model Surpassing Meta and OpenAI
Comparison of DeepSeek V3 with competitors. Data: DeepSeek.

The Hangzhou-based company trained the neural network in two months with $5.58 million, using significantly fewer computational resources (2048 GPUs) compared to larger tech companies. It promises to offer the best price-to-performance ratio in the market.

Future plans include adding multimodality and “other advanced features.”

OpenAI team member Andrej Karpathy noted that DeepSeek has demonstrated impressive research and development under limited resources.

“Does this mean you don’t need large GPU clusters for frontier LLM? No, but you should ensure you’re not wasting what you have. This looks like a good demonstration that there’s still much to be done with both data and algorithms,” he added.

Previously, DeepSeek introduced a “competitor to OpenAI’s o1″—the thinking “super-powered” AI model DeepSeek-R1-Lite-Preview.

Back in July, Chinese company Kuaishou unveiled the Kling AI model for video generation to the public.

Подписывайтесь на ForkLog в социальных сетях

Telegram (основной канал) Facebook X
Нашли ошибку в тексте? Выделите ее и нажмите CTRL+ENTER

Рассылки ForkLog: держите руку на пульсе биткоин-индустрии!

We use cookies to improve the quality of our service.

By using this website, you agree to the Privacy policy.

OK