{"id":25518,"date":"2025-07-23T13:47:12","date_gmt":"2025-07-23T10:47:12","guid":{"rendered":"https:\/\/forklog.com\/en\/alibabas-new-ai-coder-surpasses-deepseek\/"},"modified":"2025-07-23T13:47:12","modified_gmt":"2025-07-23T10:47:12","slug":"alibabas-new-ai-coder-surpasses-deepseek","status":"publish","type":"post","link":"https:\/\/forklog.com\/en\/alibabas-new-ai-coder-surpasses-deepseek\/","title":{"rendered":"Alibaba&#8217;s New AI Coder Surpasses DeepSeek"},"content":{"rendered":"<p>Alibaba has launched what it calls the &#8220;most advanced AI tool for programming&#8221;\u2014an open-source AI model named Qwen3-Coder.<\/p>\n<blockquote class=\"twitter-tweet\">\n<p lang=\"en\" dir=\"ltr\">>>> Qwen3-Coder is here! \u2705<\/p>\n<p>We\u2019re releasing Qwen3-Coder-480B-A35B-Instruct, our most powerful open agentic code model to date. This 480B-parameter Mixture-of-Experts model (35B active) natively supports 256K context and scales to 1M context with extrapolation. It achieves\u2026 <a href=\"https:\/\/t.co\/Z8HfyrVScE\">pic.twitter.com\/Z8HfyrVScE<\/a><\/p>\n<p>\u2014 Qwen (@Alibaba_Qwen) <a href=\"https:\/\/twitter.com\/Alibaba_Qwen\/status\/1947766835023335516?ref_src=twsrc%5Etfw\">July 22, 2025<\/a><\/p><\/blockquote>\n<p> <script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n<p>The neural network is designed for tasks related to programming. It can write code and manage complex workflows, showing progress in other areas:<\/p>\n<ul class=\"wp-block-list\">\n<li>following instructions;<\/li>\n<li>logical reasoning;<\/li>\n<li>text comprehension;<\/li>\n<li>mathematics;<\/li>\n<li>natural sciences;<\/li>\n<li>tool usage.<\/li>\n<\/ul>\n<p><span id=\"docs-internal-guid-cf6fe6a1-7fff-6b3f-c744-3c506bc6a645\" style=\"white-space: normal;\"><span style=\"font-size: 11pt; font-family: Arial, sans-serif; font-variant-ligatures: normal; font-variant-alternates: normal; font-variant-numeric: normal; font-variant-east-asian: normal; font-variant-position: normal; vertical-align: baseline; white-space: pre-wrap;\">The company positions the product as &#8220;particularly strong&#8221; in agentic AI tasks\u2014where artificial intelligence can operate independently without human intervention.<\/span><\/span><\/p>\n<blockquote class=\"twitter-tweet\">\n<p lang=\"zxx\" dir=\"ltr\"><a href=\"https:\/\/t.co\/M8krjTM4tq\">pic.twitter.com\/M8krjTM4tq<\/a><\/p>\n<p>\u2014 Qwen (@Alibaba_Qwen) <a href=\"https:\/\/twitter.com\/Alibaba_Qwen\/status\/1947766843311354307?ref_src=twsrc%5Etfw\">July 22, 2025<\/a><\/p><\/blockquote>\n<p> <script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n<p>According to the data presented, Qwen3-Coder has outperformed Chinese competitors, including models from DeepSeek and K2 by Moonshot AI, across several metrics. In the American Mathematics Olympiad, it scored 70.3 points compared to DeepSeek-V3-0324&#8217;s 46.6.<\/p>\n<p>In the MultiPL-E programming test, the language model scored 87.9 points, while DeepSeek scored 82.2.<\/p>\n<p>In some areas, the neural network is comparable to leading American counterparts such as Claude from Anthropic and GPT-4 from OpenAI.<\/p>\n<p>The Qwen3-235B-A22B-Instruct-2507-FP8 neural network is an update to the previous version Qwen3-235B-A22B-FP8, but operates in a non-reasoning mode\u2014producing results without step-by-step logical deductions. This allows the maximum length of processed content to increase eightfold to 256,000 tokens, enabling work with much larger texts within a single dialogue.<\/p>\n<p>Back in May, Alibaba <a href=\"https:\/\/forklog.com\/en\/news\/alibaba-reduces-ai-search-training-costs-by-88\">announced a breakthrough<\/a> in reducing the cost of training AI models focused on search.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Alibaba has launched what it calls the &#8220;most advanced AI tool for programming&#8221;\u2014an open-source AI model named Qwen3-Coder. >>> Qwen3-Coder is here! \u2705 We\u2019re releasing Qwen3-Coder-480B-A35B-Instruct, our most powerful open agentic code model to date. This 480B-parameter Mixture-of-Experts model (35B active) natively supports 256K context and scales to 1M context with extrapolation. It achieves\u2026 pic.twitter.com\/Z8HfyrVScE [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"select":"","news_style_id":"","cryptorium_level":"","_short_excerpt_text":"","creation_source":"","_metatest_mainpost_news_update":false,"footnotes":""},"categories":[3],"tags":[640,438],"class_list":["post-25518","post","type-post","status-publish","format-standard","hentry","category-news-and-analysis","tag-alibaba","tag-artificial-intelligence"],"aioseo_notices":[],"amp_enabled":true,"views":"100","promo_type":"","layout_type":"","short_excerpt":"","is_update":"","_links":{"self":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/25518","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/comments?post=25518"}],"version-history":[{"count":0,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/25518\/revisions"}],"wp:attachment":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/media?parent=25518"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/categories?post=25518"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/tags?post=25518"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}