{"id":14209,"date":"2024-06-10T14:10:00","date_gmt":"2024-06-10T11:10:00","guid":{"rendered":"https:\/\/forklog.com\/en\/alibaba-unveils-new-ai-model-qwen2\/"},"modified":"2024-06-10T14:10:00","modified_gmt":"2024-06-10T11:10:00","slug":"alibaba-unveils-new-ai-model-qwen2","status":"publish","type":"post","link":"https:\/\/forklog.com\/en\/alibaba-unveils-new-ai-model-qwen2\/","title":{"rendered":"Alibaba Unveils New AI Model Qwen2"},"content":{"rendered":"<p>Chinese tech giant Alibaba <a href=\"https:\/\/qwenlm.github.io\/blog\/qwen2\/\">announced<\/a> the release of its new artificial intelligence model, Qwen2.\u00a0<\/p>\n<p>Developed by Alibaba Cloud, it is the next generation of Tongyi Qianwen (Qwen). It includes Tongyi Qianwen LLM (or Qwen), Qwen-VL, and Qwen-Audio.<\/p>\n<p>The Qwen2 family comprises a series of five models ranging from 0.5 to 72 billion parameters, trained using data from various industries in 27 languages.<\/p>\n<figure class=\"wp-block-image\"><img decoding=\"async\" src=\"https:\/\/lh7-us.googleusercontent.com\/docsz\/AD_4nXe9cxx1xoIQP46eI9iKgSU307DLnbw6wYBroqzI0MLb0mtJogIAbbg0moPmzSMZd3NqWz-dP9vA2Y35KCmUBVjcXezzMSU33erx-jySNt4Kdn8BPruv2SLJbNtLX4pw5g0Uc6YWt1B6THK_-1HO9ip78ijw?key=Cs14DQtw0bNsqUczszrHFQ\" alt=\"Alibaba \u0432\u044b\u043f\u0443\u0441\u0442\u0438\u043b\u0430 \u043d\u043e\u0432\u0443\u044e \u0418\u0418-\u043c\u043e\u0434\u0435\u043b\u044c Qwen2\"\/><figcaption class=\"wp-element-caption\">Comparison of AI models in the Qwen2 series by Alibaba. Data: Qwen website.<\/figcaption><\/figure>\n<p>Queen2-72B is the most powerful model in the series, trained on 3 trillion <span data-descr=\"a fragment of text, such as a word or part of a word, used as the basic unit of data for processing in the model\" class=\"old_tooltip\">tokens<\/span>. For comparison, Meta&#8217;s Llama-2 is trained on 2 trillion tokens, while Llama-3 is trained on 15 trillion tokens.<\/p>\n<p>Qwen2 can handle long conversational contexts\u2014up to 128,000 tokens, comparable to OpenAI&#8217;s GPT-4o. The team claims their model surpasses Meta&#8217;s LLama3 in nearly all major synthetic tests.<\/p>\n<figure class=\"wp-block-image\"><img decoding=\"async\" src=\"https:\/\/lh7-us.googleusercontent.com\/docsz\/AD_4nXcdF7hsGXI-WA2JV8BAPQ2iSIt4ag0cUZlT85BrcOOHmjepItkJiyDJHckFAiqMjoNordGl2oodctHBzs2Pue_aA-qYuGlDoxLGhG4I-pW7D_yrD93zYUCg3FhxoBaLLx7FXChOQHxv7tVAsGnLgQt-cmv3?key=Cs14DQtw0bNsqUczszrHFQ\" alt=\"Alibaba \u0432\u044b\u043f\u0443\u0441\u0442\u0438\u043b\u0430 \u043d\u043e\u0432\u0443\u044e \u0418\u0418-\u043c\u043e\u0434\u0435\u043b\u044c Qwen2\"\/><figcaption class=\"wp-element-caption\">Comparison of Queen2-72B with competitors. Data: Qwen website.<\/figcaption><\/figure>\n<p>Independent platform Elo Arena <a href=\"https:\/\/huggingface.co\/spaces\/lmsys\/chatbot-arena-leaderboard\">rates<\/a> Qwen2-72B-Instruct slightly better than GPT-4-0314, but lower than Llama3 70B and GPT-4-0125-preview.\u00a0<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p><em>\u201cCompared to modern open-source language models, including the previously released Qwen1.5, Qwen2 has outperformed most models and demonstrated competitiveness in a range of tests targeting language understanding, language generation, multilingualism, programming, mathematics, and reasoning\u201d<\/em>, stated the Qwen team.<\/p>\n<\/blockquote>\n<p>The Qwen2 models exhibit strong comprehension of long contexts. Qwen2-72B-Instruct can flawlessly perform information retrieval tasks anywhere and nearly aced the \u201c<span data-descr=\"The essence of the task is to place a random fact or statement in the middle of a long context window and ask the model to find this information.\" class=\"old_tooltip\">Needle in a Haystack<\/span>\u201d test. Often, the performance of different models begins to degrade with continued interaction.<\/p>\n<figure class=\"wp-block-image\"><img decoding=\"async\" src=\"https:\/\/lh7-us.googleusercontent.com\/docsz\/AD_4nXeGxn4xCK0aTyYgaEAaAHUJ3XQizG3rstZdqOXc_EFhtgPaKvWl-v6S_8xFONHwIazDwvf59X7R0rzAe1mqQ5IABcY3M1TdF9HaTSsqmoy6D74ntSvAi6ryjXn3ewdtGTpMS6hiI_zPyK4HOZtM6TJ-K9o5?key=Cs14DQtw0bNsqUczszrHFQ\" alt=\"Alibaba \u0432\u044b\u043f\u0443\u0441\u0442\u0438\u043b\u0430 \u043d\u043e\u0432\u0443\u044e \u0418\u0418-\u043c\u043e\u0434\u0435\u043b\u044c Qwen2\"\/><figcaption class=\"wp-element-caption\">\u201cNeedle in a Haystack\u201d test. Data: Qwen website.<\/figcaption><\/figure>\n<p>Earlier, Alibaba announced the release of the AI chatbot Tongyi Qianwen.\u00a0<\/p>\n<p>Back in April, Meta <a href=\"https:\/\/forklog.com\/en\/news\/meta-integrates-ai-assistant-across-its-platforms\">announced the launch of a free AI assistant<\/a> Meta AI on platforms WhatsApp, Instagram, Facebook, and Messenger. It is based on the Llama 3 language model.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Chinese tech giant Alibaba announced the release of its new artificial intelligence model, Qwen2.\u00a0 Developed by Alibaba Cloud, it is the next generation of Tongyi Qianwen (Qwen). It includes Tongyi Qianwen LLM (or Qwen), Qwen-VL, and Qwen-Audio. The Qwen2 family comprises a series of five models ranging from 0.5 to 72 billion parameters, trained using [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":14208,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"select":"","news_style_id":"","cryptorium_level":"","_short_excerpt_text":"","creation_source":"","_metatest_mainpost_news_update":false,"footnotes":""},"categories":[3],"tags":[640,438],"class_list":["post-14209","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news-and-analysis","tag-alibaba","tag-artificial-intelligence"],"aioseo_notices":[],"amp_enabled":true,"views":"21","promo_type":"","layout_type":"","short_excerpt":"","is_update":"","_links":{"self":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/14209","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/comments?post=14209"}],"version-history":[{"count":0,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/14209\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/media\/14208"}],"wp:attachment":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/media?parent=14209"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/categories?post=14209"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/tags?post=14209"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}