{"id":22828,"date":"2025-04-08T15:14:21","date_gmt":"2025-04-08T12:14:21","guid":{"rendered":"https:\/\/forklog.com\/en\/meta-unveils-new-series-of-ai-models-llama-4\/"},"modified":"2025-04-08T15:14:21","modified_gmt":"2025-04-08T12:14:21","slug":"meta-unveils-new-series-of-ai-models-llama-4","status":"publish","type":"post","link":"https:\/\/forklog.com\/en\/meta-unveils-new-series-of-ai-models-llama-4\/","title":{"rendered":"Meta Unveils New Series of AI Models: Llama 4"},"content":{"rendered":"<p>Meta Corporation has <a href=\"https:\/\/ai.meta.com\/blog\/llama-4-multimodal-intelligence\/\">released<\/a> a new lineup of open AI models, Llama 4. According to internal tests, these models outperform competitors across various benchmarks. <\/p>\n<p>The series is anchored by Llama 4 Behemoth, a large language model (LLM) with 2 trillion parameters. It is currently in the training phase and has not yet been released. Two of its multimodal <span data-descr=\"simplified and reduced version\" class=\"old_tooltip\">distillations<\/span>\u2014Maverick and Scout\u2014are available to developers and users.<\/p>\n<p>Meta AI assistant, available in various company products like WhatsApp, Messenger, and Instagram, has already been updated to use Llama 4 in 40 countries. However, multimodal features are currently available only in the United States. <\/p>\n<p>It is claimed that Behemoth, the LLM teacher of the other two models, surpasses GPT-4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro in STEM-oriented <span data-descr=\"S \u2013 Science, T \u2013 Technology, E \u2013 Engineering, M \u2013 Mathematics\" class=\"old_tooltip\">benchmarks<\/span> like MATH-500 and GPQA Diamond. <\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>\u201cThis is just the beginning for the Llama 4 collection. We believe the most intelligent systems should be capable of performing generalized actions, naturally communicating with people, and solving complex tasks they have not encountered before. Empowering Llama with super capabilities in these areas will lead to the creation of better products for people on our platforms and expand developers&#8217; ability to innovate in the next major consumer and business sectors,\u201d the company announcement states. <\/p>\n<\/blockquote>\n<h2 class=\"wp-block-heading\"><strong>New Architecture<\/strong><\/h2>\n<p>Llama 4 is the first series of models to use the <span data-descr=\"an approach in machine learning where the model consists of many 'experts', and only some are activated during input processing\" class=\"old_tooltip\">Mixture of Experts (MoE)<\/span> architecture. Maverick has 128 \u201cexperts\u201d and 400 billion total parameters, but only 17 billion are active. Scout&#8217;s figures are 16, 109 billion, and 17 billion, respectively. <\/p>\n<figure class=\"wp-block-image\"><img decoding=\"async\" src=\"https:\/\/lh7-qw.googleusercontent.com\/docsz\/AD_4nXcX9U-HmfBXNyQsHqzIpPavVPhocs435RGIOFmzS9J49p2cZnG-Z3vLllHW085CWDiViDlPgZOIs6oyl_4t1R4FkKo88sgNlkpkOATEFoY8ETAR6T7s7YFpaPfmLiNV4m3b12Qr2g?key=EoXF4IWBGZDcBIXtrCQWchAA\" alt=\"Meta \u043f\u0440\u0435\u0434\u0441\u0442\u0430\u0432\u0438\u043b\u0430 \u043d\u043e\u0432\u0443\u044e \u0441\u0435\u0440\u0438\u044e \u0418\u0418-\u043c\u043e\u0434\u0435\u043b\u0435\u0439 Llama 4\"\/><figcaption class=\"wp-element-caption\">Characteristics of neural networks from the Llama 4 lineup. Data: Meta.<\/figcaption><\/figure>\n<p>According to the company&#8217;s internal tests, Maverick outperforms models like GPT-4o and Gemini 2.0 in some programming, reasoning, language support, long contexts, and image tests. However, the neural network falls short of the more powerful and modern Gemini 2.5 Pro from Google, Claude 3.7 Sonnet from Anthropic, and GPT-4.5 from OpenAI.<\/p>\n<p>Maverick is better suited for use as a general assistant and chat. Scout&#8217;s strengths lie in document summarization and reasoning over large databases. The latter can operate on a single Nvidia H100 graphics processor, while Maverick requires a Nvidia H100 DGX system or its equivalent.<\/p>\n<h2 class=\"wp-block-heading\"><strong>Controversy Surrounding Llama 4<\/strong><\/h2>\n<p>Maverick secured second place in the LLM Arena\u2014a test where people compare the performance of various models and form a \u201cuser\u201d ranking. <\/p>\n<figure class=\"wp-block-image\"><img decoding=\"async\" src=\"https:\/\/lh7-qw.googleusercontent.com\/docsz\/AD_4nXdo2AOKv3zdhdWIRTe91I9e7c5D83yGxs1x7hxhH3L0FwAsVN3sZVnLtcGPFFag4Zr5q-kekxyE5F4jJUM3E8_plDZuf7thr1Te52vr7jUKNGrlFB2kFKYfK-IJZszcKxjJcGrAXQ?key=EoXF4IWBGZDcBIXtrCQWchAA\" alt=\"Meta \u043f\u0440\u0435\u0434\u0441\u0442\u0430\u0432\u0438\u043b\u0430 \u043d\u043e\u0432\u0443\u044e \u0441\u0435\u0440\u0438\u044e \u0418\u0418-\u043c\u043e\u0434\u0435\u043b\u0435\u0439 Llama 4\"\/><figcaption class=\"wp-element-caption\">AI model rankings according to LLM Arena data. Data: <a href=\"https:\/\/lmarena.ai\/?leaderboard\">LLM Arena.<\/a><\/figcaption><\/figure>\n<p>Several researchers <a href=\"https:\/\/x.com\/suchenzang\/status\/1908938638869909724\">noted<\/a> that a specially optimized version of Maverick, unavailable to developers, participated in the tests. The version for LLM Arena uses more emojis and provides unusually long responses. <\/p>\n<blockquote class=\"twitter-tweet\">\n<p lang=\"en\" dir=\"ltr\">Okay Llama 4 is def a littled cooked lol, what is this yap city <a href=\"https:\/\/t.co\/y3GvhbVz65\">pic.twitter.com\/y3GvhbVz65<\/a><\/p>\n<p>\u2014 Nathan Lambert (@natolambert) <a href=\"https:\/\/twitter.com\/natolambert\/status\/1908893136518098958?ref_src=twsrc%5Etfw\">April 6, 2025<\/a><\/p><\/blockquote>\n<p> <script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n<p>This makes it difficult for users to predict the real performance of the neural network in \u201ceveryday\u201d conditions.<\/p>\n<h2 class=\"wp-block-heading\"><strong>Denial<\/strong><\/h2>\n<p>Meta&#8217;s Vice President for Generative Artificial Intelligence, Ahmad Al-Dahle, denied the information about model tuning for specific tests. <\/p>\n<blockquote class=\"twitter-tweet\">\n<p lang=\"en\" dir=\"ltr\">We&#8217;re glad to start getting Llama 4 in all your hands. We&#8217;re already hearing lots of great results people are getting with these models. <\/p>\n<p>That said, we&#8217;re also hearing some reports of mixed quality across different services. Since we dropped the models as soon as they were\u2026<\/p>\n<p>\u2014 Ahmad Al-Dahle (@Ahmad_Al_Dahle) <a href=\"https:\/\/twitter.com\/Ahmad_Al_Dahle\/status\/1909302532306092107?ref_src=twsrc%5Etfw\">April 7, 2025<\/a><\/p><\/blockquote>\n<p> <script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>\u201cThis is simply not true, and we would never do such a thing,\u201d he emphasized. <\/p>\n<\/blockquote>\n<p>According to the executive, \u201cthe variable quality people are observing is due to the need to stabilize the implementation.\u201d<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>\u201cSince we released the models as soon as they were ready, we expect it will take a few days for all public deployments to be configured,\u201d he added.<\/p>\n<\/blockquote>\n<p>Back in November 2024, Meta <a href=\"https:\/\/forklog.com\/en\/news\/us-military-gains-access-to-metas-ai\">opened<\/a> its AI technologies to U.S. government agencies and defense contractors, as well as allies. <\/p>\n<p>Earlier, it <a href=\"https:\/\/forklog.com\/en\/news\/meta-unveils-ai-video-generator-movie-gen\">introduced<\/a> Movie Gen\u2014an AI generator for creating new videos, editing existing ones, and adding sound to them. <\/p>\n","protected":false},"excerpt":{"rendered":"<p>Meta Corporation has released a new lineup of open AI models, Llama 4. According to internal tests, these models outperform competitors across various benchmarks. The series is anchored by Llama 4 Behemoth, a large language model (LLM) with 2 trillion parameters. It is currently in the training phase and has not yet been released. Two [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":22827,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"select":"","news_style_id":"","cryptorium_level":"","_short_excerpt_text":"","creation_source":"","_metatest_mainpost_news_update":false,"footnotes":""},"categories":[3],"tags":[438,1293,1150],"class_list":["post-22828","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news-and-analysis","tag-artificial-intelligence","tag-meta","tag-news-plus"],"aioseo_notices":[],"amp_enabled":true,"views":"82","promo_type":"","layout_type":"","short_excerpt":"","is_update":"","_links":{"self":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/22828","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/comments?post=22828"}],"version-history":[{"count":0,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/posts\/22828\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/media\/22827"}],"wp:attachment":[{"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/media?parent=22828"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/categories?post=22828"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/forklog.com\/en\/wp-json\/wp\/v2\/tags?post=22828"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}