The entire AI landscape modified in January 2025, after a Chinese Ki Startup Deepteteek, who was less well-known on the time (a subsidiary of the Hong Kong, quantitative evaluation company High-Flyer Capital Management), began his powerful open source language model Deepseek R1 in public and defeated Giants like Meta.
When Deepseek use spread quickly with researchers and firms, Meta was supposedly sent to panic mode When he learned that this latest R1 model had been trained for a fraction of the prices of many other leading models and that it had exceeded for less than several million dollars -which it pays out a few of his own AI team leaders.
Up thus far, the complete generative AI strategy of Meta had the publication of first-class open source models under its brand name “Llama” for researchers and firms on which they were capable of construct up freely (a minimum of in the event that they had fewer than 700 million monthly users and at the moment should contact META for special paid license conditions).
But Deepseek R1s surprisingly good performance in a much smaller budget had supposedly shaked the corporate management and compelled a form of billing, with the last version of Lama. 3.3After he was only published a month before December 2024 and was still old-fashioned.
Now we all know the fruits of this billing: today, Mark Zuckerberg, founder and CEO of Meta, went to his Instagram account to be announced a New Lama 4 series from ModelsWith two of them-of the 400 billion parameter Llama 4 Maverick and 109 billion parameters Llama 4 Scout-Sind available for developers today llama.com and AI code -Sharing community Hug.
An enormous 2-brillion parameter Llama 4 giant can be presented today, today, Although Metas blog post on the publications said it was still trained and gave no indication of when it could possibly be released. (Recall parameters confer with the settings that determine the behavior of the model and customarily mean a more powerful and more complex thing in regards to the model.)
An overrcription of those models is that they’re all multimodal – trained and are subsequently capable of receive and generate text, video and pictures (Hough -Audio was not mentioned).
Another is that you’ve an incredibly long context window -1 million tokens for Lama 4 Maverick and 10 million for Lama 4 Scout -which corresponds to about 1,500 or 15,000 pages of text, which the model can process with a single input/output interaction. This signifies that a user could theoretically upload or insert as much as 7,500 page-word-text and will receive a lot in return from Llama 4 Scout, which can be practical for information density comparable to medicine, science, engineering, mathematics, literature, etc.
Here is what we’ve got learned about this publication:
All-in on expert mixtures
All three models use the architectural approach “Mixing mix (MOE)” made popular in earlier model publications by Openai And Mistral, which essentially focuses on several smaller models (“experts”) in various tasks, subjects and media formats in a uniform, whole, larger model. Each LAMA 4 publication should subsequently be a combination of 128 different experts and to operate more efficiently, since only the expert is crucial for a selected task and a “common” expert for each token, as a substitute of that the complete model must be directed for each running.
As the Lama 4 blog post notes:
Both Scout and Maverick can be found to the general public for self-hosting, while no hosted API or price levels have been announced for the official META infrastructure. Instead, Meta focuses on distribution through open download and integration with Meta AI in WhatsApp, Messenger, Instagram and web.
Meta estimates the inference costs for LAMA 4 -Maverick to 0.19 to 0.49 USD per 1 million tokens (using a 3: 1 mixture of input and output). This makes it less expensive than proprietary models comparable to GPT-4O, which is estimated on tokens of 4.38 per million per million, based on community benchmarks.
All three LAMA 4 models particularly Maverick and Behemoth-Sind expressly for the argument, coding and step-by-step problem solution–they don’t seem to indicate the chains for dedicated argumentation models comparable to the Openai-O series or Deepseek R1.
Instead, they appear to be designed so that they’re more direct with “classic”, not obsessed LLMs and multimodal models comparable to Openais GPT-4O and Deepseek V3-with the exception of LAMA 4-gene that seems to threaten Deepseek R1 (more on this below!)
In addition, for LAMA 4 tailor -made pipelines built to enhance the argument after training, comparable to: B.:
- Removing over 50% of the “easy” requests in the course of the supervised tremendous -tuning.
- Introduction of a continuous reinforcement loop with increasingly harder requests.
- Use PASS@K assessment and scanning curriculum to strengthen performance in mathematics, logic and coding.
- Implementation of METAP, a brand new technology with which engineers use hyperparameters (comparable to the training rates per layer) to models and apply them to other model sizes and kinds and at the identical time maintain the intended model behavior.
Metap is of particular interest because it may be used to set hyperparameters on the model after which get many other kinds of models out, which increases the training efficiency.
Like my enterprise peat college and LLM expert Ben Dickson, the brand new Metap technique said: “This can save quite a lot of money and time. This signifies that you perform experiments on the smaller models as a substitute of doing them on the large-scale activities.”
This is especially vital when training models are used as large because the Behemoth, the 32K GPUS and FP8 precision and achieves 390 Tflops/GPU over greater than 30 trillion tokens -more than twice the LAMA -3 training data.
In other words, the researchers can largely recognize the model of the way to act, and apply this to a bigger and smaller version of the model and across different media forms.
A mighty – but not yet powerful – model family
In his Announcement video on Instagram (Of course a Meta subsidiary), Mark Zuckerberg, CEO of Meta), said that the corporate is “the corporate's goal in constructing the world's leading AI, Open Source IT and making it generally accessible so that everybody on the earth advantages.
It is a clearly fastidiously formulated statement, as is the blog post by Meta, which calls Llama 4 Scout, “the perfect multimodal model of the world and more powerful than all LAMA models of the previous generation” (highlighted by me).
In other words, these are very powerful models, near the tip of the bunch in comparison with others of their parameter size, but not necessarily latest performance records. Nevertheless, Meta desired to divided the models, including the brand new Lama 4 family -beats:
Lama 4 court
- Surpasses GPT-4.5, Gemini 2.0 Pro and Claude Sonnet 3.7 on:
- Math-500 (95.0)
- GPQA Diamond (73.7)
- MMLU for (82.2)
Call 4 Maverick
- Beats Gpt-4o and Gemini 2.0 Flash on most multimodal argumentation benchmarks:
- Chartqa, Docvqa, Mathvista, Mmmu
- Competitive with Deepseek V3.1 (45.8b params), while lower than half of the lively parameters (17b) are used
- Benchmark results:
- Chartqa: 90.0 (against GPT-4O 85.7)
- Docvqa: 94.4 (vs. 92.8)
- MMLU for: 80.5
- Cost -effective: 0.19 to 0.49 USD per 1 million tokens

Call 4 Scout
- Matches or exceeded models comparable to Mistral 3.1, Gemini 2.0 Flash-Lite and Gemma 3:
- DocVt: 94.4
- MMLU for: 74.3
- Mathvista: 70.7
- Unsurpassed 10 m token context length ideal for long documents, code bases or multiturn evaluation
- Developed for efficient provision on a single H100 GPU

But how does Lama 4 stack in spite of everything of this on Deepseek?
But in fact there may be a very different class of argumentative models comparable to Deepseek R1, Openais “O” series (comparable to GPT -4O), Gemini 2.0 and Claude Sonnet.
Use the model with the very best parameter-benchmarked-lama 4 giant and compare it with the Intial Deepseek R1 release table for R1-32B and Openai-o1 models.
Benchmark | Lama 4 court | Deepseek R1 | Openaai O1-1217 |
---|---|---|---|
Math-500 | 95.0 | 97.3 | 96.4 |
GPQA Diamond | 73.7 | 71.5 | 75.7 |
MMLU | 82.2 | 90.8 | 91.8 |
What can we complete?
- Math-500: Lama 4 giant is barely Deepseek R1 and Openai O1.
- GPQA Diamond: Behemoth is 1, but behind Openai O1.
- MMLU: Behemoth walks each, but still exceeds Gemini 2.0 Pro and GPT-4.5.
TakeAway: While Deepseek R1 and Openai O1, just a few metrics are still competitive and occurs in or near the highest of the argument in its class.
Security and fewer political bias “
Meta also emphasized the model orientation and security through the introduction of tools comparable to Lama Guard, promptly Guard and Cyberseeceval to assist developers to acknowledge uncertain input/outputs or continuous input requests and implement generative offensive agent tests (Goat) for automated Red teaming.
The company also claims that Lama 4 shows a major improvement in “political bias” and “especially, (leading LLMS) previously have leaned left previously in the case of political and social issues discussed.” Zuckerberg's hug of the Republican US President Donald J. Trump and his party after the 2024 elections.
Where Lama 4 is to this point
The Llama 4 models from Meta give efficiency, openness and high-end performance for multimodal and argumentative tasks.
With Scout and Maverick, which is now publicly available and giants are presented as a state-of-the-art teacher model, the Lama ecosystem is positioned to supply a competitive open alternative to first-class proprietary models from Openaai, Anthropic, Deepseek and Google.
Regardless of whether you construct assistants in the world of ​​company standards, AI research pipelines or Long Context Analytical Tools, Llama 4 offers flexible high-performance options with a transparent deal with the unique design.