Illustration for: Mistral launches Large 3, an Apache‑2.0 open‑source model for language, images
Open Source

Mistral launches Large 3, an Apache‑2.0 open‑source model for language, images

2 min read

Why does a new model matter when the field is already crowded? Paris‑based Mistral just announced Large 3, positioning it as a “major new open‑source AI model” in a market where licensing terms often dictate adoption. The company chose the Apache‑2.0 license, a move that signals a clear intent to keep the code freely reusable and commercially viable.

While many open models excel at text, Mistral claims this version also processes images, aiming to sit alongside the leading open‑source contenders on general language benchmarks. Early results on the LMArena leaderboard show it near the top of its class—second among open‑source non‑reasoning models and sixth overall—suggesting the effort is more than a publicity stunt. If those numbers hold, developers could finally have a single, permissively licensed system that handles both language and vision without sacrificing performance.

The following statement from Mistral sums up the ambition behind the release.

Mistral Large 3 is fully open source under the Apache-2.0 license. The company says it aims to match other leading open models on general language tasks while also handling images. On the LMArena leaderboard, it currently ranks second among open-source non-reasoning models and sixth among open-source reasoning models.

In published benchmarks, its performance lines up with other open models like Qwen and Deepseek. Still, Deepseek released V3.2 yesterday, and that update shows clear improvements over the previous version in several tests. What the new edge models mean for efficiency The smaller "Ministral 3" variants target local and edge use.

Related Topics: #Mistral #Large 3 #Apache-2.0 #open-source #AI #LMArena #Deepseek #Qwen #edge

Will the new Mistral Large 3 live up to its promise? The model is fully open‑source under Apache‑2.0, a clear shift toward broader accessibility. With 41 billion active parameters and a sparse Mixture‑of‑Experts design, it sits between the smaller “Ministral” variants—3, 8 and 14 billion parameters—and the larger family’s edge‑focused offerings.

Trained on roughly 3,000 Nvidia H200 GPUs, the effort reflects a substantial compute investment. Mistral claims the flagship can match other leading open models on general language tasks while also processing images, a capability that sets it apart from many contemporaries. On the LMArena leaderboard it currently ranks second among open‑source non‑reasoning models and sixth overall in the open‑source category, suggesting competitive performance but not outright dominance.

It remains unclear whether the model’s multimodal handling will translate into consistent results across diverse benchmarks. The open‑source licensing may encourage community scrutiny, yet real‑world adoption will ultimately test the balance between its architectural complexity and practical utility.

Further Reading

Common Questions Answered

What licensing model does Mistral Large 3 use and why is it significant?

Mistral Large 3 is released under the Apache‑2.0 license, which allows anyone to freely reuse, modify, and commercialize the code. This choice signals Mistral's intent to promote broader accessibility and avoid restrictive licensing that can limit adoption.

How does Mistral Large 3 perform on the LMArena leaderboard compared to other open‑source models?

On the LMArena leaderboard, Mistral Large 3 ranks second among open‑source non‑reasoning models and sixth among open‑source reasoning models. Its benchmark scores are comparable to other leading open models such as Qwen and Deepseek.

What are the key architectural features of Mistral Large 3, including its parameter count and design?

Mistral Large 3 contains 41 billion active parameters and uses a sparse Mixture‑of‑Experts architecture, placing it between the smaller 3, 8, and 14 billion‑parameter variants and the larger edge‑focused models. This design aims to balance performance and efficiency while supporting both language and image tasks.

What hardware was used to train Mistral Large 3 and what does this indicate about the compute investment?

The model was trained on roughly 3,000 Nvidia H200 GPUs, reflecting a substantial compute investment required for a model of its size and capabilities. This large‑scale training effort underscores Mistral's commitment to delivering a competitive open‑source AI system.