Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

EU Commission: “AI Gigafactories” to strengthen Europe as a business location

United States, China, and United Kingdom Lead the Global AI Ranking According to Stanford HAI’s Global AI Vibrancy Tool

Foundation AI: Cisco launches AI model for integration in security applications

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Facebook X (Twitter) Instagram
Advanced AI News
Home » Mistral just updated its open source Small model from 3.1 to 3.2: here’s why
Mistral AI

Mistral just updated its open source Small model from 3.1 to 3.2: here’s why

Advanced AI EditorBy Advanced AI EditorJune 20, 2025No Comments5 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more

French AI darling Mistral is keeping the new releases coming this summer.

Just days after announcing its own domestic AI-optimized cloud service Mistral Compute, the well-funded company has released an update to its 24B parameter open source model Mistral Small, jumping from a 3.1 release to 3.2-24B Instruct-2506.

The new version builds directly on Mistral Small 3.1, aiming to improve specific behaviors such as instruction following, output stability, and function calling robustness. While overall architectural details remain unchanged, the update introduces targeted refinements that affect both internal evaluations and public benchmarks.

According to Mistral AI, Small 3.2 is better at adhering to precise instructions and reduces the likelihood of infinite or repetitive generations — a problem occasionally seen in prior versions when handling long or ambiguous prompts.

Similarly, the function calling template has been upgraded to support more reliable tool-use scenarios, particularly in frameworks like vLLM.

And at the same time, it could run on a setup with a single Nvidia A100/H100 80GB GPU, drastically opening up the options for businesses with tight compute resources and/or budgets.

An updated model after only 3 months

Mistral Small 3.1 was announced in March 2025 as a flagship open release in the 24B parameter range. It offered full multimodal capabilities, multilingual understanding, and long-context processing of up to 128K tokens.

The model was explicitly positioned against proprietary peers like GPT-4o Mini, Claude 3.5 Haiku, and Gemma 3-it — and, according to Mistral, outperformed them across many tasks.

Small 3.1 also emphasized efficient deployment, with claims of running inference at 150 tokens per second and support for on-device use with 32 GB RAM.

That release came with both base and instruct checkpoints, offering flexibility for fine-tuning across domains such as legal, medical, and technical fields.

In contrast, Small 3.2 focuses on surgical improvements to behavior and reliability. It does not aim to introduce new capabilities or architecture changes. Instead, it acts as a maintenance release: cleaning up edge cases in output generation, tightening instruction compliance, and refining system prompt interactions.

Small 3.2 vs. Small 3.1: what changed?

Instruction-following benchmarks show a small but measurable improvement. Mistral’s internal accuracy rose from 82.75% in Small 3.1 to 84.78% in Small 3.2.

Similarly, performance on external datasets like Wildbench v2 and Arena Hard v2 improved significantly—Wildbench increased by nearly 10 percentage points, while Arena Hard more than doubled, jumping from 19.56% to 43.10%.

Internal metrics also suggest reduced output repetition. The rate of infinite generations dropped from 2.11% in Small 3.1 to 1.29% in Small 3.2 — almost a 2× reduction. This makes the model more reliable for developers building applications that require consistent, bounded responses.

Performance across text and coding benchmarks presents a more nuanced picture. Small 3.2 showed gains on HumanEval Plus (88.99% to 92.90%), MBPP Pass@5 (74.63% to 78.33%), and SimpleQA. It also modestly improved MMLU Pro and MATH results.

Vision benchmarks remain mostly consistent, with slight fluctuations. ChartQA and DocVQA saw marginal gains, while AI2D and Mathvista dropped by less than two percentage points. Average vision performance decreased slightly from 81.39% in Small 3.1 to 81.00% in Small 3.2.

This aligns with Mistral’s stated intent: Small 3.2 is not a model overhaul, but a refinement. As such, most benchmarks are within expected variance, and some regressions appear to be trade-offs for targeted improvements elsewhere.

However, as AI power user and influencer @chatgpt21 posted on X: “It got worse on MMLU,” meaning the Massive Multitask Language Understanding benchmark, a multidisciplinary test with 57 questions designed to assess broad LLM performance across domains. Indeed, Small 3.2 scored 80.50%, slightly below Small 3.1’s 80.62%.

Open source license will make it more appealing to cost-conscious and customized-focused users

Both Small 3.1 and 3.2 are available under the Apache 2.0 license and can be accessed via the popular. AI code sharing repository Hugging Face (itself a startup based in France and NYC).

Small 3.2 is supported by frameworks like vLLM and Transformers and requires roughly 55 GB of GPU RAM to run in bf16 or fp16 precision.

For developers seeking to build or serve applications, system prompts and inference examples are provided in the model repository.

While Mistral Small 3.1 is already integrated into platforms like Google Cloud Vertex AI and is scheduled for deployment on NVIDIA NIM and Microsoft Azure, Small 3.2 currently appears limited to self-serve access via Hugging Face and direct deployment.

What enterprises should know when considering Mistral Small 3.2 for their use cases

Mistral Small 3.2 may not shift competitive positioning in the open-weight model space, but it represents Mistral AI’s commitment to iterative model refinement.

With noticeable improvements in reliability and task handling — particularly around instruction precision and tool usage — Small 3.2 offers a cleaner user experience for developers and enterprises building on the Mistral ecosystem.

The fact that it is made by a French startup and compliant with EU rules and regulations such as GDPR and the EU AI Act also make it appealing for enterprises working in that part of the world.

Still, for those seeking the biggest jumps in benchmark performance, Small 3.1 remains a reference point—especially given that in some cases, such as MMLU, Small 3.2 does not outperform its predecessor. That makes the update more of a stability-focused option than a pure upgrade, depending on the use case.

Daily insights on business use cases with VB Daily

If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.

Read our Privacy Policy

Thanks for subscribing. Check out more VB newsletters here.

An error occured.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleAccess to AI Training Data Sparks Legal Questions
Next Article Perplexity Fires Back at BBC in AI Scraping Row
Advanced AI Editor
  • Website

Related Posts

Mistral just updated its open source Small model from 3.1 to 3.2: here’s why

June 21, 2025

Mistral just updated its open source Small model from 3.1 to 3.2: here’s why

June 21, 2025

Mistral just updated its open source Small model from 3.1 to 3.2: here’s why

June 21, 2025
Leave A Reply Cancel Reply

Latest Posts

Songtsam Resorts Launch Collaboration Inspired By Tibet’s Sacred Lake

Spanish Supreme Court Orders Heirs to Return Cathedral Statues

ARTnews Polled 10 Digital Art Experts To Find Out Their Favorite Digital Art Works

How Singapore Reimagines Care Through Design

Latest Posts

EU Commission: “AI Gigafactories” to strengthen Europe as a business location

June 21, 2025

United States, China, and United Kingdom Lead the Global AI Ranking According to Stanford HAI’s Global AI Vibrancy Tool

June 21, 2025

Foundation AI: Cisco launches AI model for integration in security applications

June 21, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.