Privacy Policy Banner

We use cookies to improve your experience. By continuing, you agree to our Privacy Policy.

Microsoft Phi-4-Reasonning: the SLM which reason

When 14 billion parameters surpass five times larger models … The evolution of SLMs crosses a strategic course with the introduction of “Phi-4” compact models capable of advanced reasoning, low latency, reflection and performance comparable to massive LLM.

Microsoft may be a very privileged partner of Openai and with most of the players to integrate their AA models in Azure Ai Foundry, the publisher does not refrain from continuing his own technological tracks, by working on innovations at the heart of neural networks (like his amazing Bitnet B1.58 model based on trit), on its own SLM in source secrets secret (May-1 project).

A year after inaugurating its range of small AI models (SLM)Phi‑3 and two months after inaugurating generation 4 with a multimodal SLM (Phi-4-Multimodal) and a tiny model ( Ph-4-min), Microsoft announces three variants of its latest generation SLM: Phi‑4‑reasoning, Phi‑4‑reasoning‑plus etPhi‑4‑mini‑reasoning.

Published on April 30, 2025, these “integrated reasoning” versions widen the open -weight offer of compact models intended for developers who must maintain a low latency while demanding a complex reasoning.

At the heart of Microsoft’s engineers’ approach to make your SLMs “reasoning”: based on supervision (SFT) from the reasoning channels from Openai O3 – Mini, and exploit a strengthening by strengthening (RL) for “more” declination. “”Thanks to distillation, strengthening learning and high quality data, these models reconcile size and performance“Explains Microsoft.

Small but gifted

And the results to the various main benchmarks on the have something to make the competition pale: typically with only 14 billion parameters, Phi‑4‑reasoning Devance Deepseek – R1 -Distill -Llama – 70b (70 billion parameters) on the series likes 2025, MMLU -PRO or HUMANEVAL – Plus, and is approaching the complete Deepseek – R1 model (671 billion parameters)! The variant Phi-4-reasoning-plusaligned on the same 14 MD of parameters but trained with 1.5 times more tokens, borders on the scores of Openai O3 – Mini on Omnimath! For information, Phi‑4‑reasoning benefits from a classic 128,000 tokens contextual window which was extended to 256,000 tokens for the version Phi-4-reasoning-plus.

Designed for embedded, Phi‑4‑mini‑reasoning Displays 3.8 billion parameters, a synthetic game of one million mathematical generated by Deepseek -R1, and achieves the performance of O1 -MINI on Math – 500 while exceeding several models of 7 to 8 billion parameters.
With its most tiny size, this model are ideal for local execution, including on mobile terminals and to meet the needs of almost immediate response. It is particularly suitable for educational uses and local chatbots.

Models open to various uses

On the deployment side, the CIOs will find these models already optimized for Copilot+ PCs: the NPU “Phi Silica” variant is preloaded in memory and provides almost instantaneous periods, guaranteeing energy -efficient cohabitation with business applications. Windows APIs allow you to integrate the offline generation in Outlook or in internal tools.

On the safe level, Microsoft claims a pipeline aligned with its principles of responsibility – responsibility, equity, reliability, security and inclusion. The models undergo a post -training combining SFT, Direct Preference Optimization and RLHF from public and internal games oriented “Helpfulness/Harmlessness”. Microsoft also publishes the “cards” of its models which detail the residual limits and attenuation measures.

Available now on Azure Ai Foundry, Hugging Face and Github Models, the three models are published under the very permissive MIT license, opening the way to local inference as with hybrid cloud deployments. For security and architecture teams, this new generation of SLM offers a credible alternative to massive LLMs, with reduced TCO, local execution as in EDGE and increased data control. These models have been proof of the incredible made by SLM for a year and their astonishing potential in a universe in search of AI less expensive and more energy - in energy and resources.

The whole of “Phi” models of Microsoft in full…

Read also:

IA based on reasoning models
Data / it

Reasoning models: the new AI border

Read also:

The company in the era of agentic AI, seen by the Microsoft Work Trend Index 2025
Data / it

The company in the of agentic AI, seen by Microsoft

Read also:

Hallucine reasoning AIs more than classic AI
Data / it

In fact, reasoning AIs can hallucinate more than conventional AI

We send you a validation email!

-

PREV More gas and less environment: but what does the Caisse des Dépôts do?
NEXT Is security really at the heart of its blurring with the royal family?