Reasoning Effort Control And Operational Gap
Sources: 1 • Confidence: High • Updated: 2026-04-13 03:50
Key takeaways
- The author reports they could not find documentation for setting reasoning effort in the Mistral API.
- Mistral Small 4 is described as a 119B-parameter Mixture-of-Experts model with 6B active parameters.
- The Mistral Small 4 model weights are reported as 242GB on Hugging Face.
- The model was tested via the Mistral API using the llm-mistral plugin with the model identifier "mistral/mistral-small-2603".
- Mistral announced Leanstral, an open-weight model tuned to produce Lean 4 formally verifiable code.
Sections
Reasoning Effort Control And Operational Gap
- The author reports they could not find documentation for setting reasoning effort in the Mistral API.
- The author expects the ability to set reasoning effort in the Mistral API may be added soon.
- Mistral Small 4 supports a setting called reasoning_effort with values "none" or "high".
- Mistral claims reasoning_effort="high" provides verbosity equivalent to previous Magistral models.
Model Architecture And Product Consolidation
- Mistral Small 4 is described as a 119B-parameter Mixture-of-Experts model with 6B active parameters.
- Mistral says Mistral Small 4 unifies reasoning, multimodal, and agentic coding capabilities previously associated with Magistral, Pixtral, and Devstral into one model.
Deployment Artifacts And Self Hosting Feasibility
- The Mistral Small 4 model weights are reported as 242GB on Hugging Face.
Tooling Reproducibility And Access Path
- The model was tested via the Mistral API using the llm-mistral plugin with the model identifier "mistral/mistral-small-2603".
Formal Verification Specialization
- Mistral announced Leanstral, an open-weight model tuned to produce Lean 4 formally verifiable code.
Watchlist
- The author reports they could not find documentation for setting reasoning effort in the Mistral API.
- The author expects the ability to set reasoning effort in the Mistral API may be added soon.
Unknowns
- What is Mistral Small 4's measured performance across reasoning, multimodal, and agentic coding tasks relative to the referenced prior model lines?
- Is reasoning_effort actually supported as an API parameter for Mistral Small 4, and if so what are the exact request/response semantics (including latency and token usage impacts)?
- What serving configuration is required to run the 119B MoE model efficiently (hardware requirements, sharding approach, and recommended inference stack)?
- Are smaller-footprint distributions (such as alternative shards or quantized weights) available for Mistral Small 4, and what quality tradeoffs do they entail?
- What is the licensing and usage constraint profile for Mistral Small 4 and Leanstral as released, and does it differ between API use and open-weight use?