model release

Mistral AI releases Mistral Small 4, claims improved performance on reasoning tasks

TL;DR

Mistral AI has released Mistral Small 4, the latest iteration of its small-scale language model. The company claims improvements in reasoning and coding capabilities, though specific benchmark scores and pricing details have not been publicly disclosed.

2 min read
0

Mistral AI Releases Mistral Small 4

Mistral AI announced the release of Mistral Small 4, the next generation of its small-scale language model. The update follows the company's strategy of offering efficient alternatives to larger foundation models.

What We Know

Mistral Small 4 is positioned as an improvement over its predecessor with a focus on reasoning and coding task performance. The company has not disclosed the model's parameter count, context window size, or training data cutoff date at this time.

Pricing and availability details have not been formally announced. Mistral AI typically offers API access through its Mistral platform, but specific pricing per 1M input and output tokens for Mistral Small 4 remains unconfirmed.

Claims vs. Confirmed Details

According to Mistral AI, the model delivers "enhanced reasoning capabilities." Benchmark scores supporting these claims have not been publicly released. The company has not specified which benchmarks were used or how Mistral Small 4 compares to competing small models like OpenAI's GPT-4 Mini or Claude Haiku.

The exact technical specifications—including whether the model handles multimodal inputs, maximum context window, and training completion date—await official documentation.

Market Context

Mistral Small 4's release continues a trend among AI companies toward releasing efficient, cost-effective models. Small models have become increasingly important as organizations seek to balance capability with inference costs and latency. Competitors in this category include:

  • OpenAI's GPT-4 Mini ($0.15 per 1M input tokens)
  • Anthropic's Claude Haiku ($0.80 per 1M input tokens)
  • Google's Gemini 1.5 Flash (pricing not yet public)

Mistral AI has built its brand on offering competitive pricing and open-weight alternatives through its partnerships with platforms like Together AI and Hugging Face.

What This Means

Mistral Small 4 represents another incremental update in the small model category, where competition is intensifying. Without concrete benchmark data, claims about "improved reasoning" cannot be independently verified. Organizations evaluating small models should wait for independent benchmarking or detailed technical specifications before making deployment decisions. The model's actual value proposition will depend on its context window, latency characteristics, and real-world performance across diverse tasks—none of which have been confirmed.

Related Articles

model release

Mistral Releases Medium 3.5: 128B Dense Model With 256k Context and Configurable Reasoning

Mistral AI released Mistral Medium 3.5, a 128B parameter dense model with a 256k context window that unifies instruction-following, reasoning, and coding capabilities. The model features configurable reasoning effort per request and a vision encoder trained from scratch for variable image sizes.

model release

NVIDIA Releases Nemotron 3 Nano Omni: 31B Multimodal Model With 256K Context and Reasoning Mode

NVIDIA released Nemotron 3 Nano Omni, a 31B parameter (30B active, 3B per token) multimodal model supporting video, audio, image, and text inputs. The model features a 256K token context window, reasoning mode with chain-of-thought, and tool calling capabilities.

model release

NVIDIA Releases Nemotron 3 Nano Omni: 31B-Parameter Multimodal Model with 256K Context and Reasoning Mode

NVIDIA has released Nemotron 3 Nano Omni 30B-A3B, a multimodal large language model with 31 billion parameters using a Mamba2-Transformer hybrid Mixture of Experts architecture. The model supports video, audio, image, and text inputs with a 256K token context window and includes a dedicated reasoning mode with chain-of-thought capabilities.

model release

Nvidia releases Nemotron 3 Nano Omni: 30B-parameter multimodal model with 256K context, free on OpenRouter

Nvidia has released Nemotron 3 Nano Omni, a 30-billion-parameter multimodal model available free on OpenRouter. The model features a 256,000-token context window, accepts text, image, video, and audio inputs, and claims 2× higher throughput for video reasoning compared to separate vision and speech pipelines.

Comments

Loading...