>
Mistral Medium 3 introduces a new class of models that balances SOTA performance 8X lower cost simpler deployability to accelerate enterprise usage The model leads in professional use cases such as coding and multimodal understanding The model delivers a range of enterprise capabilities including: Hybrid or on-premises / in-VPC deployment Custom post-training Integration into enterprise tools and systems
Notes: Benchmarks match with models like GPT-4o, Mistral's previous largest runs were ~1e25 FLOP scale, so plausibly they might have trained Medium with this much compute.