Mistral Small 3.1 AI Model Released
In a newsroom post, the AI agency detailed the brand new fashions. The Mistral Small 3.1 comes with an expanded context window of as much as 1,28,000 tokens and is alleged to ship inference speeds of 150 tokens per second. This basically means the response time of the AI mannequin is sort of quick. It arrives in two variants of chat and instruct. The former works as a typical chatbot whereas the latter is fine-tuned to comply with person directions and is beneficial when constructing an software with a selected objective.
Mistral Small 3.1 benchmark
Photo Credit: Mistral
Similar to its earlier releases, the Mistral Small 3.1 is obtainable within the public area. The open weights will be downloaded from the agency’s Hugging Face listing. The AI mannequin comes with an Apache 2.0 licence which permits tutorial and analysis utilization however forbids business use circumstances.
Mistral mentioned that the massive language mannequin (LLM) is optimised to run on a single Nvidia RTX 4090 GPU or a Mac gadget with 32GB RAM. This means fans with out an costly setup to run AI fashions may obtain and entry it. The mannequin additionally provides low-latency perform calling and performance execution which will be helpful for constructing automation and agentic workflows. The firm additionally permits builders to fine-tune the Mistral Small 3.1 to suit the use circumstances of specialized domains.
Coming to efficiency, the AI agency shared varied benchmark scores primarily based on inside testing. The Mistral Small 3.1 is alleged to outperform Gemma 3 and GPT-4o mini on the Graduate-Level Google-Proof Q&A (GPQA) Main and Diamond, HumanEval, MathVista, and the DocVQA benchmarks. However, GPT-4o mini carried out higher on the Massive Multitask Language Understanding (MMLU) benchmark, and Gemma 3 outperformed it on the MATH benchmark.
Apart from Hugging Face, the brand new mannequin can also be accessible by way of the appliance programming interface (API) on Mistral AI’s developer playground La Plateforme, in addition to on Google Cloud’s Vertex AI. It may also be made accessible on Nvidia’s NIM and Microsoft’s Azure AI Foundry within the coming weeks.