Mistral AI, in collaboration with All Arms AI, has launched up to date variations of its developer-focused giant language fashions beneath the Devstral 2507 label. The discharge consists of two fashions—Devstral Small 1.1 and Devstral Medium 2507—designed to help agent-based code reasoning, program synthesis, and structured process execution throughout giant software program repositories. These fashions are optimized for efficiency and value, making them relevant for real-world use in developer instruments and code automation techniques.
Devstral Small 1.1: Open Mannequin for Native and Embedded Use
Devstral Small 1.1 (additionally referred to as devstral-small-2507
) is predicated on the Mistral-Small-3.1 basis mannequin and comprises roughly 24 billion parameters. It helps a 128k token context window, which permits it to deal with multi-file code inputs and lengthy prompts typical in software program engineering workflows.
The mannequin is fine-tuned particularly for structured outputs, together with XML and function-calling codecs. This makes it suitable with agent frameworks corresponding to OpenHands and appropriate for duties like program navigation, multi-step edits, and code search. It’s licensed beneath Apache 2.0 and out there for each analysis and industrial use.

Efficiency: SWE-Bench Outcomes
Devstral Small 1.1 achieves 53.6% on the SWE-Bench Verified benchmark, which evaluates the mannequin’s skill to generate appropriate patches for actual GitHub points. This represents a noticeable enchancment over the earlier model (1.0) and locations it forward of different brazenly out there fashions of comparable dimension. The outcomes have been obtained utilizing the OpenHands scaffold, which gives a regular check setting for evaluating code brokers.
Whereas not on the stage of the most important proprietary fashions, this model gives a steadiness between dimension, inference price, and reasoning efficiency that’s sensible for a lot of coding duties.
Deployment: Native Inference and Quantization
The mannequin is launched in a number of codecs. Quantized variations in GGUF can be found to be used with llama.cpp
, vLLM
, and LM Studio. These codecs make it potential to run inference domestically on high-memory GPUs (e.g., RTX 4090) or Apple Silicon machines with 32GB RAM or extra. That is helpful for builders or groups that favor to function with out dependency on hosted APIs.
Mistral additionally makes the mannequin out there by way of their inference API. The present pricing is $0.10 per million enter tokens and $0.30 per million output tokens, the identical as different fashions within the Mistral-Small line.

Devstral Medium 2507: Increased Accuracy, API-Solely
Devstral Medium 2507 will not be open-sourced and is simply out there by the Mistral API or by enterprise deployment agreements. It gives the identical 128k token context size because the Small model however with larger efficiency.
The mannequin scores 61.6% on SWE-Bench Verified, outperforming a number of industrial fashions, together with Gemini 2.5 Professional and GPT-4.1, in the identical analysis framework. Its stronger reasoning capability over lengthy contexts makes it a candidate for code brokers that function throughout giant monorepos or repositories with cross-file dependencies.
API pricing is ready at $0.40 per million enter tokens and $2 per million output tokens. Fantastic-tuning is on the market for enterprise customers by way of the Mistral platform.
Comparability and Use Case Match
Mannequin | SWE-Bench Verified | Open Supply | Enter Price | Output Price | Context Size |
---|---|---|---|---|---|
Devstral Small 1.1 | 53.6% | Sure | $0.10/M | $0.30/M | 128k tokens |
Devstral Medium | 61.6% | No | $0.40/M | $2.00/M | 128k tokens |
Devstral Small is extra appropriate for native improvement, experimentation, or integrating into client-side developer instruments the place management and effectivity are necessary. In distinction, Devstral Medium gives stronger accuracy and consistency in structured code-editing duties and is meant for manufacturing companies that profit from larger efficiency regardless of elevated price.
Integration with Tooling and Brokers
Each fashions are designed to help integration with code agent frameworks corresponding to OpenHands. The help for structured operate calls and XML output codecs permits them to be built-in into automated workflows for check era, refactoring, and bug fixing. This compatibility makes it simpler to attach Devstral fashions to IDE plugins, model management bots, and inside CI/CD pipelines.
For instance, builders can use Devstral Small for prototyping native workflows, whereas Devstral Medium can be utilized in manufacturing companies that apply patches or triage pull requests based mostly on mannequin recommendations.
Conclusion
The Devstral 2507 launch displays a focused replace to Mistral’s code-oriented LLM stack, providing customers a clearer tradeoff between inference price and process accuracy. Devstral Small gives an accessible, open mannequin with adequate efficiency for a lot of use circumstances, whereas Devstral Medium caters to purposes the place correctness and reliability are important.
The provision of each fashions beneath totally different deployment choices makes them related throughout numerous levels of the software program engineering workflow—from experimental agent improvement to deployment in industrial environments.
Take a look at the Technical particulars, Devstral Small mannequin weights at Hugging Face and Devstral Medium can even be out there on Mistral Code for enterprise clients and on finetuning API. All credit score for this analysis goes to the researchers of this challenge. Additionally, be happy to comply with us on Twitter, and Youtube and don’t overlook to affix our 100k+ ML SubReddit and Subscribe to our Publication.
Sana Hassan, a consulting intern at Marktechpost and dual-degree scholar at IIT Madras, is enthusiastic about making use of know-how and AI to handle real-world challenges. With a eager curiosity in fixing sensible issues, he brings a contemporary perspective to the intersection of AI and real-life options.