Devstral Small 1.1 is a software engineering-specific large language model jointly developed by Mistral AI and All Hands AI. It is fine-tuned from Mistral-Small-3.1, with its vision encoder removed to focus solely on text-based programming tasks. Below is a detailed introduction: Technical Specifications Model Parameters and Context Window: Devstral Small 1.1 has 24B parameters and supports a 128k token context window, enabling it to handle extensive code files and long-context programming tasks. Tokenizer: It uses a custom Tekken tokenizer with a 131k vocabulary size, which helps improve the model’s understanding and processing of code-related text. Performance Metrics: On the SWE-bench …
Devstral-Small-2505: A Comprehensive Guide to Deployment, Fine-Tuning, and Practical Applications Devstral Model Example 1. Introduction and Technical Background 1.1 What is Devstral-Small-2505? Devstral-Small-2505 is a software engineering-specific large language model developed collaboratively by Mistral AI and All Hands AI. Designed for codebase exploration, multi-file editing, and engineering agent tasks, this model is fine-tuned from Mistral-Small-3.1 with its vision encoder removed, focusing solely on text-based programming. 1.2 Core Performance Metrics 128K Token Context Window: Handles extensive code files 46.8% Accuracy on SWE-bench (as of May 2025) State-of-the-art 5-shot MMLU Benchmark Performance 24B Parameters: Runs on a single RTX 4090 or 32GB …