Zelili AI

Devstral 2

Mistral AI’s State of the Art Open-Source Agentic Coding Model Leading Performance in Software Engineering Tasks
Tool Release Date

9 Dec 2025

Tool Users
6.2 Million+
0.0
πŸ‘ 54

About This AI

Devstral 2 is Mistral AI’s next-generation coding model family released on December 9, 2025, available in Devstral 2 (123B parameters) and Devstral Small 2 (24B parameters).

It is a dense transformer optimized for agentic software engineering, supporting codebase exploration, multi-file edits, architecture-level context, and tool orchestration for autonomous code agents.

With a 256K token context window, it enables processing large repositories in one pass while maintaining coherence.

The model achieves 72.2% on SWE-bench Verified (SOTA among open-weight models), outperforming many peers in real-world GitHub issue resolution and software reasoning benchmarks.

Released under modified MIT license (Devstral 2) and Apache 2.0 (Small variant), it is fully open-source with weights on Hugging Face for local deployment.

Currently free via Mistral API during initial period; post-free pricing is $0.40/$2.00 per million tokens (input/output) for the 123B version and lower for Small.

Paired with Mistral Vibe CLI (terminal-native agent), it enables end-to-end code automation, custom subagents, slash-command skills, and workflow configuration.

Ideal for developers, AI agents, software teams, and researchers building autonomous coding systems, with strong cost-efficiency (up to 7x better than some closed models) and local runnability on consumer hardware for the Small version.

Key Features

  1. 256K token context window: Handles entire large codebases and long conversations without truncation
  2. Agentic coding excellence: Explores repositories, edits multiple files, maintains architecture context, and orchestrates changes
  3. SOTA open-weight performance: 72.2% on SWE-bench Verified for real-world software engineering tasks
  4. Multi-size variants: 123B dense for maximum capability; 24B Small for local/consumer hardware deployment
  5. Tool use and reasoning: Built for autonomous agents with reliable tool calling and multi-step planning
  6. Open-source availability: Full weights under modified MIT/Apache 2.0, easy Hugging Face integration
  7. Mistral Vibe CLI integration: Native terminal agent with custom subagents, skills, and workflows
  8. High cost-efficiency: Up to 7x more efficient than competitors like Claude Sonnet in real tasks
  9. Strong benchmarks: Competitive or leading in coding, reasoning, and agentic evaluations vs open peers

Price Plans

  1. Free (Current API period $0): Free access via Mistral API during launch period; also free on Experiment plan in Mistral Studio
  2. API Paid (Post-free $0.40/$2 per M tokens input/output for 123B): Standard token pricing after promo; lower for Small 2 ($0.10/$0.30)
  3. Le Chat Pro/Team ($14.99–$24.99/Month): Access to Devstral 2 via Vibe CLI, higher limits, and integrated coding workflows

Pros

  1. Top open coding model: Sets new SOTA for open-weight agentic coding with high SWE-bench score
  2. Long context support: 256K tokens enable full-repo understanding without splitting
  3. Free initial access: Currently free via API; Small variant runs locally on consumer GPUs
  4. Strong efficiency: Delivers high performance at lower cost than many closed models
  5. Developer-focused: Pairs perfectly with Mistral Vibe for terminal-native automation
  6. Open weights: Full customization, fine-tuning, and local deployment possible
  7. Rapid innovation: Builds on Mistral's fast release cadence for coding AI

Cons

  1. High compute for full model: 123B requires significant GPUs (e.g., 4+ H100 equivalents)
  2. Free period temporary: API moves to paid ($0.40/$2 input/output per M tokens) after initial offer
  3. Verbose outputs: Can produce longer responses unless prompted for conciseness
  4. Setup for local use: Small variant easier, but full needs enterprise hardware
  5. No native multimodal: Primarily text/code-focused; limited vision compared to some rivals
  6. Recent release: Adoption metrics still emerging; no massive user base numbers yet

Use Cases

  1. Autonomous code agents: Build agents that explore repos, fix bugs, refactor code across files
  2. Software engineering tasks: Resolve GitHub issues, implement features, review PRs
  3. Local development: Run Small variant on laptop for offline coding assistance
  4. CLI automation: Use Mistral Vibe for terminal-based end-to-end workflows
  5. Research in agentic AI: Fine-tune or extend for custom software agents
  6. Team productivity: Integrate into IDEs or CI/CD for faster iteration
  7. Cost-effective coding: High performance at fraction of closed-model costs

Target Audience

  1. Software developers: Needing agentic help for coding, debugging, and refactoring
  2. AI agent builders: Creating autonomous software engineering systems
  3. Open-source enthusiasts: Running and customizing top open coding models locally
  4. Dev teams: Boosting productivity with integrated CLI agents and API
  5. Researchers: Benchmarking or advancing agentic coding frontiers
  6. Startups/enterprises: Seeking efficient, open alternatives to proprietary LLMs

How To Use

  1. Access via API: Sign up at console.mistral.ai, get API key, and call Devstral 2 endpoint
  2. Try in Mistral Studio: Use free Experiment plan or Le Chat Pro for immediate testing
  3. Run locally (Small): Download weights from Hugging Face, set up with transformers/vLLM
  4. Use Mistral Vibe CLI: Install via pip/npm, configure with API key or local model
  5. Prompt for coding: Ask agentic tasks like 'fix this bug in repo X' or 'refactor module Y'
  6. Integrate in tools: Use in IDEs via extensions or CLI for seamless workflow
  7. Monitor usage: Track tokens in console; switch to Small for cost/latency savings

How we rated Devstral 2

  • Performance: 4.8/5
  • Accuracy: 4.7/5
  • Features: 4.8/5
  • Cost-Efficiency: 4.9/5
  • Ease of Use: 4.5/5
  • Customization: 4.9/5
  • Data Privacy: 4.8/5
  • Support: 4.4/5
  • Integration: 4.7/5
  • Overall Score: 4.7/5

Devstral 2 integration with other tools

  1. Mistral API: Direct integration for web/apps via console.mistral.ai endpoints
  2. Mistral Vibe CLI: Native terminal agent for local or API-powered end-to-end coding automation
  3. Hugging Face: Model weights and inference pipelines for local deployment and fine-tuning
  4. IDE Extensions: Compatible with VS Code, Cursor, Zed via CLI or API wrappers
  5. Agent Frameworks: Works with LangChain, LlamaIndex, AutoGen for custom agentic systems

Best prompts optimised for Devstral 2

  1. Explore this GitHub repository [repo URL], understand the architecture, and suggest improvements to the main module while maintaining compatibility
  2. Fix the failing test in file tests/unit.py by editing the function in src/core.py; reason step-by-step and apply changes
  3. Refactor this legacy codebase to use modern async patterns in Python; update multiple files and add type hints
  4. Implement a new feature: add user authentication with JWT to this Express.js app; include routes, middleware, and tests
  5. Analyze this pull request diff and provide a detailed code review with suggestions for security and performance
Devstral 2 sets a new benchmark for open-source agentic coding models with 72.2% on SWE-bench Verified and a massive 256K context window. The 123B version delivers frontier performance, while Small 2 enables local use. Free initial API access and strong cost-efficiency make it ideal for developers building autonomous agents and software tools.

FAQs

  • What is Devstral 2?

    Devstral 2 is Mistral AI’s next-generation open-source coding model family (123B and 24B variants) optimized for agentic software engineering tasks, released December 9, 2025.

  • When was Devstral 2 released?

    It was officially released on December 9, 2025, alongside Mistral Vibe CLI for terminal-native coding automation.

  • What is the performance of Devstral 2?

    Devstral 2 achieves 72.2% on SWE-bench Verified, setting SOTA among open-weight models for real-world software engineering benchmarks.

  • Is Devstral 2 free to use?

    Currently free via Mistral API during initial period; post-free API pricing is $0.40/$2 per million tokens (input/output) for the 123B version.

  • What is the context window for Devstral 2?

    It supports a 256,000 token context window, enabling full codebase analysis in a single pass.

  • How does Devstral 2 compare to other models?

    It outperforms many open peers on SWE-bench and is more cost-efficient than closed models like Claude Sonnet for real tasks.

  • Can Devstral Small 2 run locally?

    Yes, the 24B Devstral Small 2 is designed for consumer hardware deployment with the same long context and strong performance.

  • What is Mistral Vibe?

    Mistral Vibe is a native CLI agent powered by Devstral 2 for end-to-end code automation, custom workflows, and terminal integration.

Newly Added Tools​

CodeRabbit

$0/Month

Code Genius

$0/Month

AskCodi

$20/Month

PearAI

$0/Month
Devstral 2 Alternatives

CodeRabbit

$0/Month

Code Genius

$0/Month

AskCodi

$20/Month

Devstral 2 Reviews

0.0
0.0 out of 5 stars (based on 0 reviews)
Excellent0%
Very good0%
Average0%
Poor0%
Terrible0%

There are no reviews yet. Be the first one to write one.