Skip to content
FAQ

Mistral Medium 3.5 Arrives With 77.6% SWE-Bench Score and Cloud-Native Coding Agents

Mistral AI launched Medium 3.5, a dense 128B open-weight model that scores 77.6% on SWE-Bench Verified, alongside Vibe remote agents that run asynchronous cloud coding sessions from the CLI or Le Chat. The dual release positions Mistral as a serious challenger in the agentic developer-tools market just ahead of Google I/O 2026.

4 min read

Mistral AI shipped two intertwined products on May 2 that together reframe what a mid-tier open-weight model can do: Mistral Medium 3.5, a dense 128-billion-parameter flagship that scores 77.6% on SWE-Bench Verified, and Vibe Remote Agents, an asynchronous cloud execution layer that lets developers spawn long-running coding sessions from a single command line.

The release is strategically timed. Google I/O is weeks away, Anthropic’s Claude 4 family is expanding, and OpenAI is competing on every front. Mistral needed a statement. Medium 3.5 is it.

One Model, Three Regimes

The core architectural bet behind Medium 3.5 is density over sparsity. Unlike Mistral Large 3, which deploys a 675B mixture-of-experts architecture with only 41B active parameters per token, Medium 3.5 keeps all 128 billion parameters active on every inference pass. The tradeoff is raw compute cost per token — but the gain is consistency, especially on multi-step agentic workflows where sparse models can produce erratic outputs across sequential reasoning steps.

The model ships with a 256K token context window and native multimodal vision support, handling text, images, and structured data without an external transcription layer. Reasoning effort is now configurable per API request: developers can dial compute down for simple lookups and up for deep multi-step analysis, using the same model across both regimes.

On benchmarks, Medium 3.5 posts numbers that challenge models significantly above its price tier. The headline figure — 77.6% on SWE-Bench Verified — outperforms Devstral 2 (Mistral’s own coding-specialist model) and beats Qwen3.5 397B, a model with more than three times the parameter count. On the Tau3-Telecom agentic benchmark, Medium 3.5 hits 91.4%, suggesting particular strength in tool-use-intensive environments.

For API users on la Plateforme, pricing sits at $0.40 per million input tokens and $1.20 per million output tokens — roughly half the cost of comparable frontier models from Anthropic and OpenAI at equivalent benchmark performance levels.

Vibe Goes Asynchronous and Cloud-Native

The second half of the announcement is arguably more disruptive to developer workflows than the model itself.

Mistral’s Vibe coding tool, previously a local CLI for interactive agentic development, now supports remote agents: coding sessions that run asynchronously in isolated cloud sandboxes, with native integrations for GitHub, Slack, and standard CI/CD pipelines. The practical upshot is that a developer can issue a task — “fix this bug class, open a PR” — and walk away. When the session completes, a pull request lands in GitHub with full diff, test results, and an explanation of every decision made.

Sessions can be spawned from either the Vibe CLI or from Le Chat, Mistral’s consumer-facing assistant. Crucially, local sessions can be “teleported” to the cloud mid-task without losing session history or task state — a feature designed for the common scenario where a developer starts debugging locally, realizes the task is longer than expected, and wants to continue without keeping their laptop open overnight.

The sandboxed execution model is not simply convenient; it is a security architecture decision. Each remote agent session runs in an ephemeral environment with no persistent filesystem access, and all credentials are injected at runtime via Mistral’s secrets management layer. For enterprises wary of persistent AI access to codebases, this addresses a real objection.

Le Chat Gets Work Mode

The announcement also previewed Work Mode in Le Chat, a parallel agentic interface for non-developer enterprise use. Work Mode connects Le Chat directly to email, calendar, Jira, Slack, and document management systems via built-in connectors. The model can execute multi-step workflows across these tools simultaneously — drafting a project update, scheduling a review meeting, and updating a Jira ticket in a single orchestrated run.

The differentiator Mistral is emphasizing: all tool calls and reasoning steps are visible to the user in real time, and explicit approval is required before any sensitive actions are taken (sending an email, modifying a calendar event). This is a deliberate counter-positioning to more autonomous agent frameworks that black-box their decision chains.

Le Chat’s Work Mode targets the same enterprise segment that Microsoft 365 Copilot and Google Workspace AI are aggressively pursuing — but with a model that is open-weight, deployable on-premises, and available under an Apache 2.0-compatible license for most use cases.

The Competitive Calculus

The timing and packaging of this release expose a clear strategic logic. Mistral’s strongest differentiators have always been openness, European regulatory positioning (critical for GDPR-sensitive enterprises), and competitive performance-per-dollar. Medium 3.5 reinforces all three.

Against Cursor and Devin, Vibe Remote Agents competes directly on the agentic coding workflow, with the advantage of being model-provider-native rather than a wrapper layer. Against Claude for Work (Anthropic’s enterprise offering) and Microsoft Copilot, Le Chat Work Mode offers open-weight optionality that neither competitor can match.

The benchmark credibility matters, too. A 77.6% SWE-Bench Verified score from a 128B model — available at API prices well below the frontier tier — changes the calculus for startups and mid-market enterprises that have been priced out of GPT-5.5 or Claude 4 at scale.

What Comes Next

Mistral has not announced a firm timeline for Medium 3.5 weights availability under an open license, though the company’s pattern has been to follow API releases with open weight releases within four to eight weeks. The Hugging Face repository for mistralai/Mistral-Medium-3.5-128B is already live with model cards, and quantized versions from Unsloth are already appearing in the community.

For developers, the operative question is not whether Medium 3.5 is competitive on benchmarks — it clearly is. The question is whether Vibe Remote Agents’ cloud execution model can accumulate the developer ecosystem gravity that competing tools have built. That is a distribution problem, not a capability problem, and it is one Mistral has historically struggled with more than the technical side.

Google I/O and whatever Google announces for Gemini’s agentic developer story will be the real stress test for this week’s launch.

Mistral AI open-weight models coding agents developer tools SWE-Bench Le Chat
Share

Related Stories

Microsoft Releases Agent Framework 1.0: Production-Ready Multi-Agent SDK for .NET and Python

Microsoft has shipped Agent Framework 1.0, the stable, production-ready release of its open-source multi-agent SDK for .NET and Python. The framework unifies Semantic Kernel's enterprise foundations with AutoGen's orchestration concepts into a single SDK, supports seven first-party model providers including Claude, Gemini, and Ollama, and introduces stable APIs with a long-term support commitment — giving enterprise developers their first mature platform for building production multi-agent AI systems.

5 min read

Vibe Coding Is Flooding the App Store: New App Releases Up 104% in April 2026

AI coding tools like Claude Code, Cursor, and Replit have triggered a historic surge in mobile app submissions: worldwide releases are up 60% year-over-year in Q1 2026 and 104% in April. The vibe-coding wave is democratizing app development but also straining Apple's review system and raising new questions about quality and discoverability.

5 min read