Run Mistral Vibe CLI with any OpenAI Compatible Server
tobrun.github.io·1d·
Discuss: r/LocalLLaMA
🦙Ollama
Preview
Report Post

Running Vibe with a Local Model (vLLM)

Vibe Mistral CLI, released this week, is Mistral’s take on the agentic code CLI. However, one small detail is not documented anywhere: how to make Vibe work with local models instead of Mistral official APIs. While there is an option local, it will only work for localhost. I however have a dedicated GPU server that I want to integrate.

In this blog, I’ll walk you through how I successfully connected Vibe CLI to a locally hosted model running on vLLM, using Devstral-2-123B-Instruct-2512 as the example.

This makes it possible to use Vibe fully offline.


🌐 Background

Mistral released Devstral-2 and created the Vibe CLI tooling. Their announcement mentions support for custom providers but **does not includ…

Similar Posts

Loading similar posts...