Mistral released new models:
https://mistral.ai/news/mistral-small-4
Small evolved into a 119B MoE model with 6B active parameters. (Small 3 used to be a dense 24 GB model instead.)
Another new model is Leanstral, which serves as a proof assistant:
https://mistral.ai/news/leanstral
---
I have been switching from Devstral Small 2 to Qwen Coder Next; but since Mistral Small 4 can be seen as a successor to Devstral Small, will be interesting to compare.
A setup that I have been using recently is to have Mistral Vibe running in a Docker container without network, mounting only a writable copy of the current working directory. 127.0.0.1:8080 is exposed, so it can still connect to a local API (llama.cpp in my case).
Since this setup reduces the blast radius, I can allow Vibe to use tools without confirmation.
Not sure if it is useful for someone else - it is tailored to my own use case and will likely not work outside of Linux - but I put the code here:




