Mistral AI builds fast, efficient language models. Ministral 8B is their latest small model with excellent multilingual support under Apache 2.0. Mistral Nemo 12B offers 128K context for document processing.
Deploy Mistral in minutes
Starting at $0.53/hr on dedicated GPU
| Model | GPU | VRAM | Price | Action |
|---|---|---|---|---|
Ministral 8B 8B (Fast) | L4 | 24 GB | $0.53/hr | Deploy |
Mistral Nemo 12B Nemo (12B) | L4 | 24 GB | $0.53/hr | Deploy |
Mistral 7B 7B (Legacy) | L4 | 24 GB | $0.53/hr | Deploy |
Prices include 30% service fee. Billed per minute while running.
Mistral requires 24GB VRAM. Consumer GPUs like the RTX 5080 (16GB) or RTX 4090 (24GB) may not have enough memory for larger variants.
On ModelPilot, deploy on a dedicated cloud GPU (up to 80GB VRAM) starting at $0.53/hr with no setup required.
Mistral requires 24GB VRAM.
Starting at $0.53/hr on a dedicated GPU. Billed per minute while running, with auto-stop when credits run out.
Text models typically deploy in 5–15 minutes including model download.
You can run smaller variants locally if your GPU has enough VRAM. For larger variants or sustained production use, cloud GPUs offer more capacity and reliability.
Pick your GPU and have it running in minutes. No infrastructure setup required.