1. Model Introduction
The largest model in the Ministral 3 family, Ministral 3 14B offers frontier capabilities and performance comparable to its larger Mistral Small 3.2 24B counterpart. A powerful and efficient language model with vision capabilities. The Ministral 3 14B Instruct model offers the following capabilities: Vision: Enables the model to analyze images and provide insights based on visual content, in addition to text. Multilingual: Supports dozens of languages, including English, French, Spanish, German, Italian, Portuguese, Dutch, Chinese, Japanese, Korean, Arabic. System Prompt: Maintains strong adherence and support for system prompts. Agentic: Offers best-in-class agentic capabilities with native function calling and JSON outputting. Edge-Optimized: Delivers best-in-class performance at a small scale, deployable anywhere. Apache 2.0 License: Open-source license allowing usage and modification for both commercial and non-commercial purposes. Large Context Window: Supports a 256k context window. For further details, please refer to the official documentation2. SGLang Installation
Please refer to the official SGLang installation guide for installation instructions.3. Model Deployment
This section provides deployment configurations optimized for different hardware platforms and use cases.3.1 Basic Configuration
Interactive Command Generator: Use the configuration selector below to automatically generate the appropriate deployment command for your hardware platform, model variant, deployment strategy, and thinking capabilities.3.2 Configuration Tips
Context length vs memory: Ministral-3 advertises a long context window; if you are memory-constrained, start by lowering —context-length (for example 32768) and increase once things are stable. Pre-installation steps: Adding the following steps after launching the dockerCommand
4. Model Invocation
4.1 Basic Usage
For basic API usage and request examples, please refer to:4.2 Advanced Usage
4.2.1 Launch the docker
Command
Command
4.2.2 Launch the server
Command
5. Benchmark
This section uses industry-standard configurations for comparable benchmark results.5.1 Speed Benchmark
Test Environment:- Hardware: MI300X GPU (8x)
- Model: mistralai/Ministral-3-14B-Instruct-2512
- Tensor Parallelism: 1
- SGLang Version: 0.5.7
- Model Deployment Command:
Command
Low Concurrency
- Benchmark Command:
Command
- Test Results:
Output
Medium Concurrency
- Benchmark Command:
Command
- Test Results:
Output
High Concurrency
- Benchmark Command:
Command
- Test Results:
Output
5.2 Accuracy Benchmark
Document model accuracy on standard benchmarks:5.2.1 GSM8K Benchmark
- Benchmark Command
Command
Output
