Mixtral 8x22B
Developed by Mistral AI
Key Capabilities
- Efficient MoE: 176B parameters but only 44B active per token
- 65K context window
- Apache 2.0 - fully open for commercial use
- Strong multilingual capabilities
- Native function calling support
VRAM Requirements by Quantization
Choose the right GPU based on your performance and quality needs.
| Model / Quantization | VRAM Required |
|---|---|
| FP16 | 352GB |
| Q4 | 100GB |
| Q2 | 55GB |
Use Cases
Mixtral 8x22B (176B total (44B active via 8-expert MoE, 2 active per token)) can be deployed for enterprise AI applications including document processing, code generation, data analysis, and conversational AI. License: Apache 2.0.
Run Mixtral 8x22B with Petronella
PTG deploys Mixtral 8x22B as a cost-effective MoE model under Apache 2.0. Ideal for businesses needing frontier-class output at lower hardware cost than dense models of equivalent quality.
Recommended Hardware
| Model Size | Recommended GPU |
|---|---|
| FP16 | DGX Spark (128GB) or 2x RTX PRO 6000 (192GB) |
| Q4 | RTX PRO 6000 Blackwell (96GB) or 2x RTX 5090 (64GB) |
Deploy Mixtral 8x22B On-Premises
Our team builds GPU-accelerated systems configured and optimized for Mixtral 8x22B. Private, secure, and fully under your control.