Yi Large
Developed by 01.AI
Key Capabilities
- Strong bilingual English and Chinese performance
- 200K extended context window
- Apache 2.0 license for open versions
- Competitive with LLaMA 2-70B at half the parameters
- Strong coding and mathematical reasoning
VRAM Requirements by Quantization
Choose the right GPU based on your performance and quality needs.
| Model / Quantization | VRAM Required |
|---|---|
| 34B FP16 | 68GB |
| 34B Q4 | 20GB |
Use Cases
Yi Large (34B (Yi-34B), with Yi-Large API at undisclosed larger size) can be deployed for enterprise AI applications including document processing, code generation, data analysis, and conversational AI. License: Apache 2.0 (Yi-34B open version).
Run Yi Large with Petronella
PTG deploys Yi models for businesses with Chinese-English bilingual AI needs. Apache 2.0 licensing and strong benchmark performance make Yi an excellent cost-effective choice.
Recommended Hardware
| Model Size | Recommended GPU |
|---|---|
| 34B FP16 | RTX PRO 6000 Blackwell (96GB) |
| 34B Q4 | RTX 5090 (32GB) or RTX PRO 5000 (48GB) |
Deploy Yi Large On-Premises
Our team builds GPU-accelerated systems configured and optimized for Yi Large. Private, secure, and fully under your control.