腾讯混元开源首个13B激活MoE大模型:推理提升超2倍,单卡可跑!

Core Insights - Tencent's Hunyuan-A13B is the first MoE model with 13B active parameters, featuring a total of 80B parameters and a native context length of 256K, achieving inference speeds over twice that of similar models, with daily calls exceeding 130 million [1][12][14] Model Performance - Hunyuan-A13B demonstrates strong general capabilities, achieving high scores on various authoritative industry benchmarks, particularly excelling in agent tool usage and long text processing [4][8][12] - In specific benchmark tests, Hunyuan-A13B outperformed other models in mathematics, reasoning, and instruction-following tasks, with notable scores such as 87.3 in AIME2024 and 91.8 in IF-Eval [7][10] Deployment and Accessibility - The model is available on open-source platforms like GitHub and Hugging Face, and its API is accessible via Tencent Cloud, allowing for easy integration and deployment [2][5][11] - Hunyuan-A13B is designed to be developer-friendly, requiring only a single mid-range GPU for deployment, and supports various quantization formats [11][12] Training Innovations - The model was pre-trained on 20 trillion tokens, covering multiple domains, which significantly enhances its general capabilities [12][13] - Tencent has developed a multi-stage training approach to improve the model's reasoning abilities while maintaining its creative and understanding capabilities [13][14] Future Developments - Tencent plans to continue its commitment to open-source by releasing more models of varying sizes and capabilities, including dense models ranging from 0.5B to 32B and MoE models with 13B active parameters [15]