Workflow
ArtifactsBench
icon
Search documents
赛道Hyper | 腾讯混元开源Hunyuan-A13B:1张AI卡搞定
Hua Er Jie Jian Wen· 2025-07-02 12:15
Core Insights - Tencent Hunyuan has open-sourced its first hybrid inference MoE model, Hunyuan-A13B, along with two new datasets, ArtifactsBench and C3-Bench, to advance the development of large models [1][6] Model Features - Hunyuan-A13B has a total of 80 billion parameters and 13 billion active parameters, providing advantages in inference efficiency compared to similar open-source models [1] - The model supports a native context window of 256K, allowing it to handle long documents effectively, which is beneficial in academic, legal, and business contexts [3] - It demonstrates superior multi-tool collaboration capabilities, enabling it to perform complex tasks across various scenarios, unlike models with single tool capabilities [2] Developer Accessibility - The model is developer-friendly, allowing deployment on mid-range GPUs, such as NVIDIA GeForce GTX series, and supports multiple quantization formats [4] - Developers can access the model through open-source communities like GitHub and Huggingface, as well as via Tencent Cloud's API [4] Training Methodology - The pre-training phase utilized a high-quality corpus of 20 trillion tokens across various fields, enhancing the model's general knowledge [5] - A multi-stage training approach was adopted to improve different capabilities, including logical reasoning and creative writing [5] Evaluation Tools - ArtifactsBench includes 1,825 tasks across nine domains for assessing code generation capabilities, while C3-Bench focuses on agent scenarios with 1,024 test data points [6] - These datasets aim to establish more comprehensive evaluation standards in the industry, facilitating model optimization [6][7] Application and Future Plans - Hunyuan-A13B is already applied in over 400 internal Tencent businesses, with an average daily request volume of 130 million [6] - Future plans include launching dense models ranging from 0.5B to 32B and continuing to open-source multimodal foundational models and plugins [6]