AI重磅!华为“黑科技”来了
Zhong Guo Ji Jin Bao·2025-08-12 07:40

Core Insights - Huawei has officially launched its AI inference technology UCM (Unified Cache Manager), aimed at addressing challenges in AI inference efficiency and user experience [1] - The AI industry is shifting focus from maximizing model capabilities to optimizing inference experiences, which directly impacts user satisfaction and commercial viability [1] Group 1: UCM Technology Overview - UCM is a KV Cache-centered inference acceleration suite that integrates various caching algorithms to manage KV Cache memory data during inference, enhancing throughput and reducing latency [2] - The growth of AI inference demands has led to an increase in KV Cache capacity, which has exceeded GPU memory limits, necessitating innovative solutions like UCM [2][3] - UCM's core value lies in providing faster inference responses and longer inference sequences, addressing the limitations of current AI models [2] Group 2: Performance Improvements - UCM enables dynamic KV unloading and position encoding expansion, achieving a tenfold increase in inference context window [3] - The technology allows for on-demand data flow across different storage media (HBM, DRAM, SSD), improving TPS (tokens per second) by 2 to 22 times, thereby reducing the cost per token [4] - Current mainstream AI models in China output tokens at a significantly lower speed compared to their international counterparts, highlighting the need for UCM's capabilities [4] Group 3: Practical Applications - Huawei's AI inference acceleration solution, in collaboration with China UnionPay, is being piloted in three business scenarios: customer voice, marketing planning, and office assistant [5] - The office assistant application can support user inputs exceeding 170,000 tokens, overcoming challenges associated with long sequence models [5]