Workflow
Arm Kleidi
icon
Search documents
端侧AI加速落地,Arm如何出招?
Core Insights - The emergence of AI agents this year has created commercial opportunities for large model vendors and chip companies, with a notable shift towards edge AI development [2][3] - AI models are becoming smarter and more compact, leading to increased demand for data centers and cloud computing, emphasizing the importance of capturing the expanding edge-cloud collaborative AI chip market [2][3] Edge AI Expansion - Three key elements are essential for building AI systems: creating a ubiquitous platform from cloud to edge, optimizing power consumption and performance per watt, and the importance of software alongside hardware [3] - The energy consumption of data centers has surged from megawatt (MW) to gigawatt (GW) levels, with over 50% of this consumption attributed to racks and semiconductor devices [3] AI Capabilities and Market Trends - The focus is shifting from model training to inference, which is crucial for realizing AI's commercial value, enabling smarter decision-making in devices like robots and smartphones [4][5] - The computational requirements for training large models are significantly higher than for inference, necessitating a substantial amount of inference operations to achieve commercial returns [5] Chip Design Challenges - The evolution of AI and the slowdown of Moore's Law are increasing the technical challenges and costs associated with chip design, making time-to-market critical [6] - Arm's strategy includes offering standardized products and platform solutions, such as the upcoming Armv9 flagship CPU, which aims to enhance performance and efficiency [6][7] Data Center Market Dynamics - Arm is actively competing in the data center market, traditionally dominated by x86 architecture, with predictions that nearly 50% of computing power for major cloud service providers will be based on Arm architecture by 2025 [8][9] - The transition from general computing to AI computing in data centers is underway, with significant efficiency improvements reported by cloud service providers using Arm-based processors [9]
Cerence AI Partners with Arm to Deliver Enhanced LLM Capabilities at the Edge
Globenewswire· 2025-05-28 15:00
Core Insights - Cerence AI has announced a strategic partnership with Arm to enhance the performance of its embedded small language model, CaLLM Edge, by utilizing Arm's Kleidi software library [1][4] - The collaboration aims to address challenges faced by automakers in optimizing compute performance for AI capabilities, particularly for large language models [2][3] Company Overview - Cerence Inc. is a global leader in creating AI-powered user experiences in the automotive sector, with over 500 million cars equipped with its technology [5] - The company focuses on integrating voice and generative AI to enhance safety and connectivity for drivers and passengers [5] Partnership Details - The partnership will enable flexible distribution and parallelization of AI computation loads between CPUs and GPUs, improving speed and performance for CaLLM Edge [3] - Arm's technology is already utilized by 94% of global automakers, providing a foundational compute architecture for AI applications in vehicles [2] Technological Advancements - The Kleidi software library is designed to accelerate machine learning and optimize neural network operations on Arm-based devices, facilitating real-time language processing [2][3] - CaLLM Edge operates efficiently on Arm-based chipsets, demonstrating high performance despite limited compute power and intensive processing requirements [3] Future Outlook - Both companies express excitement about the partnership, aiming to set new standards for performance and efficiency in automotive edge computing [4] - The collaboration is expected to lead to innovative application-specific AI models in vehicles, enhancing the in-car user experience [4]