Workflow
模型算子化
icon
Search documents
阿里云智能集团资深副总裁李飞飞:内存价格预计还会上涨两到三倍
Group 1 - The core viewpoint of the news is that Alibaba Cloud has officially launched a series of new product capabilities, including the AI Data Lake (Lakebase), which integrates large model capabilities into its database, enhancing AI-driven decision-making [1][3] - The AI-native database is seen as an inevitable direction for technological evolution, with memory prices having surged by 30% to 40%, and expected to increase by two to three times in the future [3] - Alibaba Cloud's PolarDB has been deployed at a scale exceeding 3 million cores, covering 86 availability zones globally [3] Group 2 - The integration of large models into databases is expected to evolve recognition capabilities, allowing data systems to store and query multimodal data while directly driving AI intelligent decision-making [1] - The transition to AI-driven databases is crucial as large models can generate inaccuracies when detached from real-time data, with the key to addressing this being the real-time conversion of hot data to tokens [3] - Future token usage may increase by 100 to 1000 times, relying on various agentic AI to achieve contextual applications and unlock their value [3]
阿里云重新定义AI时代数据库
Hua Er Jie Jian Wen· 2026-01-21 10:18
Core Viewpoint - Alibaba Cloud's approach to the "AI Native" trend is more pragmatic, focusing on being "AI Ready" rather than rushing to label their products as "AI Native" [3][4]. Group 1: AI Readiness - The concept of "AI Ready" is explained through a "4+1" formula, emphasizing the need for databases to evolve from traditional structured data storage to a more versatile "Lakebase" that can handle various data types [4][5]. - The first step towards "AI Ready" is transforming databases into a "Lakebase" that can store both structured and unstructured data, allowing for better data management [4][8]. - The second key aspect is unified metadata management, which is crucial for handling the diverse and large volumes of data generated in the AI era [8][9]. - The third capability involves multi-modal retrieval and processing, integrating structured, semi-structured, and unstructured data [9][11]. - The fourth aspect includes model operatorization and support for AI agents, enabling real-time data processing and interaction with AI models [11][12]. Group 2: Cost Efficiency - Alibaba Cloud emphasizes cost efficiency through resource pooling, multi-tenancy, and elastic scaling, which are essential in the context of rising hardware prices [13][14]. - The "Serverless" model allows for extreme elasticity, enabling businesses to only pay for resources when needed, thus reducing costs during periods of low demand [15][16]. - The company highlights the importance of scale in achieving cost advantages, as larger operations can better absorb costs and provide savings to customers [36][38]. Group 3: Future of AI Native Databases - The transition from "AI Ready" to "AI Native" is seen as a gradual process, with specific criteria needed to define a database as "AI Native," such as a significant portion of users being AI agents and outputs being predominantly tokens [23][24]. - The future landscape is expected to be dominated by AI agents utilizing databases, with a focus on token-based outputs rather than traditional data formats [24][26]. - The integration of various AI capabilities, including natural language processing and multi-modal interactions, is essential for enhancing user experience and database functionality [20][21]. Group 4: Industry Trends and Challenges - The current trend in the AI landscape is characterized by rapid evolution, making it premature for companies to claim they have achieved "AI Native" status [22][30]. - The ongoing rise in memory and storage prices is expected to be a long-term challenge, impacting the overall cost structure of cloud services [39][40]. - Companies are encouraged to leverage cloud and AI platforms to maximize value, especially during periods of rising costs, as they can provide greater efficiency and scalability compared to traditional self-managed resources [40].