Core Viewpoint - The article discusses the emergence of extreme events in digital mirrors, emphasizing that these events are not due to code vulnerabilities but rather arise from the spontaneous emergence of systems. A research team from Shanghai AI Laboratory and several universities aims to dissect the evolution of these "black swan" events within multi-agent systems (MAS) [2][4]. Group 1: Emergence of Multi-Agent Systems - The year 2023 marked the rise of large language models (LLMs) driving MAS simulations of human society, with Stanford's "Smallville" gaining significant attention [5]. - Various complex MAS sandboxes have been developed to replicate macroeconomic systems, financial markets, and social networks, effectively creating digital mirrors of human society [6]. - As system complexity increases, concerning phenomena such as inflation, stock market crashes, and group polarization have been observed, mirroring real-world "black swan" events [7]. Group 2: The Black Box Challenge - The intricate non-linear interactions among agents create a significant "black box" challenge, making it difficult to pinpoint the origins of crises within these systems [11]. - The research team introduced a diagnostic framework for extreme events in MAS, utilizing the Shapley Value from game theory to allocate disaster risk among agents based on their actions [13]. - The framework categorizes risk contributions along three dimensions: time, agent, and behavior pattern, allowing for precise quantification of marginal impacts on crises [13]. Group 3: Findings on Extreme Event Evolution - The research identified five common evolutionary patterns of extreme events across different scenarios, indicating that such events are systematic and understandable rather than random [17]. - Discovery 1: Extreme events exhibit differentiated temporal evolution characteristics, either accumulating risks over time or triggering instantaneously [19]. - Discovery 2: A small number of high-risk agents often drive extreme events [20]. - Discovery 3: Agents contributing significantly to system collapse tend to display high instability in their daily behaviors [20]. - Discovery 4: Agents develop implicit agreements, leading to synchronized increases or decreases in system risk [20]. - Discovery 5: Most risks leading to system collapse stem from a few specific behavior patterns [20]. Group 4: Implications for Risk Management - Experimental results show that by removing high-risk actions identified through the framework, the overall risk of system collapse can significantly decrease [21]. - The findings suggest that targeted regulation and intervention of high-risk agents and behaviors can prevent crises in both AI-simulated environments and real-world scenarios [22]. Conclusion - The article emphasizes the importance of understanding and explaining the emergence phenomena in multi-agent systems to create a safer future [23].
在AI社会抓「内鬼」?上海AI Lab推出首个多智能体极端事件解释框架
机器之心·2026-03-04 09:15