AlphaZero
Search documents
10 years of AlphaGo: The turning point for AI | Thore Graepel & Pushmeet Kohli
Google DeepMind· 2026-03-10 17:28
Welcome back to Google Deep Mind the podcast. I'm Professor Hannah Fry. Picture this scene. It's March 2016. Inside a hotel suite in Soul, South Korea, two players are playing the ancient game of Go. A game of unimaginable complexity, long thought impossible for a machine to master. On one side is Lisa Doll, a legendary 18time Go world champion. on the other, Alph Go, a neural networkbased AI system built on a powerful technique called reinforcement learning. >> Welcome to the Deep Mind challenged live in S ...
DeepMind强化学习掌门人David Silver离职创业,Alpha系列AI缔造者,哈萨比斯左膀右臂
3 6 Ke· 2026-02-02 08:21
Core Insights - David Silver, a prominent researcher in reinforcement learning, has left DeepMind after 15 years to establish his own AI company, Ineffable Intelligence [1][5]. Company Formation - Ineffable Intelligence was quietly founded in November 2025, with Silver officially appointed as a director on January 16, 2026 [2]. - The company is headquartered in London and is actively recruiting AI research talent while seeking venture capital [3]. Contributions at DeepMind - Silver was a key figure in the development of DeepMind's "Alpha series," leading or significantly contributing to major projects such as AlphaGo, AlphaZero, MuZero, and AlphaStar [7][9]. - His work on AlphaGo, which defeated world champion Lee Sedol in 2016, marked a significant milestone in AI history [9]. - Silver has received multiple accolades, including the ACM Prize in Computing in 2019 and the Royal Academy of Engineering Silver Medal in 2017 [10]. Academic and Research Impact - Silver is one of the most published authors among DeepMind employees, with over 280,000 citations and an h-index of 104 according to Google Scholar [11]. - His research has focused on advancing AI capabilities beyond human knowledge, advocating for a new "Age of Experience" where AI learns from its own experiences [17][19]. Vision for AI - Silver aims to tackle the challenge of creating superintelligent AI that can learn independently from first principles, moving away from reliance on human knowledge [17][19].
AlphaGo之父David Silver离职创业,目标超级智能
机器之心· 2026-01-31 02:34
Core Viewpoint - David Silver, a prominent AI researcher from Google DeepMind, has left the company to establish a new startup named Ineffable Intelligence, focusing on solving complex AI challenges and pursuing superintelligence [1][3][4]. Group 1: Company Formation and Background - Ineffable Intelligence is being founded in London, with active recruitment for AI researchers and seeking venture capital [3]. - Silver was a key figure at Google DeepMind, contributing to significant achievements such as AlphaGo, AlphaStar, and AlphaZero, which demonstrated the capabilities of AI in complex games [9][12][14]. - The company was officially registered in November 2025, with Silver appointed as a director in January 2026 [4]. Group 2: Silver's Contributions and Vision - Silver's work includes developing AI systems that surpassed human capabilities in games, showcasing the potential of AI to learn and adapt [12][14]. - He emphasizes the need for AI to explore and discover knowledge independently, moving beyond human limitations and biases [18][23]. - The vision for Ineffable Intelligence is to create a self-learning superintelligence that can autonomously uncover foundational knowledge [23]. Group 3: Industry Context and Trends - Silver's departure follows a trend where notable AI researchers are leaving established labs to pursue startups focused on superintelligence, with significant funding being raised in the sector [15]. - Other notable figures, such as Ilya Sutskever and Yann LeCun, are also venturing into similar domains, indicating a growing interest in the pursuit of advanced AI capabilities [15][16].
DeepMind强化学习掌门人David Silver离职创业!Alpha系列AI缔造者,哈萨比斯左膀右臂
量子位· 2026-01-31 01:34
Core Viewpoint - David Silver, a prominent figure in reinforcement learning and a key researcher at DeepMind for 15 years, has left the company to establish his own AI startup, Ineffable Intelligence, aiming to tackle the challenges of achieving superintelligence in AI [2][21]. Group 1: Departure from DeepMind - David Silver has officially left DeepMind and has been appointed as the director of his new company, Ineffable Intelligence, which was quietly established in November 2025 [3][2]. - Silver had been on leave for several months prior to his departure from DeepMind [4]. - Google DeepMind confirmed Silver's departure and expressed gratitude for his contributions during his tenure [9]. Group 2: Achievements at DeepMind - Silver was instrumental in the development of several landmark AI projects at DeepMind, including AlphaGo, which defeated world champion Lee Sedol in 2016, marking a significant milestone in AI history [14]. - He also led the development of AlphaZero, which achieved superhuman performance in Go, chess, and shogi without relying on human game data [14]. - Silver contributed to the creation of MuZero, which learns to play games without being informed of the rules, and AlphaStar, which defeated top players in StarCraft II [15][16]. - He has received numerous accolades, including the ACM Prize in Computing in 2019 and the Royal Academy of Engineering Silver Medal in 2017 [18]. Group 3: Vision for the Future - Silver's motivation for founding Ineffable Intelligence is to return to the awe and wonder of solving the most challenging problems in AI, with a focus on creating a superintelligent AI that can learn endlessly [21]. - He advocates for a new "Age of Experience" in AI, where systems learn from experiences through reinforcement learning, moving beyond reliance on human knowledge [24]. - Silver believes that achieving true superintelligence requires AI to learn from first principles, independent of human intuition and knowledge [25].
Hinton加入Scaling Law论战,他不站学生Ilya
量子位· 2026-01-01 02:13
Core Viewpoint - The article discusses the ongoing debate surrounding the "Scaling Law" in AI, highlighting contrasting perspectives from key figures in the field, particularly Ilya Sutskever and Geoffrey Hinton, regarding the future and limitations of scaling AI models [1][8][21]. Group 1: Perspectives on Scaling Law - Ilya Sutskever expresses skepticism about the continued effectiveness of Scaling Law, suggesting that merely increasing model size may not yield significant improvements in AI performance [23][40]. - Geoffrey Hinton, on the other hand, maintains that Scaling Laws are still valid but face challenges, particularly due to data scarcity, which he believes can be addressed by AI generating its own training data [10][21]. - Demis Hassabis, CEO of DeepMind, supports Hinton's view, emphasizing the importance of scaling for achieving advanced AI systems and the potential for self-evolving AI through data generation [15][19]. Group 2: The Debate on Data and Model Scaling - The article outlines the historical context of Scaling Law, which posits that increasing model parameters, training data, and computational resources leads to predictable improvements in AI performance [26][27]. - Recent discussions have shifted towards concerns about data limitations, with Ilya arguing that the era of pre-training is coming to an end due to diminishing returns from scaling [32][41]. - Yann LeCun also shares skepticism about the assumption that more data and computational power will automatically lead to smarter AI, indicating a broader questioning of the Scaling Law's applicability [46][48]. Group 3: Future Directions and Research Focus - The article suggests that while current paradigms may still yield significant economic and social impacts, achieving Artificial General Intelligence (AGI) or Artificial Superintelligence (ASI) will likely require further research breakthroughs [53]. - There is a consensus among leading researchers that while AGI is not a distant fantasy, the nature and speed of necessary breakthroughs remain uncertain [53].
四周2亿人围观,诺奖凭什么颁给他,都在这一个半小时里
3 6 Ke· 2025-12-29 11:45
Core Insights - The documentary "The Thinking Game" provides an in-depth look at the operations behind a general artificial intelligence (AGI) laboratory, showcasing the journey that led to groundbreaking projects like AlphaFold [4][5][34] - It emphasizes the transformative potential of AGI, suggesting that humanity is on the brink of creating a new form of intelligence that transcends biological limitations [5][7] Group 1: Background and Formation of DeepMind - Initially, the term "artificial intelligence" was taboo, leading to skepticism in academic circles [8] - Demis Hassabis and Shane Legg founded DeepMind after realizing traditional academic paths were insufficient for their ambitions, leading to a bold decision to create a company focused on AGI [10][13] - The early days of DeepMind were characterized by secrecy and a lack of public presence, as they pursued a vision that few investors understood [13][15] Group 2: Development of AI Capabilities - DeepMind's approach involved using games as a testing ground for AI, allowing the system to learn without predefined rules [17][19] - The AI's ability to learn and adapt was demonstrated through its performance in various Atari games, culminating in a moment where it surpassed human capabilities [21] - The development of AlphaGo marked a significant milestone, as it defeated human champions in Go, a game previously thought to be a domain of human intelligence [22][26] Group 3: Breakthroughs in Life Sciences - AlphaFold emerged as a solution to the complex problem of protein folding, a challenge that had stumped scientists for decades [34][36] - The model achieved unprecedented accuracy in predicting protein structures, leading to a major breakthrough in life sciences [39][40] - DeepMind's decision to make 200 million protein structures publicly available signifies a commitment to advancing scientific research [41] Group 4: Ethical Considerations and Future Implications - The rapid advancement of AI capabilities raises ethical questions about the implications of AGI, with researchers expressing concerns about the potential consequences of their work [43] - The documentary draws parallels between the development of AGI and historical events, suggesting that society must collectively decide how to handle the emergence of such technology [45] - The narrative concludes with a call for humanity to take responsibility for the future of AGI, emphasizing that it is a shared challenge that transcends individual interests [45]
辛顿高徒压轴,谷歌最新颠覆性论文:AGI不是神,只是「一家公司」
3 6 Ke· 2025-12-22 08:13
Core Viewpoint - Google DeepMind challenges the traditional notion of Artificial General Intelligence (AGI) as a singular, omnipotent entity, proposing instead that AGI may emerge from a distributed network of specialized agents, termed "Patchwork AGI" [5][15][16]. Group 1: Concept of AGI - The prevailing narrative of AGI as a singular, all-knowing "super brain" is deeply rooted in science fiction and early AI research, leading to a focus on controlling this hypothetical entity [3][5]. - DeepMind's paper, "Distributed AGI Safety," argues that the assumption of a singular AGI is fundamentally flawed and overlooks the potential for intelligence to emerge from complex, distributed systems [5][8]. Group 2: Patchwork AGI - Patchwork AGI suggests that human society's strength comes from diverse roles and collaboration, similar to how AI could function through a network of specialized models rather than a single omnipotent model [15][16]. - This model is economically advantageous, as training multiple specialized models is more cost-effective than developing a single, all-encompassing model [16][19]. Group 3: Economic and Social Implications - The emergence of AGI may not be gradual but could occur suddenly when numerous specialized agents connect seamlessly, leading to a collective intelligence that surpasses human oversight [26][27]. - The paper emphasizes the need to shift focus from psychological alignment of a singular entity to sociological and economic stability of a network of agents [9][76]. Group 4: Risks and Challenges - Distributed systems introduce unique risks that differ from those associated with a singular AGI, including potential for collective "loss of control" rather than individual malice [30][31]. - The concept of "tacit collusion" among agents could lead to unintended consequences, such as price fixing or coordinated actions without explicit communication [31][38]. Group 5: Regulatory Framework - DeepMind proposes a multi-layered security framework to manage the interactions of distributed agents, emphasizing the need for a "virtual agent sandbox economy" to regulate their behavior [59][64]. - The framework includes mechanisms for monitoring agent interactions, ensuring baseline security, and integrating legal oversight to prevent monopolistic behaviors [67][70]. Group 6: Future Outlook - The paper serves as a call to action, highlighting the urgency of establishing robust infrastructure to manage the complexities of a distributed AGI landscape before it becomes a reality [70][78]. - It warns that if friction in AI connections is minimized, the resulting complexity could overwhelm existing safety measures, necessitating proactive governance [79].
AI被严重低估,AlphaGo缔造者罕见发声:2026年AI自主上岗8小时
3 6 Ke· 2025-11-04 12:11
Core Insights - The public's perception of AI is significantly lagging behind its actual advancements, with a gap of at least one generation [2][5][41] - AI is evolving at an exponential rate, with predictions indicating that by mid-2026, AI models could autonomously complete tasks for up to 8 hours, potentially surpassing human experts in various fields by 2027 [9][33][43] Group 1: AI Progress and Public Perception - Researchers have observed that AI can now independently complete complex tasks for several hours, contrary to the public's focus on its mistakes [2][5] - Julian Schrittwieser, a key figure in AI development, argues that the current public discourse underestimates AI's capabilities and progress [5][41] - The METR study indicates that AI models are achieving a 50% success rate in software engineering tasks lasting about one hour, with an exponential growth trend observed every seven months [6][9] Group 2: Cross-Industry Evaluation - The OpenAI GDPval study assessed AI performance across 44 professions and 9 industries, revealing that AI models are nearing human-level performance [12][20] - Claude Opus 4.1 has shown superior performance compared to GPT-5 in various tasks, indicating that AI is not just a theoretical concept but is increasingly applicable in real-world scenarios [19][20] - The evaluation results suggest that AI is approaching the average level of human experts, with implications for various sectors including law, finance, and healthcare [20][25] Group 3: Future Predictions and Implications - By the end of 2026, it is anticipated that AI models will perform at the level of human experts in multiple industry tasks, with the potential to frequently exceed expert performance in specific areas by 2027 [33][39] - The envisioned future includes a collaborative environment where humans work alongside AI, enhancing productivity significantly rather than leading to mass unemployment [36][39] - The potential transformation of industries due to AI advancements is profound, with the possibility of AI becoming a powerful tool rather than a competitor [39][40]
马斯克刚关注了这份AI报告
Sou Hu Cai Jing· 2025-09-19 04:35
Core Insights - The report commissioned by Google DeepMind predicts that by 2030, the cost of AI compute clusters will exceed $100 billion, capable of supporting training tasks equivalent to running the largest AI compute cluster continuously for 3,000 years [3][5] - AI model training is expected to consume power at a gigawatt level, with the computational requirements reaching thousands of times that of GPT-4 [3][5] - Despite concerns about potential bottlenecks in scaling, recent AI models have shown significant progress in benchmark tests and revenue growth, indicating that the expansion trend is likely to continue [4][9] Cost and Revenue - The training costs for AI are projected to exceed $100 billion, with power consumption reaching several gigawatts [5] - Revenue growth for companies like OpenAI, Anthropic, and Google DeepMind is expected to exceed 90% in the second half of 2024, with annualized growth rates projected to be over three times [9] - If AI developers' revenues continue to grow as predicted, they will be able to match the required investments of over $100 billion by 2030 [19] Data Availability - The report suggests that publicly available text data will last until 2027, after which synthetic data will fill the gap [5][12] - The emergence of synthetic data has been validated through models like AlphaZero and AlphaProof, which achieved expert-level performance through self-generated data [15] Algorithm Efficiency - There is an ongoing improvement in algorithm efficiency alongside increasing computational power, with no current evidence suggesting a sudden acceleration in algorithmic advancements [20] - The report indicates that even if there is a shift towards more efficient algorithms, it may further increase the demand for computational resources [20] Computational Distribution - The report states that the computational resources for training and inference are currently comparable and should expand synchronously [24] - Even with a potential shift towards inference tasks, the growth in inference scale is unlikely to hinder the development of training processes [27] Scientific Advancements - By 2030, AI is expected to assist in complex scientific tasks across various fields, including software engineering, mathematics, molecular biology, and weather forecasting [27][30][31][33][34] - AI will likely become a research assistant, aiding in formalizing proofs and answering complex biological questions, with significant advancements anticipated in protein-ligand interactions and weather prediction accuracy [33][34]
X @Demis Hassabis
Demis Hassabis· 2025-08-04 18:26
AI & Games - Games serve as a valuable testing environment for AI development, including the company's work on AlphaGo & AlphaZero [1] - The company anticipates rapid advancements in AI through the addition of more games and challenges to the Arena [1]