预测与解释脱节
Search documents
哈佛&MIT:AI能预测,但它还解释不了“why”
3 6 Ke· 2025-10-22 00:56
Core Insights - The core question in the field of Artificial General Intelligence (AGI) is whether large language models (LLMs) can learn a "world model" or if they are merely playing a "next word prediction" game [1][2] - A recent experiment by Harvard and MIT tested LLMs using orbital mechanics to determine if they could derive the underlying laws of physics, specifically Newton's laws, from their predictions [2][4] - The results indicated a disconnection between prediction and explanation, as the AI models could accurately predict planetary trajectories but failed to encode the underlying physical laws [4][6] Experiment Design and Findings - The research utilized 10 million simulated solar system coordinate sequences (totaling 20 billion tokens) to train a small Transformer model [4] - The hypothesis was that if the model could make accurate predictions without understanding Newton's laws, it would not possess a complete "world model" [2][4] - The findings showed that while the AI could predict trajectories well, the derived force vectors were chaotic and unrelated to Newton's laws, indicating a lack of a stable guiding framework [6][8] Implications for AI Development - The inability of AI models to maintain consistent errors across different samples suggests they do not possess a stable world model, which is essential for scientific discovery [8][9] - The research highlights a fundamental limitation in current AI models, as they can achieve high accuracy in predictions but lack the capability to construct a reality-based world model [10][11] - Future AI development may require a combination of larger models and new methodologies to enhance understanding and generalization capabilities [12][13] Broader Context - The study reflects a classic scientific debate about whether the essence of science lies in precise predictions or in understanding the underlying reasons for phenomena [12][14] - The quest for AI to evolve from being merely a "prediction machine" to a "thinker" capable of understanding the logic of the world is crucial for its future impact on scientific discovery [14]
AI大家说 | 哈佛&MIT:AI能预测,但它还解释不了“why”
红杉汇· 2025-10-22 00:06
Core Insights - The article discusses a significant experiment conducted by Harvard and MIT to explore whether large language models (LLMs) can learn a "world model" or if they merely predict the next word based on probabilities [3][4][5] - The experiment utilized orbital mechanics as a testing ground, aiming to determine if AI could derive Newton's laws from its predictions of planetary motion [4][5] - The findings revealed that while AI models could accurately predict planetary trajectories, they did not encode the underlying physical laws, indicating a disconnect between prediction and explanation [6][10] Group 1: Experiment Design and Findings - The research team trained a small Transformer model on 10 million simulated solar system coordinates, totaling 20 billion tokens, to assess its ability to utilize Newton's laws for predicting planetary movements [8] - The results showed that the AI model could generate precise trajectory predictions but relied on specific situational heuristics rather than understanding the fundamental laws of physics [10][11] - The study also highlighted that the AI's predictions could not be generalized to untrained scenarios, demonstrating a lack of a stable world model [10][11] Group 2: Implications for AI Development - The research raises questions about the fundamental limitations of AI models, particularly regarding their ability to construct a coherent world model necessary for scientific discovery [11][12] - The article suggests that while LLMs are not entirely useless, they are currently insufficient for achieving scientific breakthroughs [13] - Future AI development may require a combination of larger models and new methodologies to enhance their understanding and predictive capabilities [13][14] Group 3: Philosophical Considerations - The article reflects on a classic scientific debate: whether the essence of science lies in precise predictions or in understanding the underlying reasons for phenomena [14] - It emphasizes the importance of developing AI that can not only predict but also comprehend the logic of the world, which will determine its ultimate impact on scientific history [14]