Workflow
Hierarchical Reasoning Model (HRM)
icon
Search documents
三星 TRM 论文:少即是多,用递归替代深度,挑战 Transformer 范式
3 6 Ke· 2025-11-03 12:51
Core Insights - The study reveals that smaller networks can outperform large language models in reasoning tasks, with the Tiny Recursive Model (TRM) achieving superior results using only 7 million parameters and a two-layer neural network [4][11] - TRM's architecture eliminates self-attention layers, demonstrating that for small-scale fixed input tasks, a multi-layer perceptron (MLP) can reduce overfitting [4][11] - The model's recursive updating mechanism allows for multi-round self-correction, leading to higher accuracy compared to traditional large models that rely on chain-of-thought reasoning [6][12] Model Performance - TRM achieved an accuracy of 45% on ARC-AGI-1 and 8% on ARC-AGI-2, outperforming most large models [6] - In the Sudoku-Extreme task, TRM set a record with an accuracy of 87.4%, while in the Maze-Hard task, it reached 85.3% accuracy [6][14] - The model's performance is significantly better than the Hierarchical Reasoning Model (HRM), with TRM's accuracy in Maze-Hard being 10 percentage points higher [8][12] Design Philosophy - TRM's design is a reflection on the previous HRM model, simplifying the recursive updating process by using a single network with deep supervision [8][12] - The model reduces parameters by approximately 74% compared to HRM and halves the number of forward passes while improving accuracy [8][12] - TRM replaces the traditional adaptive computational time mechanism with a simple binary decision to stop reasoning, enhancing training speed without sacrificing accuracy [9][10] Implications for AI Development - The findings challenge the notion that larger models are inherently better, suggesting that depth of recursive thinking can replace the need for increased model size [11][14] - The research indicates a new direction for lightweight AI reasoning, particularly beneficial for edge AI and low-resource applications [14] - The study emphasizes that intelligent depth may stem from repeated thinking rather than sheer scale, providing insights for future AI model development [14]
又是王冠:27M小模型超越o3-mini!拒绝马斯克的00后果然不同
Sou Hu Cai Jing· 2025-08-10 04:21
Core Insights - The article discusses the development of a new AI model called the Hierarchical Reasoning Model (HRM) by Sapient Intelligence, which has achieved superior performance compared to larger models with fewer parameters [3][5][18] - HRM utilizes innovative techniques inspired by brain functions, allowing it to perform complex reasoning tasks efficiently without relying on traditional pre-training methods [4][12][14] Model Performance - HRM, with only 27 million parameters, surpassed larger models like o3-mini-high and Claude 3.7 in various tests, achieving a 40.3% accuracy rate in the ARC-AGI challenge [16][18] - In extreme Sudoku tasks, HRM demonstrated near-perfect accuracy, while traditional models struggled significantly [16][18] Technical Innovations - HRM employs a dual-layer cyclical module design that mimics the brain's hierarchical processing and time-scale separation, enhancing both global direction and local execution efficiency [4][7] - The model incorporates a layered convergence mechanism to avoid premature convergence, allowing it to adaptively set new goals based on high-level updates [9][11] - It utilizes approximate gradient techniques to optimize memory usage and computational efficiency, aligning with biological learning patterns [12] - A deep supervision mechanism is integrated, allowing for periodic evaluations and adjustments during the learning process, which helps in correcting deviations promptly [13][14] Developer Background - The model's creator, Wang Guan, is a young entrepreneur who previously declined offers from major tech figures like Elon Musk, aiming instead to revolutionize AI architecture [20][22] - Wang co-founded Sapient Intelligence in 2024, focusing on developing models with advanced reasoning and planning capabilities [22]