Workflow
OpenChat
icon
Search documents
又是王冠:27M小模型超越o3-mini!拒绝马斯克的00后果然不同
Sou Hu Cai Jing· 2025-08-10 04:21
Core Insights - The article discusses the development of a new AI model called the Hierarchical Reasoning Model (HRM) by Sapient Intelligence, which has achieved superior performance compared to larger models with fewer parameters [3][5][18] - HRM utilizes innovative techniques inspired by brain functions, allowing it to perform complex reasoning tasks efficiently without relying on traditional pre-training methods [4][12][14] Model Performance - HRM, with only 27 million parameters, surpassed larger models like o3-mini-high and Claude 3.7 in various tests, achieving a 40.3% accuracy rate in the ARC-AGI challenge [16][18] - In extreme Sudoku tasks, HRM demonstrated near-perfect accuracy, while traditional models struggled significantly [16][18] Technical Innovations - HRM employs a dual-layer cyclical module design that mimics the brain's hierarchical processing and time-scale separation, enhancing both global direction and local execution efficiency [4][7] - The model incorporates a layered convergence mechanism to avoid premature convergence, allowing it to adaptively set new goals based on high-level updates [9][11] - It utilizes approximate gradient techniques to optimize memory usage and computational efficiency, aligning with biological learning patterns [12] - A deep supervision mechanism is integrated, allowing for periodic evaluations and adjustments during the learning process, which helps in correcting deviations promptly [13][14] Developer Background - The model's creator, Wang Guan, is a young entrepreneur who previously declined offers from major tech figures like Elon Musk, aiming instead to revolutionize AI architecture [20][22] - Wang co-founded Sapient Intelligence in 2024, focusing on developing models with advanced reasoning and planning capabilities [22]
又是王冠:27M小模型超越o3-mini!拒绝马斯克的00后果然不同
量子位· 2025-08-10 04:11
Core Viewpoint - The article discusses the breakthrough of the Hierarchical Reasoning Model (HRM) developed by Sapient Intelligence, which outperforms larger models like o3-mini-high and Claude 3.7 with only 27 million parameters and 1,000 training samples, showcasing its potential in AI reasoning capabilities [4][6][27]. Group 1: Model Performance - HRM achieves a 40.3% accuracy rate in the ARC-AGI challenge, surpassing larger models such as o3-mini-high (34.5%) and Claude 3.7 (21.2%) [27]. - In extreme Sudoku tasks, HRM demonstrates near-perfect accuracy, while existing models struggle, achieving 0% accuracy [27]. - For complex maze navigation, HRM maintains stable performance, while a 175 million parameter Transformer model shows less than 20% accuracy [29]. Group 2: Model Design and Technology - HRM is inspired by the brain's hierarchical processing and time-scale separation, featuring two complementary loop modules for abstract planning and detail computation [10][12]. - The model employs a hierarchical convergence mechanism to avoid premature convergence, allowing continuous task progression [16][18]. - HRM utilizes approximate gradient techniques to enhance efficiency, requiring less memory and aligning with biological learning patterns [19]. - A deep supervision mechanism is introduced, allowing for periodic assessments and adjustments during the learning process [20][22]. - Adaptive computation time enables HRM to allocate thinking time flexibly, optimizing performance based on task complexity [23][24]. Group 3: Developer Background - The founder of Sapient Intelligence, Wang Guan, is a Tsinghua University alumnus who declined offers from major tech figures like Elon Musk to focus on challenging the Transformer architecture [2][33][37]. - Wang has a strong background in programming and has independently developed the OpenChat project, which contributed to his recognition in the AI community [34][35].
马斯克挖不动的清华学霸,一年造出 “反内卷 AI”!0.27B参数硬刚思维链模型,推理完爆o3-mini-high
AI前线· 2025-08-04 06:43
Core Viewpoint - The article discusses the launch of a new AI model named HRM by Sapient Intelligence, which, despite its smaller parameter size of 27 million, demonstrates superior reasoning capabilities compared to larger models like ChatGPT and Claude 3.5, particularly in complex reasoning tasks [2][7]. Group 1: Model Performance and Comparison - HRM outperformed advanced chain-of-thought models in complex reasoning tasks, achieving near-perfect accuracy with only 1,000 training samples, while traditional models failed completely in tests like "extreme Sudoku" and "high-difficulty mazes" [6][7]. - In the ARC-AGI benchmark test, HRM scored 40.3%, surpassing larger models such as o3-mini-high (34.5%) and Claude 3.7 Sonnet (21.2%) [7]. Group 2: Model Architecture and Innovation - HRM's architecture is inspired by human brain functions, utilizing a dual recursive module system that allows for both slow, abstract planning and fast, detailed calculations, thus enabling deep reasoning without extensive data [11][14]. - The model employs "implicit reasoning," which avoids the limitations of traditional token-based reasoning, allowing for more efficient processing and reduced reliance on large datasets [13][16]. Group 3: Economic and Practical Implications - The efficiency of HRM translates to significant economic benefits, with the potential to complete tasks 100 times faster than traditional models, making it suitable for applications in environments with limited data and resources [18][19]. - Initial successes in fields such as healthcare, climate prediction, and robotics indicate the model's versatility and potential for broader applications beyond text-based systems [19].
X @Token Terminal 📊
Token Terminal 📊· 2025-07-02 16:32
RT Fabio (@Zero2HeroZombie)In June, @OpenChat has added 13.7K new Users (12.9K registered with POH) 📈Hopefully, the number of new users in the $ICP Ecosystem should rapidly increase after the release of ☕️Data: @tokenterminal https://t.co/Y2PSRx8hv1 ...