思维链(CoT)技术
Search documents
AAAI 2026 Oral | 大模型「爱你在心口难开」?深度隐藏认知让推理更可靠
机器之心· 2026-01-09 02:53
Core Insights - The article discusses the advancements in large language models (LLMs) in reasoning tasks, particularly emphasizing the Chain-of-Thought (CoT) technique, which enhances model performance by generating intermediate reasoning steps before arriving at a final answer [2][6] - A research team from Hefei University of Technology proposes that LLMs possess a "hidden cognition" that allows them to internally assess the correctness of their reasoning, even if this is not reflected in the token probabilities during generation [2][10] - The paper introduces a framework that enables models to score their reasoning steps based on this hidden cognition, thereby improving the reliability of CoT [2][10] Summary by Sections Introduction - The article highlights the growing application of LLMs in various reasoning tasks and the importance of maintaining stable and reliable reasoning quality throughout the generation process [6][8] - It identifies factors that can affect the reliability of reasoning chains, such as subtle biases in understanding, expression noise, and cumulative errors in long chains [6][8] Research Motivation - The research aims to determine if there are internal signals within the model that can reflect the reliability of current reasoning steps, potentially guiding the model to continue with more reliable paths [7][15] - The study focuses on two key questions regarding the existence of discernible signals in internal activations and the feasibility of constructing a mechanism to utilize these signals [8][15] Methodology and Innovations - The proposed method involves detecting "truth sensitivity" from multiple attention heads and training a simple probe on internal representations to assess which layers are most sensitive to reasoning correctness [10][11] - A confidence predictor is constructed using the most sensitive attention heads to output reliability scores for each reasoning step, based on deep internal representations rather than token probabilities [12][21] - The research introduces a confidence-guided search strategy that combines model generation probabilities with confidence scores to filter the most reliable reasoning paths [13][16] Experimental Results - The study evaluates the effectiveness of the confidence predictor and its application in guiding reasoning paths across various benchmarks, including both single-modal and multi-modal reasoning tasks [22][24] - Results indicate that the proposed method consistently outperforms baseline models, achieving significant improvements in reasoning accuracy across different datasets [23][24] - Ablation studies confirm the critical role of the confidence predictor in enhancing reasoning performance, with random selection of reasoning steps leading to a notable decline in effectiveness [25][27]
为大模型思考装上“猎鹰重装引擎” :腾讯混元 SEAT 重塑深度思考
AI科技大本营· 2025-07-15 11:30
Core Viewpoint - Tencent's Hunyuan team has introduced the SEAT adaptive parallel reasoning framework, transforming complex reasoning tasks from a "single-engine airship" into a "multi-engine rocket," enhancing the capabilities of large models in handling intricate reasoning challenges [7][44]. Group 1: SEAT Framework Overview - The SEAT framework integrates both sequential and parallel scaling paradigms, allowing for extensive exploration and deep refinement of reasoning processes [15][43]. - It employs a multi-round parallel reasoning approach, significantly enhancing the model's exploration capabilities by generating multiple independent reasoning paths simultaneously [16][20]. - The framework is designed to be plug-and-play, enabling easy integration with existing large language models without requiring additional training [29][44]. Group 2: Performance Enhancements - Initial experiments show that even with a minimal parallel setup (N=2), the SEAT framework can achieve a remarkable accuracy improvement of +14.1% for a 32B model and +24.5% for a 7B model [28]. - As the number of parallel paths increases (up to N=8), performance continues to improve, demonstrating the framework's powerful exploration capabilities [23]. Group 3: Semantic Entropy as Navigation - The SEAT framework introduces semantic entropy as a self-supervised metric to gauge the consistency of reasoning outputs, acting as a "navigation sensor" to determine when to stop computations [27][32]. - Two navigation strategies are implemented: a predefined threshold approach and an adaptive threshold-free mechanism, both aimed at optimizing the reasoning process [35][36]. Group 4: Safety Mechanisms - The SEAT framework includes a safety mechanism to prevent "semantic entropy collapse," which can lead to overconfidence and erroneous outputs in smaller models [38][40]. - By monitoring semantic entropy, the framework can issue stop commands before the model's performance deteriorates, ensuring stable reasoning outcomes [40][44].
只用2700万参数,这个推理模型超越了DeepSeek和Claude
机器之心· 2025-06-30 10:23
Core Insights - The article discusses the need for transformation in the architecture of large language models (LLMs), particularly focusing on the limitations of current chain-of-thought (CoT) techniques, which face challenges such as task complexity, high data requirements, and latency issues [2][4]. Group 1: Hierarchical Reasoning Model (HRM) - The Hierarchical Reasoning Model (HRM) is introduced as a novel cyclic architecture inspired by the human brain's layered and multi-timescale processing mechanisms, achieving high computational depth while maintaining training stability and efficiency [3][6]. - HRM operates through two interdependent cyclic modules: a high-level module for slow, abstract planning and a low-level module for fast, detailed computations, achieving remarkable performance on complex reasoning tasks with only 27 million parameters and 1,000 training samples [4][5]. - HRM does not require pre-training or CoT data, yet it performs nearly perfectly on challenging tasks such as complex Sudoku puzzles and optimal pathfinding in large mazes, outperforming larger models with longer context windows [5][6]. Group 2: Design and Mechanisms - The core design of HRM is based on hierarchical processing and time-scale separation, where high-level brain regions integrate information over longer time scales while low-level regions handle immediate sensory information [12][13]. - HRM incorporates feedback loops similar to the brain's dense recurrent neural network connections, enhancing representation accuracy and contextual adaptability while avoiding issues related to backpropagation through time (BPTT) [14][19]. - The model introduces approximate gradients and deep supervision, allowing for efficient memory usage and improved training dynamics, which contrasts with traditional methods that require extensive memory and time [20][23]. Group 3: Performance and Adaptability - HRM demonstrates hierarchical convergence, with the high-level module stabilizing while the low-level module converges repeatedly, leading to rapid convergence and minimal residuals compared to deep neural networks [17][36]. - The model features adaptive computation time (ACT), enabling it to dynamically adjust computational resources based on task complexity, thus optimizing performance without significant resource expenditure [25][27]. - HRM can seamlessly extend inference computation by adjusting parameters without the need for retraining or architectural changes, showcasing its flexibility in handling complex reasoning tasks [28][36]. Group 4: Experimental Results - Experimental results indicate that HRM excels in complex reasoning tasks, raising questions about the underlying reasoning algorithms it employs, which is crucial for enhancing model interpretability [31][39]. - Visualizations of HRM's reasoning processes reveal its strategies in maze and Sudoku tasks, demonstrating a combination of exploration and optimization techniques that resemble depth-first search methods [31][38]. - The hierarchical structure of HRM emerges as a natural characteristic during the learning of complex reasoning tasks, rather than being an inherent property of the model architecture [34].