Workflow
结构化推理
icon
Search documents
思维链可无限延伸了,MIT等打破大模型上下文天花板
量子位· 2025-08-20 01:13
Core Viewpoint - The article discusses the introduction of the Thread Inference Model (TIM) by MIT and its associated inference engine TIMRUN, which allows for nearly unlimited long-range reasoning in large models by overcoming the physical limitations of context windows [2][3][5]. Group 1: TIM Architecture - TIM transforms the reasoning process into a recursive task tree structure, enabling the model to handle complex problems by breaking them down into simpler subtasks [11][12]. - Each task unit in TIM consists of four key components: thought, tool use, subtasks, and conclusion [12]. - The model employs a pruning mechanism to discard unnecessary subtasks after completion, which can reduce KV cache usage by over 50% [13]. Group 2: TIMRUN Engine - TIMRUN is designed to address the challenges of deploying TIM, particularly in managing GPU memory and position encoding for "infinite" reasoning [15][16]. - The engine utilizes dynamic memory management and reuses position encoding, allowing for continuous content generation within fixed output window limits [18]. - TIMRUN can initiate tool calls internally during runtime, significantly reducing the token cost complexity from O(n²) to O(n) [22]. Group 3: Performance Metrics - In benchmark tests, the TIM-8b model achieved a 69% accuracy rate on the MATH500 task and 46.7% on the more challenging AIME 2024 task, demonstrating that pruning does not compromise performance [26]. - TIM's performance in multi-hop reasoning tasks was validated, achieving a 67.9% accuracy rate on the Datacommons QA benchmark, comparable to methods requiring extensive token prompts [27]. - Efficiency tests showed that TIMRUN improved throughput by approximately 20% compared to baseline systems, maintaining stability even with increased tool calls [29][30].
KAG-Thinker:「结构化」思考新范式,支持逻辑严谨的大模型复杂推理
机器之心· 2025-07-08 06:54
Core Viewpoint - The article discusses the release of the KAG-Thinker model by Ant Group's Knowledge Engine team in collaboration with Zhejiang University and Tongji University, focusing on structured reasoning for complex tasks, enhancing logical consistency and stability in reasoning processes. Group 1: Model Development and Features - KAG-Thinker is an important upgrade of the KAG framework, designed to construct a stable and interpretable reasoning paradigm for complex tasks in both general and specialized fields [1][3] - The model utilizes a dual semantic representation mechanism of natural language and logical functions to better leverage structured knowledge [3] - It combines breadth splitting and depth solving to improve the rigor of problem-solving, introducing a knowledge boundary determination mechanism centered on knowledge point alignment [3][10] Group 2: Performance and Evaluation - Experimental results show that KAG-Thinker outperforms state-of-the-art deep search methods by an average of 4.1% across seven single-hop and multi-hop reasoning datasets [6][24] - In single-hop datasets, KAG-Thinker achieved an average improvement of 4.5%, while in multi-hop datasets, the improvement was 3.9% [25] - The model demonstrated effectiveness in specialized fields, particularly in medical question-answering tasks, indicating its potential for fine-tuning in other professional domains [6][39] Group 3: Framework Integration and Stability - The KAG framework version 0.8 enhances knowledge base capabilities, supporting structured and unstructured data integration, and allows developers to customize indexing [28][29] - KAG-Thinker, integrated with the KAG framework, shows an average performance improvement of 3.0% in EM and 3.8% in F1 metrics compared to the standalone Thinker model [31] - Stability tests indicate that KAG-Thinker 7B outperforms previous versions in terms of consistent problem decomposition, achieving an average improvement of 17.9% and 7.6% under common temperature parameters [33]