Test-time Scaling
Search documents
CVPR 2026 Workshop征稿|从感知到推理,ViSCALE 2.0 邀你重塑计算机视觉的 System 2
机器之心· 2026-02-13 04:19
Core Insights - The article discusses the evolution of computer vision towards a new paradigm, emphasizing the transition from basic pixel perception to complex spatial reasoning and world modeling, facilitated by Test-time Scaling (TTS) [2][5] - The upcoming ViSCALE 2026 conference aims to gather leading scholars to explore breakthroughs in visual models through computational expansion, focusing on deep reasoning rather than mere static outputs [4][5] Group 1: Conference Highlights - ViSCALE 2026 will feature discussions on spatial intelligence and world models, with contributions from top scholars including Sergey Levine, Manling Li, and Ziwei Liu [5] - The conference encourages innovative research submissions that challenge existing visual model limitations, providing a platform for both theoretical and application-focused studies [7] Group 2: Key Topics of Discussion - The conference will cover various topics, including: - Enhancing video generation's physical consistency and long-term causal reasoning through TTS [6] - Breaking 2D limitations to enable models to navigate and operate in 3D spaces like humans [6] - Developing visual reasoning chains that allow models to self-correct and engage in multi-step reasoning [6] - Exploring scaling laws that relate computational load during testing to visual reasoning performance [6] Group 3: Submission Details - The conference invites submissions in two tracks: Full Papers (8 pages) and Extended Abstracts (up to 4 pages), with specific formatting requirements [9] - Important deadlines include submission by March 10, 2026, and notification of acceptance by March 18, 2026 [9]
更多非共识,Test-time Scaling 能否一直大力出奇迹?
机器之心· 2025-12-07 01:30
Group 1 - The article discusses the ongoing debate in the industry regarding Test-time Scaling (TTS) and its effectiveness in enhancing the performance of large language models (LLMs) [6][7]. - TTS has gained significant attention since Q3 2024, as it represents a crucial paradigm for improving LLM performance by dynamically allocating more computational resources during the inference phase [7][8]. - Various research institutions, including Google and UC Berkeley, have explored how increasing computational resources at test time can enhance LLM capabilities, leading to a focus on inference processes [8][9]. Group 2 - The article outlines four dimensions for systematically reviewing TTS methods: "What to scale," "How to scale," "Where to scale," and "How well to scale" [8][10]. - "What to scale" focuses on the objects of expansion, such as the length of the chain of thought (CoT), sample size, path depth, or internal states [9]. - "How to scale" examines the methods of expansion, including approaches like Prompt, Search, Reinforcement Learning (RL), or Mixture-of-Models [10]. Group 3 - The article highlights that the industry has developed a deeper understanding of TTS mechanisms and implementation methods over the past year, although there are still significant disagreements and reflections on improvement strategies [12]. - Research from Fudan University suggests that the popular "Sequential" approach of extending CoT does not consistently improve accuracy, proposing a "Parallel" method as a potential improvement [12][13]. - The "Parallel" method allows models to perform parallel reasoning to generate multiple inference paths, aggregating these paths to derive the final answer, thus enhancing the breadth of thought [13]. Group 4 - The article notes that as the industry continues to explore TTS, previously unrecognized limitations of certain approaches are being confirmed [14]. - There is a growing trend towards External (parallel, hybrid, etc.) TTS methods as Internal (Sequential) approaches approach their limits [14]. - The future of TTS may not lie solely in increased computational power but rather in smarter search techniques, indicating a shift in focus [14][15].