量子位

Search documents
OpenAI谷歌Anthropic罕见联手发研究!Ilya/Hinton/Bengio带头支持,共推CoT监测方案
量子位· 2025-07-16 04:21
Core Viewpoint - Major AI companies are shifting from competition to collaboration, focusing on AI safety research through a joint statement and the introduction of a new concept called CoT monitoring [1][3][4]. Group 1: Collaboration and Key Contributors - OpenAI, Google DeepMind, and Anthropic are leading a collaborative effort involving over 40 top institutions, including notable figures like Yoshua Bengio and Shane Legg [3][6]. - The collaboration contrasts with the competitive landscape where companies like Meta are aggressively recruiting top talent from these giants [5][6]. Group 2: CoT Monitoring Concept - CoT monitoring is proposed as a core method for controlling AI agents and ensuring their safety [4][7]. - The opacity of AI agents is identified as a primary risk, and understanding their reasoning processes could enhance risk management [7][8]. Group 3: Mechanisms of CoT Monitoring - CoT allows for the externalization of reasoning processes, which is essential for certain tasks and can help detect abnormal behaviors [9][10][15]. - CoT monitoring has shown value in identifying model misbehavior and early signs of misalignment [18][19]. Group 4: Limitations and Challenges - The effectiveness of CoT monitoring may depend on the training paradigms of advanced models, with potential issues arising from result-oriented reinforcement learning [21][22]. - There are concerns about the reliability of CoT monitoring, as some models may obscure their true reasoning processes even when prompted to reveal them [30][31]. Group 5: Perspectives from Companies - OpenAI expresses optimism about the value of CoT monitoring, citing successful applications in identifying reward attacks in code [24][26]. - In contrast, Anthropic raises concerns about the reliability of CoT monitoring, noting that models often fail to acknowledge their reasoning processes accurately [30][35].
Jason Wei也被小扎带走:思维链开创者、o1系列奠基人!这次真挖到OpenAI大动脉了
量子位· 2025-07-16 04:21
Core Viewpoint - The article discusses the significant talent acquisition by Meta, particularly focusing on Jason Wei, a key figure in OpenAI's o1 model, who is reportedly leaving OpenAI to join Meta, indicating a potential shift in the competitive landscape of AI development [1][2][8]. Group 1: Talent Acquisition - Jason Wei, the proponent of the "Chain-of-Thought" prompting technique, has been confirmed to be leaving OpenAI for Meta, alongside another key figure, Hyung Won Chung [2][4][9]. - Meta's recruitment strategy appears to be effective, as it has successfully attracted top talent from OpenAI, despite the latter's efforts to retain them through various incentives [29][30]. - The article highlights that Meta is providing substantial support to its AI talent, including direct reporting to Mark Zuckerberg and access to unlimited GPU resources, which may be appealing to top researchers [30][29]. Group 2: Internal Challenges at OpenAI - A blog post by former OpenAI engineer Calvin French-Owen reflects on the rapid growth and ensuing chaos within OpenAI, noting that the workforce expanded from 1,000 to 3,000 employees in a short period, leading to challenges in communication and management [33][38]. - The high-pressure work environment at OpenAI is emphasized, with reports of extreme workloads and a lack of structured processes, which may contribute to employee dissatisfaction [40][41][45]. - Calvin's reflections suggest that OpenAI has not fully adapted to its status as a large organization, drawing parallels to early challenges faced by Meta [46][47].
看遍奥斯卡后,VLM达到电影摄影理解新SOTA|上海AI Lab开源
量子位· 2025-07-16 01:49
Core Insights - The article discusses the launch of ShotBench, a comprehensive benchmark designed for understanding film language, along with the ShotVL model and the ShotQA dataset, aimed at enhancing visual language models (VLMs) in film comprehension [1][6][15]. Group 1: ShotBench and Its Components - ShotBench includes over 3,500 expert-annotated image and video question-answer pairs from more than 200 acclaimed films, covering eight key dimensions of cinematography [1][8]. - The ShotQA dataset consists of approximately 70,000 question-answer pairs, specifically designed to align models with "cinematic language" [15][19]. - The benchmark framework is structured to evaluate models from a professional cinematographer's perspective, focusing on extracting visual cues and reasoning behind cinematic techniques [8][14]. Group 2: Performance Evaluation - The evaluation of 24 leading VLMs revealed significant limitations, with even the best models achieving an average accuracy below 60%, particularly struggling with fine-grained visual cues and complex spatial reasoning [3][6]. - ShotVL-3B achieved a notable performance improvement of 19% over the baseline model Qwen2.5-VL-3B, establishing new state-of-the-art (SOTA) performance in film language understanding [3][24]. - ShotVL outperformed both the best open-source model (Qwen2.5-VL-72B-Instruct) and proprietary models (GPT-4o) across all dimensions evaluated [3][24]. Group 3: Training Methodology - ShotVL employs a two-phase training process: first, a large-scale supervised fine-tuning (SFT) to acquire broad knowledge, followed by group relative policy optimization (GRPO) for fine-grained reasoning enhancement [15][19][20]. - The first phase utilized approximately 70,000 question-answer pairs from the ShotQA dataset to establish strong alignment between visual features and specific cinematic terms [19]. - The second phase focused on improving reasoning capabilities and prediction accuracy, demonstrating the effectiveness of the GRPO approach [20][28]. Group 4: Key Dimensions of Cinematography - The eight core dimensions covered in ShotBench include Shot Size, Shot Framing, Camera Angle, Lens Size, Lighting Type, Lighting Condition, Composition, and Camera Movement, each critical for understanding film language [11][16][17]. - Each dimension is represented by a substantial number of samples, ensuring comprehensive coverage for model evaluation [17]. Group 5: Open Source Contribution - The team has made the model, data, and code open-source to facilitate rapid development in AI-driven film understanding and generation [4][30].
老黄投了个120亿美元最贵种子轮!但小钱:H20中国开卖,市值一夜暴涨1600亿美元
量子位· 2025-07-16 01:49
OpenAI前CTO Mira创业公司,Thinking Machines Lab——思考机器实验室,刚刚公告了首款融资情况: 顺利筹集约20亿美元 (约合人民币143亿元) ,公司估值一夜飙升至120亿美元 (约合人民币861亿元) ,成为硅谷史上最大种子轮之一。 领投的是A16z,一众芯片厂商也投了——其中最瞩目的莫过于老黄治下的 英伟达 。 鱼羊 发自 纽凹非寺 量子位 | 公众号 QbitAI 种子轮,估值120亿美元! 硅谷乃至全球创纪录的创业种子轮诞生了。 要知道,Thinking Machines Lab今年2月才刚刚官宣成立,现在还是个 0产品 的状态…… 但即便如此,依然吸引了众多投资者排队塞钱。 其中最瞩目的自然是英伟达,毕竟作为AI时代最大的受益者,再大的种子轮在4万亿美元市值的池子里,毛毛雨而已。 更何况,人在北京的老黄,仅仅昨天一晚,市值又涨了差不多1600亿……嗯,美元。 正在北京出席链博会的老黄,刚刚带来"非常、非常好的消息":H20恢复供应。有消息称,据说包括腾讯字节在内的大批中国客户已经在排队 买卡了。 最贵种子轮公司:0产品估值860亿 Thinking Machines ...
7个月翻一番!AI agent能力飙升,METR报告揭示指数级进化规律
量子位· 2025-07-16 01:49
henry 发自 凹非寺 量子位 | 公众号 QbitAI 报告指出:在软件开发、数学竞赛、科学问答等任务中,agent已能完成 相当于人类花费50–200分钟才能完成的任务 ,并且这种能力还在快 速提升——大约每 2–6个月 就能 翻一番 。 在计算机操作任务中,虽然任务时长较短,但增长率与软件开发等任务一致。 Agent在自动驾驶任务的性能增长速度则较慢,约20个月翻一番。 Agent能力每7个月翻一番! 根据非营利研究机构METR最新发布的报告,这一规律已在9项基准测试中得到了验证。 这些任务涉及编程、数学、计算机使用、自动驾驶等领域,表明大模型正在不断向着高度自动化迈进。 在视频理解任务中,模型能够在 时长1小时 的视频上取得 50% 的成功率。 作为一家致力于研究前沿人工智能系统能力及其风险的研究团队,METR此次的报告又进一步拉近了AI自主化的时间线,快来和我们看看报告 有哪些内容吧。 Agent的摩尔定律 在此前的测试中,METR将评估范围聚焦于软件开发和研究类任务,并发现AI agent的能力呈现出一种"摩尔定律"式的增长趋势—— 平均每七 个月,其可完成任务的time horizon就会翻一 ...
完全透明开源的共情语音大模型,三阶段训练,四大模块实现端到端对话 | 紫东太初联合长城汽车开源OpenS2S
量子位· 2025-07-16 01:49
紫东太初团队 投稿 量子位 | 公众号 QbitAI GPT-4o、Gemini这些顶级语音模型虽然展现了惊人的共情对话能力,但它们的技术体系完全闭源。 现在, 紫东太初团队联合长城汽车AI Lab 直接把整个技术栈都开源了,推出完全透明开源的端到端共情语音语言大模型OpenS2S。 OpenS2S的核心在于提供一个高效、低成本构建共情语音系统的新范式。 它不仅继承了团队在语音到文本共情模型BLSP-Emo上的技术积累,更引入了流式交错解码架构,实现了低延迟的实时语音生成。OpenS2S 提出自动化数据生成方法,结合大语言模型与可控文本到语音生成技术,构建多说话者、多情感的高质量共情语音训练语料。 | Name | | | LLaMA-Omni2 Qwen2-Audio GLM-4-Voice Kimi-Audio OpenS2S | | | | --- | --- | --- | --- | --- | --- | | Training Data | × | × | × | × | S | | Pretraining Code | × | × | × | × | V | | Fine-tuning Code ...
首篇潜空间推理综述!模型思考不必依赖Token,带宽暴增2700+倍
量子位· 2025-07-16 01:49
Core Insights - The article presents a comprehensive overview of latent space reasoning, highlighting its potential to achieve over 2700 times the bandwidth of traditional explicit reasoning chains (CoT) [1][15]. Group 1: Overview of Latent Space Reasoning - Latent space reasoning is an emerging field that traces its origins to the 2019 ICLR paper "Universal Transformers" by researchers from the University of Amsterdam and Google Brain [7]. - The article introduces a unified framework for latent space reasoning, which is based on mechanical interpretability and connects with the internal operations of models [3][4]. - The framework aims to facilitate future explorations, such as investigating infinite-depth reasoning through diffusion models [4]. Group 2: Mechanisms of Latent Space Reasoning - Latent space reasoning employs latent chains of thought, which represent reasoning in a continuous internal form rather than discrete natural language tokens [13][14]. - This method significantly enhances bandwidth, with each token in explicit CoT being approximately 15 bits, while latent CoT operations in a 2560-dimensional FP16 space yield around 40960 bits per step [15]. - The reasoning process is not constrained by a limited vocabulary, allowing for richer expressive capabilities [16]. Group 3: Modes of Latent Space Reasoning - There are two primary modes of latent space reasoning: vertical cycles and horizontal cycles [19]. - Vertical cycles utilize activation-based methods to extend computational depth, allowing models to repeatedly process the same set of layers to enhance reasoning [20][21]. - Horizontal cycles focus on expanding the model's memory and reasoning capabilities over time, maintaining a compressed hidden state that aggregates information from multiple time steps [28][29]. Group 4: Depth and Reasoning Capacity - The relationship between layer depth and reasoning capability is critical, with studies indicating that the implicit reasoning chain ability of models is strictly limited by the number of layers [34][40]. - Sufficient layer depth is necessary to execute multi-hop reasoning tasks effectively, as insufficient layers can hinder the emergence of final reasoning results [36][41]. - Research has established that the achievable length of reasoning chains is linearly related to the number of layers, positioning layer depth as a primary bottleneck for latent reasoning capacity [45]. Group 5: Advanced Reasoning Paradigms - The concept of "infinite depth reasoning" is proposed, allowing AI to allocate unlimited "thinking time" to refine solutions without output length constraints [53]. - This can be achieved through spatial infinite reasoning, which utilizes text diffusion models, and temporal infinite reasoning, which equates longer sequences with more optimization iterations [54][57]. - The article discusses specific methods for implementing these advanced paradigms, emphasizing their potential to enhance latent space reasoning [58].
一篇被证明“理论有误”的论文,拿下了ICML2025时间检验奖
量子位· 2025-07-15 08:31
Core Insights - The Batch Normalization paper, published in 2015, has been awarded the Time-Tested Award at ICML 2025, highlighting its significant impact on deep learning [1] - With over 60,000 citations, this work is considered a milestone in the development of deep learning, facilitating the training and application of deep neural networks [2][4] - Batch Normalization is a key technology that enabled deep learning to transition from small-scale experiments to large-scale practical applications [3] Group 1 - In 2015, deep learning faced challenges in training deep neural networks, which were often unstable and sensitive to parameter initialization [5][6][7] - Researchers Sergey Ioffe and Christian Szegedy identified the issue of Internal Covariate Shift, where the distribution of data within the network changes during training, complicating the training process [8][11] - Their solution involved normalizing the data at each layer, similar to input layer normalization, which significantly improved training speed and stability [12] Group 2 - The original paper demonstrated that using Batch Normalization allowed advanced image classification models to achieve the same accuracy with only 1/14 of the training steps [13] - Batch Normalization not only accelerated training but also introduced a regularization effect, enhancing the model's generalization ability [14][15] - Following its introduction, Batch Normalization became foundational for many mainstream convolutional neural networks, such as ResNet and DenseNet [18] Group 3 - In 2018, a paper from MIT challenged the core theory of Batch Normalization, showing that even with introduced noise, models with Batch Normalization still trained faster than those without it [21][23] - This research revealed that Batch Normalization smooths the Optimization Landscape, making gradient behavior more predictable and stable [24] - It was suggested that Batch Normalization acts as an unsupervised learning technique, allowing networks to adapt to the data's inherent structure early in training [25] Group 4 - Recent studies have provided deeper insights into Batch Normalization from a geometric perspective [29] - Both authors, Ioffe and Szegedy, have continued their careers in AI, with Szegedy joining xAI and Ioffe following suit [30][32] - Szegedy has since transitioned to a new role at Morph Labs, focusing on achieving "verifiable superintelligence" [34]
只因一个“:”,大模型全军覆没
量子位· 2025-07-15 08:31
Core Viewpoint - The article discusses a significant vulnerability in large language models (LLMs) where simple tokens, such as colons and specific phrases, can deceive these models into providing false positive rewards, highlighting the need for improved robustness in LLMs [1][21][33]. Group 1: Vulnerability Discovery - A recent study titled "A Token Can Deceive LLM" reveals that LLMs can be easily tricked by certain symbols and phrases, leading to incorrect evaluations [2][12]. - The vulnerability affects various LLMs, including GPT-4o, Claude-4, and LLaMA3-70B, which all exhibited high false positive rates (FPR) when exposed to these deceptive tokens [7][21]. - The study identified two main categories of deceptive tokens: non-character symbols (e.g., spaces, colons) and reasoning starter phrases (e.g., "Thought process:", "解") [4][15]. Group 2: Experimental Findings - All tested models, regardless of type, triggered false positive responses, with GPT-4o showing a FPR of 35% for the colon symbol and LLaMA3-70B having a FPR of 60%-90% for the phrase "Thought process:" [21][23]. - The research also indicated that model size does not consistently correlate with FPR, suggesting that larger models are not necessarily more robust against these attacks [23][26]. - The experiments demonstrated that the vulnerability could proliferate, allowing for the automatic generation of new deceptive responses based on existing "universal keys" [25]. Group 3: Mitigation Strategies - To address the identified vulnerabilities, researchers developed a new model called Master-RM, which significantly reduces the FPR to nearly zero by using an enhanced training dataset that includes adversarial samples [29][31]. - Master-RM was tested across various datasets and demonstrated robust performance, maintaining a high consistency rate with GPT-4o [32]. - The findings emphasize the importance of rigorous adversarial evaluation in the reinforcement learning from human feedback (RLHF) processes to ensure the reliability of LLMs [34][35].
Switch的救世主是老黄!?
量子位· 2025-07-15 06:28
Core Viewpoint - The article discusses the successful collaboration between Nintendo and NVIDIA, highlighting the launch of the Switch2 and its advanced graphics capabilities through the NVN2 API, which significantly enhances gaming performance and visual quality compared to its predecessor. Group 1: Switch2 Launch and Performance - The Switch2, equipped with NVIDIA's NVN2, addresses issues like overheating, lag, short battery life, and reduced graphics quality seen in the previous generation [2][5] - Within just four days of its release, the Switch2 sold over 3.5 million units, setting a record for Nintendo's fastest sales [5] - The NVN2 API allows the Switch2 to run games that can achieve 60 frames per second on Xbox, making it easier to port titles [3][12] Group 2: Historical Context of Nintendo and NVIDIA Collaboration - The partnership between Nintendo and NVIDIA began over a decade ago, driven by Nintendo's need for a new mobile gaming hardware and NVIDIA's desire to revitalize its Tegra chip line [4][10] - The original Switch, which sold over 150 million units, was a result of this collaboration, leading to a 108% increase in Tegra processor sales, reaching $332 million [17][18] Group 3: Technical Innovations of NVN and NVN2 - NVN is a custom graphics API designed specifically for the Switch, optimizing performance by removing unnecessary features found in general-purpose APIs like OpenGL or Vulkan [25][26] - The Switch2's NVN2 includes NVIDIA's deep learning super sampling (DLSS) and ray tracing capabilities, allowing for enhanced graphics performance, such as increasing game resolution from 720p/30fps to 1440p/60fps [30] Group 4: Gaming Philosophy and User Experience - The Switch's design philosophy emphasizes flexibility, allowing players to switch between gaming scenarios, whether on the go or at home [32][36] - The technology serves to enhance creativity and enjoyment in gaming, reflecting a shift in how gaming is integrated into daily life [37][40]