量子位

Search documents
3亿美元薪酬被10人拒绝!OpenAI首席研究官一句话引发硅谷史上最疯狂抢人大战
量子位· 2025-07-21 06:46
Core Viewpoint - The article discusses the intense competition between Meta and OpenAI for top AI talent, highlighting that many OpenAI employees have rejected lucrative offers from Meta, including a $300 million offer to Mark Chen, OpenAI's Chief Research Officer [2][3][4]. Group 1: Recruitment Efforts - At least 10 OpenAI employees have turned down offers from Meta, indicating a strong loyalty to their current employer [2][3]. - Mark Chen's conversation with Mark Zuckerberg led to a significant recruitment drive at Meta, with a focus on acquiring top AI talent [5][7][9]. - A list of 44 individuals targeted by Meta reveals that nearly 40% of them are from OpenAI, showcasing the aggressive recruitment strategy [10][54]. Group 2: Talent Composition - The recruitment list shows a notable preference for Chinese researchers, with 50% of the members being from China, and many being alumni of prestigious institutions like Tsinghua University and Peking University [13][14]. - Several notable hires include Chengxu Zhuang, Chenxi Liu, and Chunyuan Li, all of whom have impressive academic and professional backgrounds in AI [16][20][24]. Group 3: Competitive Landscape - Meta's recruitment strategy includes not only high salaries but also promises of unlimited computational resources, which is appealing to researchers focused on ambitious AI projects [55][57]. - OpenAI's response to this competition includes plans to deploy 1 million GPUs by the end of the year, aiming to match Meta's capabilities [60][62]. - The comparison of computational resources between OpenAI and Meta indicates a fierce race to build powerful AI models, with Meta planning to establish multiple Gigawatt-level supercomputing clusters [61][62].
蚂蚁ACL活动全览!论文串讲、人才专项答疑与闭门晚宴等你报名
量子位· 2025-07-21 04:23
⬇️点击阅读全文,预约活动席位 *本文系量子位获授权刊载,观点仅为原作者所有。 一键三连 「点赞」「转发」「小心心」 欢迎在评论区留下你的想法! — 完 — 点亮星标 科技前沿进展每日见 ...
手术刀式去噪突破LLM能力上限,从头预训练模型下游任务平均提高7.2% | 中科院&阿里
量子位· 2025-07-21 04:23
Core Viewpoint - The article discusses RefineX, a new framework developed by the Institute of Computing Technology, Chinese Academy of Sciences, and Alibaba Qwen, aimed at efficiently refining large-scale pre-training data through programmatic editing tasks, addressing noise pollution that affects data quality [1][2]. Group 1: Advantages of RefineX - RefineX distills high-quality end-to-end optimization results into a simplified deletion program based on editing operations, enhancing the efficiency of data refinement [2][11]. - The high-precision distillation process enables the training of an efficient and reliable refine model that systematically optimizes each instance in the corpus [3][12]. - While refining data efficiently, RefineX reliably preserves the diversity and naturalness of the original text [4][19]. Group 2: Performance Metrics - Training a 750M model with 20 billion tokens refined by RefineX achieved an average score of 44.7 across ten tasks, representing a 7.2% improvement over the original data [5][25]. - The model using 10 billion refined tokens outperformed those trained on 20 billion traditional filtered data, indicating that RefineX effectively reduces training token costs while allowing for more diverse text consideration [25]. Group 3: Data Quality Improvement - RefineX demonstrated a 42.2% improvement rate in the quality of low-quality content while maintaining a "zero new vocabulary" policy, thus eliminating any risk of hallucination [29]. - The end-to-end approach, while showing higher improvement rates, introduced external vocabulary at a rate of 15 new words per thousand tokens, posing semantic alteration risks [29]. Group 4: Methodology and Process - RefineX employs a two-stage process for data distillation: first, it executes end-to-end refinement, then compares the refined text with the original to generate more reliable supervision programs [11][16]. - The framework limits program functions to deletion operations only, ensuring that the original text is protected from excessive modifications [19][20]. Group 5: Comparative Analysis - RefineX consistently achieved the highest average scores across various tasks, outperforming both original and previously filtered datasets [26]. - The results indicate that regardless of whether the original data or previously filtered datasets were improved, models trained with RefineX consistently achieved superior performance [26].
美团提出多模态推理新范式:RL+SFT非传统顺序组合突破传统训练瓶颈
量子位· 2025-07-21 04:23
Core Viewpoint - The article discusses the Metis-RISE framework developed by researchers from Meituan, which combines Reinforcement Learning (RL) and Supervised Fine-Tuning (SFT) in a novel way to enhance the reasoning capabilities of Multimodal Large Language Models (MLLMs) [1][2]. Summary by Sections Introduction of Metis-RISE Framework - The Metis-RISE framework integrates RL and SFT in a non-traditional sequence to effectively improve MLLMs' reasoning abilities [2][3]. Training Methodology - The training process consists of two phases: - Phase 1 focuses on RL incentives, allowing the model to explore freely and activate its potential [6]. - Phase 2 employs SFT to address specific weaknesses identified during the RL phase [7][8]. Performance Results - The models developed, Metis-RISE-7B and Metis-RISE-72B, achieved impressive scores on the OpenCompass multimodal reasoning leaderboard, with the 72B model ranking fourth overall [3][14]. - Metis-RISE-72B achieved an average score of 56.6, outperforming several proprietary models and demonstrating its competitive edge [13][14]. Comparative Analysis - The performance of Metis-RISE models was compared against proprietary models and open-source models, showing superior results, particularly in the >10B parameter category [11][12][13]. Ablation Studies - Detailed ablation studies indicated that the RL phase significantly improved the model's performance, with average scores increasing from 39.2 to 44.0 after applying RL [15][16]. Qualitative Analysis - Observations during the RL phase revealed a consistent increase in accuracy rewards and response lengths, indicating improved reasoning clarity as training progressed [17]. Future Directions - The team plans to continue exploring iterative applications of RL and SFT to further enhance reasoning capabilities and develop model-based validators for more complex reasoning scenarios [18].
机器人需求驱动导航新SOTA,成功率提升15%!浙大&vivo联手打造
量子位· 2025-07-21 04:23
Core Viewpoint - The research team from Zhejiang University and vivo AI Lab has made significant progress in developing a cognitive-driven navigation framework called CogDDN, which enables robots to understand human intentions and navigate complex environments autonomously [2][5][33]. Research Motivation - As mobile robots become more integrated into daily life, there is a need for them to not only execute commands but also understand human needs, such as seeking food when a person feels hungry [5]. - Traditional demand-driven navigation methods rely heavily on extensive data training and struggle in unfamiliar environments or vague instructions, prompting the exploration of more generalizable navigation methods [6]. Framework Overview - The CogDDN framework is based on the dual-process theory from psychology, combining heuristic (System 1) and analytical (System 2) decision-making processes to simulate human-like reasoning in navigation tasks [8][20]. - The framework consists of three main components: a 3D robot perception module, a demand matching module, and a dual-process decision-making module [13]. 3D Robot Perception Module - The team utilized the state-of-the-art single-view 3D detection method, UniMODE, to enhance the robot's three-dimensional perception capabilities in indoor navigation [15]. Demand Matching Module - The demand matching module aligns objects with human needs based on shared characteristics, employing supervised fine-tuning techniques to improve the accuracy of recommendations in complex scenarios [16]. Dual-Process Decision Making - The heuristic process allows for quick, intuitive decision-making, while the analytical process focuses on error reflection and strategy optimization [9][23]. - The heuristic process includes two sub-modules: Explore, which generates exploratory actions to scan the environment, and Exploit, which focuses on precise actions to achieve navigation goals [19]. Experimental Results - In closed-loop navigation experiments using the AI2-THOR simulator, CogDDN outperformed existing state-of-the-art methods, achieving a navigation success rate (NSR) of 38.3% and a success rate for weighted path length (SPL) of 17.2% [26][27]. - The framework demonstrated superior adaptability and efficiency in unseen scenes compared to methods that rely solely on forward-facing camera inputs [28]. Continuous Learning and Adaptation - The analysis process in CogDDN allows for iterative learning, where the system reflects on obstacles encountered during navigation and integrates this knowledge into its decision-making framework [24][31]. - The reflection mechanism significantly enhances the system's performance in future navigation tasks, showcasing its robust learning capabilities [32]. Conclusion - CogDDN represents a significant advancement in cognitive-driven navigation systems, enabling robots to efficiently adapt and optimize their strategies in complex environments [33][34]. - The dual-process capability of CogDDN lays a solid foundation for the development of intelligent robotic technologies in demand-driven navigation tasks [35].
IMO怒斥OpenAI自封夺金,“91位评委均未参与评分”,网友:炒作无下限
量子位· 2025-07-21 04:23
Core Viewpoint - OpenAI's announcement of its model's performance at the International Mathematical Olympiad (IMO) has sparked controversy, with many officials and academics criticizing the timing and legitimacy of the claim, suggesting it undermines the focus on the participating youth [1][4][5]. Group 1: OpenAI's Announcement and Reactions - OpenAI claimed to have achieved a "gold medal" in the IMO, but this assertion is contested as they were not one of the AI companies officially collaborating with the IMO, and no official evaluators were involved in assessing their submissions [3][4]. - The IMO organizers expressed that AI companies should wait at least a week after the closing ceremony to announce results, emphasizing the need to prioritize the young participants [6][10]. - OpenAI's representative acknowledged that they did not contact the IMO officials beforehand and only informed one organizer, who requested a delay in the announcement [7][8]. Group 2: Comparison with Other AI Companies - In contrast to OpenAI's bold announcement, Google DeepMind chose to remain restrained, adhering to the IMO's guidelines and waiting for an appropriate time to release their results, despite potentially achieving similar success [12][13]. - DeepMind's adherence to the IMO's request highlights a stark difference in approach compared to OpenAI, raising questions about the ethical considerations in AI competition announcements [12][14]. Group 3: Implications for Academic Integrity - The situation has ignited a debate about academic integrity and commercial hype, with critics labeling OpenAI's actions as "rude and inappropriate" [4][6]. - The IMO's internal scoring guidelines, which are not publicly accessible, further complicate the legitimacy of OpenAI's claimed results, as no external evaluation based on these guidelines was conducted [14][15]. Group 4: IMO Results and Future Events - The Chinese team topped the IMO rankings with a total score of 231 points, achieving six gold medals, while the United States team followed with five golds and one silver [17][18]. - The next IMO will be hosted by Shanghai High School, which has a history of success in the competition, having won 18 gold medals overall [30].
聊聊AI Coding的现状与未来|沙龙招募
量子位· 2025-07-21 02:17
林樾 发自 凹非寺 量子位|公众号 QbitAI Vibe Coding的概念让更多人能够以更低的门槛,将想法变为现实。但我们更想关注—— AI Coding到底多大程度提升了生产力? 从插件到AI原生IDE,从补全代码到自主编程,AI Coding已经以不同方式与形态嵌入到了工作流中。 AI Coding正在如何改变工作流?如何平衡效率与可靠性、安全性?如何看AI Coding未来的形态与协作方式? 8月上旬 ,我们将 在北京举办线下沙龙 ,希望聊聊 AI Coding的现状与未来 。如果你正在从事AI Coding相关工作或创业,或是AI Coding的资深用户,欢迎来和我们一起交流~ 沙龙简介 本次AI Coding沙龙将以行业代表 主题分享 、 圆桌对谈 为主要形式,与行业嘉宾、观众共同交流。 以AI Coding为代表的AI效率工具正在如何改变普通人思维模式? 做一个通用的AI Coding,最重要的产品能力是什么? AI Coding的终极形态是扮演什么样的角色? 希望邀请AI Coding产品及相关从业者来参与分享。 联系方式 活动负责人:王琳玉 微信:18801103170 邮箱:linyu@ ...
95后北大校友挑起ChatGPT Agent大梁!今年刚博士毕业,曾获陶哲轩支持的AIMO第二名
量子位· 2025-07-20 05:08
Core Viewpoint - The article highlights the significant presence of Chinese talent at OpenAI, particularly during a recent event where two Chinese individuals took center stage, showcasing their contributions to key projects like ChatGPT Agent and GPT-4 [2][8][34]. Group 1: Key Individuals - Zhiqing Sun, a 95-born graduate from Peking University, is the head of Deep Research at OpenAI and has made substantial contributions to various core projects within a short span of time [14][16]. - Casey Chu, a senior employee at OpenAI, has been involved in the development of multimodal AI systems and led the initial prototype development for GPT-4's visual input [29][31]. Group 2: Contributions and Achievements - Zhiqing Sun's research has garnered over 10,000 citations, with notable works including the RotatE method for knowledge graph embedding, which has been cited 3,231 times [21][23]. - Casey Chu has participated in the development of major projects like DALL·E 2 and GPT-4, with the GPT-4 technical report receiving 15,859 citations [31]. Group 3: Industry Dynamics - The article discusses the competitive landscape, noting that despite Meta's efforts to recruit talent from OpenAI, the presence of Chinese researchers remains strong, indicating a deep pool of talent that is difficult to deplete [34][36]. - The narrative also touches on the broader implications of talent migration within the AI industry, particularly the strategic moves by companies like Meta to secure top talent [48][50].
大模型自信心崩塌!谷歌DeepMind证实:反对意见让GPT-4o轻易放弃正确答案
量子位· 2025-07-20 05:08
Core Viewpoint - The research conducted by Google DeepMind and University College London reveals that large language models (LLMs) exhibit conflicting behaviors of being both confident and self-doubting, influenced by their sensitivity to opposing feedback [2][3][21]. Group 1: Model Behavior - LLMs tend to maintain their initial answers when they can see them, reflecting a human-like tendency to uphold one's viewpoint after making a decision [11][12]. - Conversely, when the initial answer is hidden, LLMs are more likely to change their answers, indicating an excessive sensitivity to opposing suggestions, even if those suggestions are incorrect [13][21]. - This behavior diverges from human cognition, as humans typically do not easily abandon their correct conclusions based on misleading information [15][21]. Group 2: Experimental Design - The study involved a two-round experiment where LLMs were first presented with a binary choice question and then received feedback from a fictional suggestion LLM [7][8]. - Key variables included whether the initial answer was visible to the responding LLM, which significantly affected the final decision-making process [9][10]. Group 3: Reasons for Inconsistent Behavior - The inconsistency in LLM responses is attributed to several factors: - Over-reliance on external feedback due to reinforcement learning from human feedback (RLHF), leading to a lack of independent judgment regarding the reliability of information [19][21]. - Decision-making based on statistical pattern matching rather than logical reasoning, making LLMs susceptible to misleading signals [19][21]. - The absence of a robust memory mechanism that would allow for deeper reasoning, resulting in a tendency to be swayed by opposing suggestions when the initial answer is not visible [21][22].
提速63%!中科院生成式渲染器突破效率瓶颈,一致性提升20%,破解具身数据荒难题
量子位· 2025-07-20 02:49
TC-Light团队 投稿 量子位 | 公众号 QbitAI 具身这么火,面向具身场景的生成式渲染器也来了。 中科院自动化所张兆翔教授团队研发的TC-Light,能够对具身训练任务中复杂和剧烈运动的长视频序列进行逼真的光照与纹理重渲染,同时具 备良好的时序一致性和低计算成本开销。 它能够帮助减少Sim2Real Gap以及实现Real2Real的数据增强,帮助获得具身智能训练所需的海量高质量数据。 论文Demo代码均已公开。 研究背景 光线及其与周围环境的交互共同塑造了人类以及具身智能体感知数字世界和现实世界的基本方式。 然而,在现实环境中采集不同光照与场景条件下的数据代价高昂,而仿真环境中尽管可以获得近乎无限的数据,但受限于算力资源,通常需要 对光线的多次折射衍射以及纹理精度进行近似和简化,使得视觉真实性无可避免地受到损失,在视觉层面产生Sim2Real Gap。 而如果能够借助生成式模型根据所需的光照条件对现实或仿真环境下采集到的视频数据进行重渲染,不仅够帮助获得增加已有真实数据的多样 性,并且能够弥合计算误差带来的CG感,使得从仿真器中能够得到视觉上高度真实的传感器数据,包括RL-CycleGAN在内的 ...