Workflow
信息瓶颈
icon
Search documents
LeCun离职前的吐槽太猛了
量子位· 2025-12-21 05:45
Core Viewpoint - LeCun expresses skepticism about the potential of large language models (LLMs) to achieve artificial general intelligence (AGI), arguing that the path to superintelligence through LLMs is fundamentally flawed [2][78]. Group 1: Departure from Meta - LeCun is leaving Meta after nearly 12 years, criticizing the company's increasingly closed approach to research and its focus on short-term projects [3][11][26]. - He plans to establish a new company named Advanced Machine Intelligence (AMI), which will prioritize open research and focus on world models [10][19]. Group 2: World Models vs. LLMs - LeCun believes that world models, which handle high-dimensional and continuous data, are fundamentally different from LLMs, which excel at discrete text data [28][29]. - He argues that relying solely on text data will never allow AI to reach human intelligence levels, as the complexity of real-world data is far greater than that of text [31][32]. Group 3: Research Philosophy - LeCun emphasizes the importance of open research and publication, stating that without sharing results, research lacks validity [15][17]. - He critiques Meta's shift towards short-term projects, suggesting that true breakthroughs require long-term, open-ended research [18][26]. Group 4: Future of AI - LeCun envisions that the development of world models and planning capabilities could lead to significant advancements in AI, but achieving human-level intelligence will require substantial foundational work and theoretical innovation [84][85]. - He asserts that the most challenging aspect of AI development is not reaching human intelligence but rather achieving the intelligence level of dogs, as this requires a deep understanding of foundational theories [88][89]. Group 5: Personal Mission - At 65, LeCun remains committed to enhancing human intelligence, viewing it as the most scarce resource and a key driver for societal progress [92][94]. - He reflects on his career, expressing a desire to continue contributing to the field and emphasizing the importance of open collaboration in scientific advancement [103].
倒计时3周离职,LeCun最后警告:硅谷已陷入集体幻觉
3 6 Ke· 2025-12-16 07:11
Core Viewpoint - LeCun criticizes the obsession with large language models (LLMs) in Silicon Valley, asserting that this approach is a dead end and will not lead to artificial general intelligence (AGI) [1][3][26] Group 1: Critique of Current AI Approaches - LeCun argues that the current trend of stacking LLMs and relying on extensive synthetic data is misguided and ineffective for achieving true intelligence [1][3][26] - He emphasizes that the real challenge in AI is not achieving human-like intelligence but rather understanding basic intelligence, as demonstrated by simple creatures like cats and children [3][12] - The focus on LLMs is seen as a dangerous "herd mentality" in the industry, with major companies like OpenAI, Google, and Meta all pursuing similar strategies [26][30] Group 2: Introduction of World Models - LeCun is advocating for a different approach called "world models," which involves making predictions in an abstract representation space rather than relying solely on pixel-level outputs [3][14] - He believes that world models can effectively handle high-dimensional, continuous, and noisy data, which LLMs struggle with [14][12] - The concept of world models is tied to the idea of planning, where the system predicts the outcomes of actions to optimize task completion [14][12] Group 3: Future Directions and Company Formation - LeCun plans to establish a new company, Advanced Machine Intelligence (AMI), focusing on world models and maintaining an open research tradition [4][5][30] - AMI aims to not only conduct research but also develop practical products related to world models and planning [9][30] - The company will be global, with headquarters in Paris and offices in other locations, including New York [30] Group 4: Perspectives on AGI and AI Development Timeline - LeCun dismisses the concept of AGI as meaningless, arguing that human intelligence is highly specialized and cannot be replicated in a single model [31][36] - He predicts that significant advancements in AI could occur within 5-10 years, potentially achieving intelligence levels comparable to dogs, but acknowledges that unforeseen obstacles may extend this timeline [31][33] Group 5: Advice for Future AI Professionals - LeCun advises against pursuing computer science as a primary focus, suggesting instead to study subjects with long-lasting relevance, such as mathematics, engineering, and physics [45][46] - He emphasizes the importance of learning how to learn and adapting to rapid technological changes in the AI field [45][46]
AAAI 2026 Oral:明略科技开创稀疏数据「信息瓶颈动态压缩」,精度+速度双SOTA
机器之心· 2025-12-02 06:47
Core Insights - The article discusses the challenges of "Efficient AI," particularly in the context of transformer models becoming larger and more general, while also becoming computationally heavy for edge devices like robots [1][2] - A paper titled "CompTrack," accepted for oral presentation at AAAI 2026, addresses the issue of whether models need to process all input data, showcasing how compression techniques can significantly reduce computational costs while maintaining or even improving model performance [2][14] Redundancy Challenges - Current AI models face "Dual-Redundancy" challenges, which include: 1. Spatial Redundancy: Unrelated background points and blank areas are processed, wasting computational resources and degrading accuracy [3][5] 2. Informational Redundancy: Even in relevant foreground targets, there is a prevalence of redundant and low-value information, which can lead to inefficiencies [5][7] CompTrack Framework - CompTrack proposes an end-to-end framework that addresses both types of redundancy simultaneously [7] - The framework includes: 1. A Spatial Foreground Predictor (SFP) that filters out low-information background noise using information entropy theory [8] 2. An Information Bottleneck-guided Dynamic Token Compression (IB-DTC) module designed to dynamically compress information redundancy in the foreground [10][11] Efficiency and Performance - The IB-DTC module is significant for Efficient AI as it: 1. Is based on the Information Bottleneck principle, retaining only valuable information for predictions [11] 2. Utilizes online Singular Value Decomposition (SVD) for dynamic compression rates based on the intrinsic rank of input data [12] 3. Allows for end-to-end training by using SVD as a guide for optimal compression rates [12] Application and Results - CompTrack has been applied to challenging 3D point cloud tracking tasks, demonstrating that systematic compression of information redundancy is highly effective [14] - The framework not only enhances efficiency but also sets a precedent for addressing information redundancy in various fields, including sensor fusion in robotics and multimodal processing in visual-language models [14][15] - Performance metrics show that CompTrack achieves real-time performance at 80 FPS on RTX 3090, surpassing state-of-the-art methods, with a significant reduction in computational load to 0.94G FLOPs [15]
中科大提出动作价值表征学习新方法,率先填补长期决策信息的缺失
量子位· 2025-03-31 04:35
Core Viewpoint - The article discusses the introduction of a novel robust action value representation learning method called ROUSER, which addresses the lack of long-term information in visual reinforcement learning by utilizing the Information Bottleneck framework [2][9]. Group 1: ROUSER Methodology - ROUSER maximizes the mutual information between the representation and action value to retain long-term information while minimizing the mutual information between the representation and state-action pairs to filter out irrelevant features [4][10]. - The method decomposes the robust representation of state-action pairs into representations that include single-step rewards and the robust representation of the next state-action pair, allowing for effective learning despite unknown action values [5][10]. Group 2: Experimental Results - In experiments involving 12 tasks with background and color interference, ROUSER outperformed various advanced methods in 11 of the tasks, demonstrating its effectiveness in enhancing generalization capabilities [6][18]. - ROUSER is compatible with both continuous and discrete control tasks, as evidenced by experiments conducted in the Procgen environment, which showed improved generalization performance when combined with value-based VRL methods [21][22]. Group 3: Theoretical Foundations - The theoretical proof indicates that ROUSER can accurately estimate action values using the learned vectorized representations, thereby improving the robustness of various visual reinforcement learning algorithms [3][17].