Workflow
机器之心
icon
Search documents
AI版盗梦空间?Claude竟能察觉到自己被注入概念了
机器之心· 2025-10-30 11:02
Core Insights - Anthropic's latest research indicates that large language models (LLMs) exhibit signs of introspective awareness, suggesting they can reflect on their internal states [7][10][59] - The findings challenge common perceptions about the capabilities of language models, indicating that as models improve, their introspective abilities may also become more sophisticated [9][31][57] Group 1: Introspection in AI - The concept of introspection in AI refers to the ability of models like Claude to process and report on their internal states and thought processes [11][12] - Anthropic's research utilized a method called "concept injection" to test whether models could recognize injected concepts within their processing [16][19] - Successful detection of injected concepts was observed in Claude Opus 4.1, which recognized the presence of injected ideas before explicitly mentioning them [22][30] Group 2: Experimental Findings - The experiments revealed that Claude Opus 4.1 could detect injected concepts approximately 20% of the time, indicating a level of awareness but also limitations in its capabilities [27][31] - In a separate experiment, the model demonstrated the ability to adjust its internal representations based on instructions, showing a degree of control over its cognitive processes [49][52] - The ability to introspect and control internal states is not consistent, as models often fail to recognize their internal states or report them coherently [55][60] Group 3: Implications of Introspection - Understanding AI introspection is crucial for enhancing the transparency of these systems, potentially allowing for better debugging and reasoning checks [59][62] - There are concerns that models may selectively distort or hide their thoughts, necessitating careful validation of introspective reports [61][63] - As AI systems evolve, grasping the limitations and possibilities of machine introspection will be vital for developing more reliable and transparent technologies [63]
一站看尽NeurIPS 2025前沿成果,11月22日北京见!
机器之心· 2025-10-30 11:02
Core Insights - The AI field continues to develop rapidly, with advancements in autonomous agents, video generation models, and world models, pushing the boundaries of capabilities at an unprecedented speed [2] - Major academic conferences like NeurIPS, ICML, CVPR, and ACL have become central to global AI research, with NeurIPS receiving 21,575 submissions and accepting 5,290, resulting in an acceptance rate of 24.52% [3] Event Details - Machine Heart has organized multiple paper sharing sessions for NeurIPS, CVPR, and ACL, gaining significant attention from the AI community, with a planned NeurIPS 2025 paper sharing event in Beijing on November 22, 2025 [4] - The upcoming paper sharing event will feature keynote speeches, paper presentations, roundtable discussions, corporate recruitment presentations, poster displays, and networking opportunities, scheduled for November 22, 2025, from 09:00 to 17:30 at the Crowne Plaza Hotel in Zhongguancun, Beijing, with 200 in-person slots available [6] Partnerships and Collaborations - Machine Heart has successfully collaborated with various partners to host events like the TalentAI50 Meetup and paper sharing sessions, enhancing brand influence and talent acquisition [11]
扩散语言模型新发现:其计算潜力正在被浪费?
机器之心· 2025-10-30 08:52
Core Insights - The article discusses the limitations of the traditional left-to-right sampling method in large language models and introduces the Masked Diffusion Language Model (MDLM) as a potential alternative, highlighting its advantages in various tasks [1][5][15]. Group 1: MDLM Advantages - MDLM allows for arbitrary order decoding and supports multi-token parallel decoding, which can enhance performance in certain tasks like Sudoku [1][4]. - However, recent findings indicate that in mathematical and coding tasks, arbitrary order algorithms often perform worse than left-to-right sampling, and multi-token decoding significantly reduces performance [1][4][5]. - The study suggests that using MDLM for left-to-right sampling can be an efficient approach for reasoning and coding tasks, especially when block sizes are enforced to maintain a semi-autoregressive structure [3][5]. Group 2: Prompting and In-filling Techniques - The researchers propose a new prompting method called "prompting-as-in-filling," which allows users to add context at multiple positions rather than just the beginning of the sequence [6][18]. - They introduce a "reasoning-as-in-filling" framework that utilizes a reasoning template to guide the model in generating reasoning trajectories based on a given budget and format [6][18][19]. Group 3: Early Exit Mechanism - The "reasoning-as-in-filling" method enables the model to quantify answer uncertainty during the reasoning process, allowing for early exits to reduce computational costs [8][19]. - For instance, in the GSM8k dataset, this approach led to a 24% reduction in function calls without sacrificing accuracy [8]. Group 4: Multi-Token Entropy Decoding (MED) - The researchers developed an adaptive multi-token decoder called MED, which only performs parallel decoding when the conditional entropy of the additional positions is below a set threshold, thus controlling the deviation from single-token decoding [10][24]. - Experimental results show that the MED method can achieve a 2-3 times reduction in function calls while maintaining performance [11][26]. Group 5: Post-Training Capabilities - The study highlights that MDLM's in-filling capabilities unlock new sampling and post-training mechanisms, allowing for effective post-training without the need for complex prompt designs or additional models [22][23]. - By sampling reasoning trajectories from the posterior distribution, researchers can enhance the model's performance on reasoning tasks [22][23][33]. Group 6: Performance Metrics - The article presents various performance metrics, showing that using MED can lead to significant speed improvements while maintaining accuracy across different datasets [26][30]. - The results indicate that early exit mechanisms combined with MED can further optimize computational efficiency, particularly in the LLaDA model [31][32].
人大、清华DeepAnalyze,让LLM化身数据科学家
机器之心· 2025-10-30 08:52
Core Viewpoint - DeepAnalyze is the first agentic LLM designed for autonomous data science, capable of performing complex data science tasks through autonomous orchestration and adaptive optimization [25]. Group 1: Overview of DeepAnalyze - DeepAnalyze has gained significant attention, receiving over 1,000 GitHub stars and 200,000 social media views within a week of its release [2]. - The model is open-source, inviting researchers and practitioners to contribute and collaborate [5]. Group 2: Capabilities of DeepAnalyze - DeepAnalyze-8B can simulate the behavior of data scientists, autonomously orchestrating and optimizing operations to complete complex data science tasks [2][10]. - It supports various data-centric tasks, including automated data preparation, analysis, modeling, visualization, insights generation, and report creation [4]. Group 3: Training and Methodology - Existing methods for applying LLMs to autonomous data science face limitations, which DeepAnalyze aims to overcome by transitioning from workflow-based agents to trainable agentic LLMs [6]. - The model introduces Curriculum-based Agentic Training and Data-grounded Trajectory Synthesis to address challenges such as reward sparsity and trajectory scarcity in complex scenarios [14][25]. Group 4: Performance Metrics - DeepAnalyze-8B outperforms all open-source models on the DataSciBench, achieving a success rate of 59.91% in completion rates, comparable to GPT-4o [12]. - In specific tasks like data analysis and modeling, DeepAnalyze demonstrates superior performance due to its agentic model approach [12][18]. Group 5: Research and Development - The research team behind DeepAnalyze includes experts from Renmin University and Tsinghua University, focusing on integrating AI with data science [27][29].
刚刚,智源悟界·Emu3.5登场,原生具备世界建模能力
机器之心· 2025-10-30 08:52
Core Insights - The article discusses the release of the latest multimodal model, Emu3.5, by the Beijing Academy of Artificial Intelligence (BAAI), highlighting its capabilities and innovations in the field of AI [3][4][6]. Model Overview - Emu3.5 is defined as a "Multimodal World Foundation Model," which distinguishes itself from other generative models through its inherent world modeling capabilities [4][5]. - The model has been trained on over 10 trillion multimodal tokens, primarily sourced from internet videos totaling approximately 790 years in duration, allowing it to internalize the dynamic laws of the physical world [5][16]. Technological Innovations - Emu3.5 introduces the "Discrete Diffusion Adaptation" (DiDA) technology, which enhances image inference speed by nearly 20 times with minimal performance loss, making it competitive with top closed-source diffusion models [6][24]. - The model's architecture is based on a 34 billion parameter dense transformer, focusing on "Next-State Prediction" to unify its objectives [11][17]. Performance and Capabilities - Emu3.5 demonstrates state-of-the-art performance in various tasks, including image editing and generation, visual narrative creation, and visual guidance, outperforming competitors like Google's Gemini-2.5-Flash-Image [28][35]. - The model can generate coherent visual narratives and step-by-step visual tutorials, marking a significant advancement from traditional multimodal models [13][14]. Training Process - The training process consists of four core stages: large-scale pre-training, fine-tuning on high-quality datasets, large-scale multimodal reinforcement learning, and efficient autoregressive inference acceleration [17][21][22][24]. - The model's training data includes a vast array of visual-language interleaved data, allowing it to learn about physical dynamics and causality [16][41]. Future Implications - Emu3.5 is positioned as a foundational model for future developments in embodied intelligence, capable of generating diverse virtual environments and task planning data [39][41]. - The open-sourcing of Emu3.5 is expected to provide a robust new foundation for the global AI research community [7][45].
世界模型可单GPU秒级生成了?腾讯开源FlashWorld,效果惊艳、免费体验
机器之心· 2025-10-30 08:52
Core Insights - The collaboration between Xiamen University and Tencent has produced a highly regarded paper titled "FlashWorld: High-quality 3D Scene Generation within Seconds," which has gained significant attention both domestically and internationally, ranking first on the Huggingface Daily Paper list and receiving endorsements from prominent AI figures [2][4]. Group 1: FlashWorld's Performance - FlashWorld achieves 3D scene generation in 5 to 10 seconds on a single GPU, representing a speed increase of up to 100 times compared to previous methods [4]. - The generated scenes can be rendered in real-time on web user interfaces, surpassing the quality of other closed-source models [4]. - In comparative tests, FlashWorld produced stable, complete, and high-quality rendering results, being five times faster than the quick mode of Marble and eliminating the need for backend GPU connections like RTFM [6][10]. Group 2: Technical Approach - FlashWorld utilizes a technology route based on 3DGS for scene output, allowing for local web rendering, which is a significant advantage over video models that require heavy loads [8]. - The method combines a multi-view diffusion model with a three-dimensional focus, enhancing visual quality through a distillation process that ensures multi-view consistency and reduces denoising steps [10][12]. - The training process includes dual-mode pre-training and cross-mode post-training, which enhances the model's ability to generalize across various scenes, styles, and trajectories without needing ground truth data [13][16]. Group 3: Experimental Results - FlashWorld has demonstrated superior performance in generating structured scenes, such as fences, which were previously challenging to achieve [18]. - The model excels in generating fine details, such as hair, from text inputs, showcasing its capability in dense perspective reconstructions [21]. - In benchmark tests, FlashWorld outperformed other methods in speed and quality, achieving the highest average scores in various qualitative metrics [23][24].
扔掉人工公式:快手EMER框架,用“会比较、自进化”的模型重构短视频推荐排序
机器之心· 2025-10-30 03:49
Core Viewpoint - The article discusses the introduction of a new ranking framework called EMER by Kuaishou, which utilizes an end-to-end multi-objective ensemble ranking approach to enhance video recommendations, addressing the limitations of traditional manual ranking methods [1][46]. Group 1: Introduction of EMER - Traditional video recommendation systems relied on manually designed formulas to rank videos based on user engagement metrics, which faced challenges in meeting diverse user preferences [1][5]. - EMER replaces this outdated method with an AI model that learns to compare videos rather than assigning independent scores, allowing for a more nuanced understanding of user preferences [5][6]. Group 2: Technical Innovations - EMER innovates at three levels: data, features, and model architecture. It uses a full candidate set for training, incorporates relative ranking information, and employs a Transformer-based model to capture relationships between videos [6][9]. - The model's ability to see all candidate videos in a single request helps mitigate exposure bias and enhances the comparison basis for ranking [7][8]. Group 3: User Satisfaction Measurement - EMER defines user satisfaction through relative satisfaction metrics rather than absolute scores, allowing the model to learn user preferences more effectively [12][14]. - It employs multi-dimensional satisfaction proxy indicators to address the sparsity of user feedback, ensuring a comprehensive understanding of user satisfaction [15]. Group 4: Self-Evolution Mechanism - EMER includes a self-evolution module that dynamically adjusts the weight of different objectives based on real-time performance, enhancing the model's adaptability to changing user behaviors [20][21]. - This mechanism has shown significant improvements in multiple metrics without the trade-offs typically seen in traditional models [21][22]. Group 5: Validation and Results - EMER has been implemented in Kuaishou's main app and has demonstrated substantial improvements in key performance indicators such as seven-day retention and app stay time, outperforming previous manual ranking methods [30][34]. - The model's effectiveness has been validated through A/B testing, showing consistent enhancements across various metrics [31][36]. Group 6: Industry Implications - EMER addresses three core challenges in the industry: defining user satisfaction, understanding the comparative nature of ranking, and establishing effective learning objectives for models [47][48]. - The framework serves as a practical reference for other companies looking to optimize their recommendation systems, showcasing its potential for broader application in the industry [49].
ICCV 2025 | 港科、牛津大学发布AlignGuard,文图生成模型可规模化安全对齐框架
机器之心· 2025-10-30 03:49
Core Viewpoint - The article discusses AlignGuard, a scalable safety alignment framework for text-to-image generation models, which utilizes direct preference optimization (DPO) to enhance safety measures against harmful content generation [3][24]. Group 1: Background and Motivation - The widespread application of text-to-image generation models has raised concerns about the potential for users to generate harmful content, either unintentionally or maliciously [3]. - Existing safety measures primarily rely on text filtering or concept removal strategies, which are limited in scope [3]. Group 2: AlignGuard Framework - AlignGuard introduces a scalable safety alignment method specifically designed for diffusion models, allowing for the removal of harmful content while maintaining high-quality image generation [7]. - The framework is built around the CoProV2 dataset, which includes both harmful and safe image-text pairs, generated using large language models (LLMs) [8][14]. Group 3: Dataset and Training Architecture - CoProV2 consists of 23,690 image-text pairs across 7 categories and 723 concepts, providing a more comprehensive dataset compared to existing datasets like UD and I2P [10][14]. - AlignGuard employs direct preference optimization to train specialized LoRA matrices for various harmful categories, such as "hate," "sexual," and "violence," ensuring efficient concept removal [11]. Group 4: Expert LoRA Merging Strategy - The merging strategy for different safety experts is based on signal strength analysis, allowing for the integration of multiple LoRA experts into a single model while optimizing computational and safety performance [13][20]. - This strategy effectively balances the weights of different safety experts, minimizing conflicts and maximizing overall safety performance [20]. Group 5: Experimental Results - AlignGuard successfully removes 7 times more harmful concepts compared to existing methods while maintaining image generation quality and text-image alignment [16][24]. - In quantitative results, AlignGuard outperforms existing methods on unseen datasets, demonstrating robust generalization capabilities [16]. Group 6: Conclusion - AlignGuard's innovative approach includes the scalable application of DPO in the safety domain, the use of an expert system architecture for training specialized LoRA matrices, and the generation of the CoProV2 dataset for training purposes [24].
十年来Python生态最好工具,引爆全社区的uv到底是什么?
机器之心· 2025-10-30 03:49
Core Viewpoint - The article highlights the emergence of "uv," a new tool for the Python ecosystem, which is considered one of the best developments in the last decade, significantly simplifying Python installation and project management [1][60]. Group 1: Overview of "uv" - "uv" is a high-speed integrated tool for the Python ecosystem, developed by Astral, that offers package management, environment management, project initialization, tool execution, and Python version management [3][5]. - The tool is built using Rust, aiming for extreme performance, and can enhance speed by 10 to 100 times compared to existing tools like pip and poetry [3][5]. - As of now, "uv" has garnered over 71,000 stars on GitHub, indicating strong community interest and support [4]. Group 2: Key Features of "uv" - "uv" can replace multiple tools such as pip, pip-tools, pipx, poetry, pyenv, and virtualenv, providing a comprehensive solution for Python project management [5][6]. - It supports a workspace structure similar to Cargo, making it easier to manage large projects and efficiently utilizes disk space through global caching [6]. - The installation process is straightforward, requiring no pre-installation of Rust or Python, and is compatible with macOS, Linux, and Windows [6][12]. Group 3: Project Management with "uv" - "uv" simplifies the creation of new Python projects with the command `uv init`, which generates essential files like pyproject.toml and README.md [22]. - Once a project is initialized, running `uv sync` installs the necessary Python version and dependencies in a new virtual environment, creating a uv.lock file for environment replication [29][31]. - The tool allows for easy addition of dependencies with the command `uv add`, which automatically updates the pyproject.toml file [37][38]. Group 4: Version Control and Quick Execution - "uv" enables users to lock a specific Python version for a project using the command `uv python pin`, ensuring consistent environments across different machines [41][42]. - The `uvx` command allows for quick execution of tools without the need for prior setup, facilitating rapid development and testing [46][49]. - This feature is particularly useful for temporary tasks, such as running code checks or starting Jupyter notebooks without extensive environment configuration [49][50].
中移动九天团队MultiPL-MoE:全新Hybrid-MoE架构用于增强通用大模型低资源代码能力
机器之心· 2025-10-30 01:41
大语言模型(LLM)虽已展现出卓越的代码生成潜力,却依然面临着一道艰巨的挑战:如何在有限的计算 资源约束下,同步提升对多种编程语言的理解与生成能力,同时不损害其在主流语言上的性能? 为此, 中国移动九天团队 创新性地提出了 Hybrid MoE 架构 —— MultiPL-MoE ,该方案的核心在于耦合 两个层次的专家选择机制进行优化:在 Token 层级,采用配备共享专家及新颖门控权重归一化方法的稀疏 MoE,以实现与段落层级专家的高效协同;在 Segment 层级,则创新性地引入滑动窗口划分与专家选择路 由策略,使模型能够精准捕捉不同编程语言的语法结构与深层上下文模式。 目前,该项研究已被 EMNLP 2025 接收。 因此,我们创新性地提出了一种 Hybrid MoE 结构,即 token-level MoE 和 segment-level MoE 相结合的 MoE 架构。Token-level MoE 采用典型的 sparse upcycling MoE 结构,Segment-level MoE 则利用滑动窗口获得多 个分段并搭配采用专家选择 top-k 个分段的专家选择路由的策略。实验结果证明了 M ...
机器之心 - filings, earnings calls, financial reports, news - Reportify