Workflow
机器之心
icon
Search documents
刚刚,OpenAI推出学习模式,AI教师真来了,系统提示词已泄露
机器之心· 2025-07-30 00:48
Core Viewpoint - ChatGPT has introduced a new feature called Study Mode, which aims to enhance user learning by guiding them through problem-solving rather than simply providing answers [1][2][4]. Summary by Sections Features of Study Mode - The Study Mode includes interactive prompts that encourage active learning through Socratic questioning and hints, rather than direct answers [5]. - Responses are organized into understandable sections, highlighting key connections between topics to reduce cognitive load [5]. - The mode offers personalized support tailored to the user's skill level and previous interactions [5]. - Knowledge assessments, including quizzes and open-ended questions, are provided to track progress and reinforce learning [5]. - Users can easily switch to Study Mode during conversations, allowing for flexible learning objectives [5]. User Experience - Initial feedback on the Study Mode has been overwhelmingly positive, indicating its effectiveness in enhancing the learning experience [6]. - A practical example demonstrated how ChatGPT assesses the user's understanding before tailoring the teaching approach to their knowledge level [9]. Development Insights - OpenAI has collaborated with educators and experts to create a system of prompts that support deeper learning behaviors, such as encouraging active participation and providing actionable feedback [13]. - The underlying principles of the Study Mode are based on extensive research in learning sciences [13]. Prompt Engineering - OpenAI has openly shared the key components of the system prompts used in Study Mode, emphasizing the importance of understanding user goals and building on existing knowledge [16][17][18]. - The approach focuses on guiding users through questions and prompts rather than providing direct answers, fostering a collaborative learning environment [19][22].
AI安全上,开源仍胜闭源,Meta、UCB防御LLM提示词注入攻击
机器之心· 2025-07-30 00:48
Core Viewpoint - Meta and UCB have developed the first industrial-grade secure large language model, Meta-SecAlign-70B, which demonstrates superior robustness against prompt injection attacks compared to existing closed-source solutions like gpt-4o and gemini-2.5-flash, while also exhibiting enhanced agentic abilities [1][17]. Group 1: Background on Prompt Injection Attacks - Large Language Models (LLMs) have become crucial components in AI systems, interacting with both trusted users and untrusted environments [4]. - Prompt injection attacks pose a significant threat, where LLMs may be misled by malicious instructions embedded within the data they process, leading to unintended actions [5][10]. - The OWASP security community has identified prompt injection attacks as a primary threat to LLM-integrated applications, successfully targeting industrial AI systems like Google Bard and Slack AI [10]. Group 2: Defense Mechanisms Against Prompt Injection - The core objective of the defense strategy is to train LLMs to distinguish between prompts and data, ensuring that only the prompt is followed while treating the data as pure information [11][12]. - The SecAlign++ method involves adding special delimiters to separate prompts from data, followed by training the LLM to prefer safe outputs and avoid unsafe responses [12][14]. - Meta-SecAlign-70B, trained using the SecAlign++ method, is the first industrial-grade secure LLM that surpasses the performance of existing closed-source models [17][21]. Group 3: Performance and Robustness - Meta-SecAlign-70B shows a lower attack success rate across seven prompt injection benchmarks compared to existing closed-source models, while maintaining competitive utility in agent tasks [19][20]. - The model exhibits significant robustness, achieving an attack success rate of less than 2% in most scenarios after fine-tuning on a 19K instruction dataset, and this robustness generalizes to tasks outside the training data [20][21]. - The open-source nature of Meta-SecAlign-70B aims to break the monopoly of closed-source models on defense methods, facilitating rapid advancements in AI security research [21].
WAIC 2025大黑马,一个「谢耳朵AI」如何用分子式超越Grok-4
机器之心· 2025-07-29 10:31
Core Insights - The article highlights the launch of the Intern-S1 multimodal model by Shanghai AI Laboratory, which is positioned as a leading open-source model in the field of scientific research, showcasing significant advancements in AI for science [5][12][17]. Group 1: Model Capabilities - Intern-S1 is recognized for its superior performance in scientific reasoning tasks, outperforming leading closed-source models like Grok-4, particularly in fields such as chemistry, materials science, and biology [12][17]. - The model integrates a 235 billion parameter MoE language model and a 6 billion vision encoder, trained on 5 trillion tokens, with over 2.5 trillion tokens specifically from scientific domains [25][21]. - Intern-S1 demonstrates a 70% improvement in compression rates for chemical formulas compared to previous models, indicating enhanced efficiency in processing complex scientific data [26]. Group 2: Technological Innovations - The model employs a dynamic tokenizer and temporal signal encoder to effectively handle various complex scientific modalities, addressing challenges posed by data heterogeneity and semantic understanding [26]. - Intern-S1's training costs for reinforcement learning have been reduced by tenfold due to collaborative breakthroughs in system and algorithm optimization [30]. - The model's architecture allows for a unique "cross-modal scientific analysis engine," enabling it to interpret complex scientific data such as chemical structures and seismic signals accurately [16][17]. Group 3: Open Source and Community Engagement - Since its initial release in 2023, the "ShuSheng" model family has been continuously upgraded and expanded, fostering an active open-source community with participation from hundreds of thousands of developers [32][33]. - The Shanghai AI Laboratory has launched a comprehensive open-source toolchain that includes frameworks for data processing, pre-training, fine-tuning, deployment, and evaluation, aimed at lowering barriers for research and application [32]. - The Intern-Discovery platform, based on Intern-S1, has been introduced to enhance collaboration among researchers, tools, and research subjects, promoting a new phase of scientific discovery [6][33].
通义实验室大火的 WebAgent 续作:全开源模型方案超过GPT4.1 , 收获开源SOTA
机器之心· 2025-07-29 10:31
Core Insights - The article introduces WebShaper, a new paradigm for synthesizing training data for information-seeking (IS) tasks, achieving a state-of-the-art (SOTA) score of 60.1 on the GAIA benchmark using an open-source model [1][6][30] - WebShaper addresses the lack of high-quality training data for GAIA and Browsecomp, reflecting a deeper understanding of IS tasks from heuristic to formalized definitions [2][7] Group 1: Formalization and Methodology - WebShaper proposes a formalized model for IS tasks based on set theory, introducing the concept of Knowledge Projection (KP) to control reasoning paths and task complexity [13][14] - The formalization allows for precise control over reasoning complexity and logical structure, aligning information structure with reasoning structure to minimize errors in data synthesis [10][16] - The process begins with pre-constructed seed tasks, which are expanded into final synthesized data through a dedicated Expander module, ensuring broad coverage and task correctness [18][25] Group 2: Data Generation and Training - The article emphasizes the importance of systematically constructing high-quality training data to enhance the information retrieval capabilities of intelligent agents [9] - WebShaper's approach transitions from an "information-driven" synthesis paradigm to a "formalization-driven" one, enabling broader task coverage and knowledge generation [15][31] - The training of agents is conducted using supervised fine-tuning (SFT) combined with reinforcement learning strategies, resulting in 5,000 training trajectories and significant performance improvements on the GAIA benchmark [26][31] Group 3: Performance and Comparisons - WebShaper's performance surpasses that of closed-source models, with the highest score of 60.1 compared to 40.7 for GPT4.1 and 58.2 for Claude Sonnet4 [30] - The article highlights that the task-solving capabilities of WebShaper require more agent actions compared to baseline data, indicating a higher complexity in the tasks generated [32] Group 4: Implications and Future Directions - The formalized task synthesis approach of WebShaper can be extended to more complex tasks beyond IS, suggesting a broader application in AI research [35] - The article advocates for open-source data and models as a means to achieve high performance in AI tasks, promoting a collaborative ecosystem for advancing AI research [34]
全球首个全链式空间天气AI预报模型“风宇”!国家卫星气象中心牵头,联合南昌大学、华为共同研发
机器之心· 2025-07-29 09:58
Core Viewpoint - The article highlights the development and significance of the "Fengyu" model, which is the world's first full-chain artificial intelligence forecasting model for space weather, enhancing China's capabilities in space weather monitoring and prediction [2][9]. Group 1: Importance of Space Weather Monitoring - The current solar activity poses threats to satellites, aircraft, and critical ground infrastructure due to unpredictable events like solar flares, likened to an invisible "cosmic tsunami" [4]. - Traditional forecasting methods rely heavily on numerical models, which are complex and time-consuming, making real-time responses challenging [5]. Group 2: Introduction of the "Fengyu" Model - The "Fengyu" model was officially launched on July 26, 2025, at the World Artificial Intelligence Conference, developed by the National Satellite Meteorological Center in collaboration with Nanchang University and Huawei [8]. - The model integrates physical models, numerical forecasting, and artificial intelligence, significantly improving China's space weather forecasting capabilities [9]. Group 3: Technological Innovations of the "Fengyu" Model - The model features a pioneering "chain training structure" that integrates forecasting processes into a collaborative system, addressing the limitations of previous isolated AI models [12]. - It introduces a unique "intelligent coupling optimization mechanism" that allows for real-time collaborative optimization among different regional models, enhancing forecasting accuracy [14]. - The model is built on the MindSpore Science suite and Ascend hardware, achieving superior training efficiency and predictive accuracy compared to traditional platforms [11][18]. Group 4: Performance and Applications - The "Fengyu" model has demonstrated exceptional short-term forecasting capabilities, maintaining prediction errors for global electron density within approximately 10% during significant geomagnetic storm events [25]. - It can guide spacecraft design and operational management, optimizing satellite fuel usage and flight posture in response to predicted space weather changes [27][28]. Group 5: Future Directions - The release of the "Fengyu" model marks a significant advancement in space weather monitoring and prediction, serving as a successful case in the AI for Science domain [30]. - Future developments aim to deploy AI capabilities directly on satellites for autonomous decision-making, representing a critical evolution in aerospace AI applications [31][32].
LeCun出手,造出视频世界模型,挑战英伟达COSMOS
机器之心· 2025-07-29 09:58
Core Viewpoint - The article discusses the development and advantages of a new video world model called DINO-world, which aims to improve the efficiency and effectiveness of predicting future frames in various environments, particularly in the context of artificial intelligence and machine learning [9][10]. Data Challenges - The acquisition of large-scale, high-quality video datasets is costly, especially when action annotations are required. Current successful applications of world models are limited to specific fields like autonomous driving and video games [5]. - Accurately modeling physical laws and behaviors in unconstrained, partially observable environments remains a significant challenge, even for short time scales. Advanced pixel-based generative models consume enormous computational resources, with training times reaching up to 22 million GPU hours for models like COSMOS [6]. Model Development - DINO-world utilizes a frozen visual encoder (DINOv2) to pre-train the video world model in a latent space, followed by fine-tuning with action data for planning and control [9]. - The architecture of DINO-world significantly reduces resource consumption during both training and inference phases compared to current state-of-the-art models [10]. Training and Evaluation - DINO-world was trained on a large dataset of approximately 60 million uncleaned network videos, enabling it to learn transferable features across different domains [11]. - In the VSPW segmentation prediction task, DINO-world achieved a mean Intersection over Union (mIoU) improvement of 6.3% when predicting future frames, outperforming the second-best model [13]. Methodology - The model employs a frame encoder that does not directly model pixels but instead uses latent representations based on video patches, which significantly lowers the computational cost of training the predictor [19]. - The training objective is set as "next frame prediction," allowing for efficient parallelization and focusing on the most relevant tokens for loss calculation [27]. Action-Conditioned Fine-Tuning - DINO-world can be adapted for action-conditioned tasks by incorporating an action module that updates the query vector based on the corresponding actions, which can be trained on a small dataset of action-conditioned trajectories [30][33]. Experimental Results - DINO-world demonstrated superior performance in dense prediction tasks across various datasets, including Cityscapes, VSPW, and KITTI, validating the effectiveness of the proposed paradigm [37][38]. - The model's performance in intuitive physics tests showed a strong understanding of physical behaviors, comparable to larger models like V-JEPA [40][41]. Planning Evaluation - The action-conditioned model was trained on offline trajectories, showing significant performance improvements compared to models trained from scratch, particularly in more complex environments [44].
这家国内公司,在给具身智能技术栈做「通解」
机器之心· 2025-07-29 09:58
Core Viewpoint - The article discusses the advancements in embodied intelligence in robotics, highlighting the integration of perception, cognition, and manipulation capabilities to enable robots to perform complex tasks in real-world environments [1][2][3]. Group 1: Embodied Intelligence and Robotics - Embodied intelligence is a highly discussed area in AI, aiming to give large models a physical body to perceive the real world and perform complex tasks [1]. - The recent WAIC 2025 conference showcased various robots demonstrating advanced capabilities, indicating progress towards a "robotic ChatGPT" [2][5]. - Robots are now capable of handling soft objects and performing tasks like folding clothes with precision, showcasing their improved dexterity [6][7][8]. Group 2: Technological Advancements - Robots can autonomously identify and categorize a wide range of real-life objects, including reading handwritten labels and executing tasks based on natural language instructions [9][11][12]. - The technology behind these robots, developed by Mech-Mind, includes a multimodal large model, high-precision 3D cameras, and bio-inspired dexterous hands, enabling them to understand and interact with their environment effectively [23][26][29]. - The robots can also perform complex decision-making tasks, such as sorting and classifying items at high speeds, demonstrating their versatility and efficiency [13][21][22]. Group 3: Market Position and Future Prospects - Mech-Mind has established itself as a leader in the robotics field, with over 15,000 units deployed globally and a significant market share in the domestic sector [45][46]. - The company aims to enhance robots' understanding, reasoning, and learning capabilities, facilitating their application across various industries, including logistics, manufacturing, and household tasks [47]. - The ongoing investment from major tech companies in embodied intelligence indicates a strong belief in the future potential of robotics in various sectors [41][42].
ACL首届博士论文奖公布,华人学者李曼玲获荣誉提名
机器之心· 2025-07-29 09:58
Core Insights - The article discusses the announcement of the ACL's new award for outstanding doctoral dissertations in computational linguistics, highlighting the significance of the award and its impact on the field of natural language processing [1][2][4]. Group 1: Award Details - The inaugural recipient of the ACL Doctoral Dissertation Award is Sewon Min from the University of Washington, recognized for her thesis titled "Rethinking Data Use in Large Language Models" [2][4]. - The award committee emphasized that Min's research provides critical insights into the behavior and capabilities of large language models, particularly in the area of in-context learning [4][14]. Group 2: Research Contributions - Min's dissertation discusses the understanding and advancement of large language models, focusing on their use of extensive training datasets [14]. - She demonstrates that the in-context learning ability of these models is largely determined by the content learned from training data [15]. - Min introduces a new class of language models called nonparametric language models, which utilize training data as a storage mechanism to retrieve information, enhancing accuracy and updatability [16][18]. Group 3: Other Nominated Works - The article also mentions three additional nominees for the award: Manling Li from the University of Illinois Urbana-Champaign, Ashish Sharma from the University of Washington, and Thomas Rishi Sherborne from the University of Edinburgh [8][20]. - Manling Li's work focuses on event-centric multimodal knowledge acquisition, proposing methods to transition from entity-centric to event-centric knowledge extraction [26][30]. - Ashish Sharma explores human-AI collaboration to improve mental health support, demonstrating how AI can enhance empathy in conversations and assist users in self-help interventions [45][51]. - Thomas Rishi Sherborne's research addresses cross-lingual transfer for semantic parsing, proposing strategies for effective adaptation of semantic parsers to new languages [62][64].
从数字人到「有温度的」机器人,京东把 AI 深度应用的路线图「摸透」了
机器之心· 2025-07-29 07:44
Core Viewpoint - The article discusses the evolution of JD's AI model, now branded as "JoyAI," emphasizing its transition from a focus on model training to practical applications in various industries, marking the beginning of a new phase in AI development [1][2][3]. Group 1: JoyAI Model Upgrade - JD has upgraded its AI model brand to "JoyAI," which now encompasses a wide range of modalities including language, voice, image, and video, showcasing advancements in AI technology [6][7]. - The upgraded JoyAI features models ranging from 3 billion to 750 billion parameters, achieving a 30% increase in inference efficiency and a 70% reduction in training costs [7][8]. Group 2: Application and Impact - JoyAI has been applied in hundreds of scenarios within JD, with a significant increase in model usage during the recent "618" shopping festival, showing a 130% increase compared to the previous "11.11" event [11][12]. - The digital human technology based on JoyAI has been widely adopted, with over 20,000 brands utilizing it, demonstrating its commercial viability [16][19]. Group 3: Future of AI in Industry - The article highlights the potential growth of China's core AI industry, projected to reach a market value of $140 billion by 2030, indicating a shift from AI as a supplementary tool to a critical component in various industries [14]. - JD's AI technology is not only advancing in digital humans but also expanding into logistics, retail, and healthcare, with applications that enhance productivity and service quality [44][53]. Group 4: JoyInside Platform - JD introduced the JoyInside platform, which integrates AI capabilities into smart hardware, allowing for personalized interactions and enhancing user engagement [29][32]. - JoyInside has been adopted by numerous companies, including educational and industrial applications, showcasing its versatility across different sectors [33][34]. Group 5: Investment and Ecosystem Development - JD is actively investing in the embodied intelligence sector, having recently invested in four companies, indicating a strategic focus on building a robust ecosystem around AI technologies [41][42]. - The company is opening its JoyInside platform to various brands, providing a comprehensive solution that includes software, hardware, and content integration [43][56].
开启RL Scaling新纪元,siiRL开源:完全分布式强化学习框架,支持超千卡规模高效训练
机器之心· 2025-07-29 07:44
Core Insights - The article emphasizes the importance of overcoming scalability bottlenecks in Reinforcement Learning (RL) frameworks as a key to unlocking advanced AI reasoning capabilities and achieving stronger general intelligence [2][31] - The introduction of the siiRL framework by the Shanghai Institute of Intelligent Technology is highlighted as a significant advancement in supporting large-scale and efficient RL training [3][31] Group 1: Scalability Challenges - Traditional RL frameworks often rely on a centralized controller architecture, which leads to performance bottlenecks and memory overflow when scaled to hundreds or thousands of GPUs [8][9] - The centralized design is manageable at smaller scales but becomes a critical limitation as the system expands, resulting in high I/O and communication overhead [9][10] Group 2: siiRL Framework Features - siiRL employs an innovative multi-controller paradigm and fully distributed architecture, effectively removing the central node and distributing tasks across all worker nodes [11][31] - The framework demonstrates near-linear scalability, achieving a 7-fold increase in end-to-end training throughput and maintaining performance even at 1024 GPU scales [21][31] - The architecture includes three core components: DAG Planner for workflow definition, DAG Workers for task execution, and Data Coordinators for managing data flow [13][14][15] Group 3: Performance Validation - Experimental results show that siiRL outperforms baseline frameworks, achieving up to 2.62 times higher throughput in data-intensive scenarios [19][26] - In long-context tasks, the performance advantage of siiRL increases significantly as context length grows, demonstrating its efficiency in handling larger data communication volumes [26][27] - Convergence tests indicate that performance improvements do not compromise model accuracy, with reward and entropy curves closely aligning with baseline frameworks [28][31] Group 4: Future Plans - The framework is designed to support complex multi-agent systems, with plans to enhance compatibility with multi-agent reinforcement learning (MARL) algorithms and improve agent-environment interaction mechanisms [29][31]