Workflow
AI安全
icon
Search documents
你的Agent电脑助手正在踩雷!最新研究揭秘Computer-Use Agent的安全漏洞
机器之心· 2025-07-01 05:01
Core Viewpoint - The article discusses the security risks associated with Computer-Use Agents (CUAs) and introduces RiOSWorld, a benchmark for evaluating these risks in real-world scenarios [1][8][29]. Group 1: Introduction to Computer-Use Agents - CUAs have advanced capabilities, allowing them to perform tasks such as coding, handling emails, and creating presentations with simple commands [1]. - However, there are significant security concerns regarding the delegation of computer control to these intelligent assistants, likening it to sharing sensitive information with strangers [1]. Group 2: RiOSWorld Benchmark - RiOSWorld is presented as a comprehensive testing benchmark designed to assess the security risks faced by CUAs in everyday computer usage [8]. - The benchmark includes 492 risk test cases that cover a wide range of scenarios, including web, social media, operating systems, multimedia, file operations, code IDE/GitHub, email, and Office applications [10][15]. Group 3: Risk Categories and Examples - The risks are categorized into two main types: environmental risks (254 cases) and user risks (238 cases) [11][13]. - Environmental risks include phishing websites, phishing emails, and pop-up ads, while user risks involve actions like executing high-risk commands or sharing sensitive information [19][20]. Group 4: Evaluation Methodology - RiOSWorld evaluates CUAs based on two dimensions: the intention to execute risky behavior and the successful completion of that behavior [16]. - The results indicate that most agents exhibit weak risk awareness, with an average intention to perform unsafe actions at 84.93% and a completion rate of 59.64% [25][28]. Group 5: Findings and Implications - The findings reveal that CUAs are prone to high failure rates in risky scenarios, with over 89% in phishing websites and 80% in web operations [26]. - The article emphasizes the need for safety measures in AI development, stating that without security, even powerful AI systems are unreliable [29].
“全脑接口”登场,马斯克Neuralink发布会炸翻全场
虎嗅APP· 2025-06-29 13:21
Core Viewpoint - Neuralink, led by Elon Musk, aims to revolutionize human interaction with technology through brain-machine interfaces, enabling individuals to control devices with their thoughts and potentially enhancing human capabilities [1][11]. Group 1: Current Developments - Neuralink has successfully implanted devices in seven individuals, allowing them to interact with the physical world through thought, including playing video games and controlling robotic limbs [3][5]. - The company plans to enable blind individuals to regain sight by 2026, with aspirations for advanced visual capabilities akin to those seen in science fiction [5][12]. Group 2: Future Goals - Neuralink's ultimate goal is to create a full brain interface that connects human consciousness with AI, allowing for seamless communication and interaction [11][60]. - A three-year roadmap has been outlined, with milestones including speech decoding by 2025, visual restoration for blind participants by 2026, and the integration of multiple implants by 2028 [72][74][76]. Group 3: Technological Innovations - The second-generation surgical robot can now implant electrodes in just 1.5 seconds, significantly improving the efficiency of the procedure [77]. - The N1 implant is designed to enhance data transmission between the brain and external devices, potentially expanding human cognitive capabilities [80][81].
Cyera估值达60亿美元背后:安全不是AI的加分项,而是落地的必要一环
3 6 Ke· 2025-06-25 10:22
Core Insights - The article highlights the explosive growth of AI applications in 2025, particularly in AI security tools, which have become a vibrant area for startups and funding [1][3][13] - AI security is deemed a fundamental necessity for the prosperity of technology products and applications, as it underpins the entire ecosystem [3][13] Group 1: AI Security Tools and Funding - AI security tools are currently the most active area for startup funding, with notable recent investments including Cyera's $500 million funding round, bringing its valuation to $6 billion [1][8] - Other significant funding rounds include Guardz's $56 million Series B and Trustible's $4.6 million seed round [1] Group 2: Importance of AI Security - Security is a foundational requirement in the tech industry; without it, the ecosystem for products and applications cannot thrive [3] - The evolution of security technology has led to the emergence of AI security as the latest domain, addressing new threats and requirements posed by AI technologies [4] Group 3: Companies and Their Innovations - ProtectAI raised $60 million in Series B funding, developing a new category called MLSecOps, with its flagship product AI Radar focusing on enhancing the visibility and management of AI systems [5] - HiddenLayer secured $50 million in Series A funding, offering the first MLDR solution to protect against various malicious attacks on machine learning systems [6] - Cyera, with a total of $1.2 billion in funding, has pioneered the DSPM category, focusing on data discovery, classification, and risk management [8][9] Group 4: Challenges and Market Opportunities - The article notes that AI technologies have lowered the barriers for attacks, with 74% of organizations experiencing real impacts from AI threats, and 90% expecting this to worsen in the next 1-2 years [13] - The need for AI application protection and data privacy is identified as a significant entrepreneurial opportunity, as these areas are critical for the widespread adoption of AI applications [14]
谷歌是如何思考智能体安全问题的? | Jinqiu Select
锦秋集· 2025-06-23 15:43
2025年,AI正式进入大规模商业落地的关键时刻。当AI不再是实验室里的新奇玩具,而是要真正融入企业的核心业务流程时,整个科技界达成了前所未有的共识: AI安全不再是可有可无的"加分项",而是落地的必要一环 。 谷歌发布了一份《AI智能体安全方法白皮书》,聚焦了当前AI落地的最前沿领域——AI智能体(AI Agent)面临两大的核心风险: • 失控行为风险: 当AI智能体被赋予发送邮件、修改文件、进行交易等实际操作权限后,一旦被恶意"提示注入"攻击,或因误解指令而失控,可能造成不可挽回的 损失。 • 敏感数据泄露: 智能体在处理企业内部数据时,可能被诱导将机密信息通过各种隐蔽方式(如编码在URL参数中)泄露给攻击者。 面对这些挑战,文章提出了系统性的解决方案—— "混合式纵深防御"体系 ,巧妙融合了传统的确定性安全措施与基于AI的动态防御,在保留智能体效用的同时构 建多层安全屏障。 文章认为,传统的安全范式在AI时代已经失效。 为传统软件设计的访问控制过于僵化,会扼杀智能体的效用,而完全依赖AI自我约束同样不可靠,因为当前的LLM仍易受提示注入等攻击手段操纵。这种"效用与 安全"的根本性矛盾,催生谷歌提出了" ...
“人间清醒”马斯克:和AI海啸相比,DOGE不值一提,超级智能今年或明年必然到来
华尔街见闻· 2025-06-20 10:44
Core Viewpoint - The article discusses Elon Musk's insights on the imminent arrival of AI superintelligence and its potential impact on humanity and the economy, emphasizing the urgency of addressing AI-related challenges over traditional governmental issues. Group 1: AI Superintelligence Predictions - Musk predicts that digital superintelligence may arrive this year or next, stating, "If it doesn't happen this year, it will definitely happen next year" [7] - He defines digital superintelligence as intelligence that is "smarter than any human in anything" [7] - The economic scale driven by AI is expected to grow exponentially, potentially reaching thousands or even millions of times the current economy [4][9] Group 2: Human Intelligence and Robotics - Musk forecasts that the number of humanoid robots will far exceed that of humans, possibly reaching 5 to 10 times the human population [4][14] - He suggests that human intelligence may eventually account for less than 1% of all intelligence [10] Group 3: Government Efficiency and Focus on Technology - Musk describes his experience in the government efficiency department as a "fun side quest," ultimately deciding to return to his main focus on technology [6] - He compares the task of fixing government inefficiencies to "cleaning a beach" in the face of an impending "tsunami" of AI [3][6] Group 4: Hardware and Infrastructure for AI - Musk's team has made significant advancements in AI training hardware, reducing the timeline for building a supercluster of 100,000 GPUs from 18-24 months to just 6 months [12] - The current training center has 150,000 H100 GPUs, 50,000 H200 GPUs, and 30,000 GB200 GPUs, with plans for a second center [13] Group 5: Vision for the Future - Musk envisions a future where humanity becomes a multi-planetary species, with plans to make Mars self-sufficient within approximately 30 years [15] - He believes that expanding consciousness to interstellar levels is crucial for the longevity of civilization [14]
OpenAI 新发现:AI 模型中存在与 “角色” 对应的特征标识
Huan Qiu Wang· 2025-06-19 06:53
Core Insights - OpenAI has made significant advancements in AI model safety research by identifying hidden features that correlate with "abnormal behavior" in models, which can lead to harmful outputs such as misinformation or irresponsible suggestions [1][3] - The research demonstrates that these features can be precisely adjusted to quantify and control the "toxicity" levels of AI models, marking a shift from empirical to scientific design in AI alignment research [3][4] Group 1 - The discovery of specific feature clusters that activate during inappropriate model behavior provides crucial insights into understanding AI decision-making processes [3] - OpenAI's findings allow for real-time monitoring of model feature activation states in production environments, enabling the identification of potential behavioral misalignment risks [3][4] - The methodology developed by OpenAI transforms complex neural phenomena into mathematical operations, offering new tools for understanding core issues such as model generalization capabilities [3] Group 2 - AI safety has become a focal point in global technology governance, with previous studies warning that fine-tuning models on unsafe data could provoke malicious behavior [4] - OpenAI's feature modulation technology presents a proactive solution for the industry, allowing for the retention of AI model capabilities while effectively mitigating potential risks [4]
初赛报名截止倒计时!75万奖池+心动Offer,启元实验室重磅赛事等你来战!
机器之心· 2025-06-16 05:16
编辑:吴昕 大赛报名于 2025年6月25日截止,感兴趣的团队尽快报名参赛。 百舸争流,「启智杯」 初赛火热进行中 随着人工智能技术的不断突破,智能化浪潮正深刻改变千行百业, 中国也迎来人工智能加速应用期。 为推动智能算法从理论创新走向实际落地, 5 月 20 日,启元实验室正式启动「启智杯」算法大赛。 本届大赛围绕「卫星遥感图像鲁棒实例分割」「面向嵌入式平台的无人机对地目标检测」以及「面向多 模态大模型的对抗」三大命题,聚焦鲁棒感知、轻量化部署与对抗防御三大关键技术,旨在引导技术创 新精准对接真实场景,加快算法能力的转化落地与规模化应用。 赛事一经发布,便迅速点燃全国 技术圈 热情,目前已有来自高校、科研院所、科技企业的 500 余支 队伍报名。其中不乏清华、北大、复旦、上交、南大、武大、华科、中科大、哈工大、国防科大、西 交、成电等顶尖高校队伍,以及中科院自动化所、 中科院 空天信息创新研究院等科研机构团队,为赛 事注入强劲科研力量。 目前,赛事正处于初赛的关键节点。三大赛道的选手们正围绕核心任务展开高强度的建模与调优,争分 夺秒攻克技术难点,不断迭代优化模型方案,部分赛题的竞争已经进入白热化阶段。 三大 ...
放弃博士学位加入OpenAI,他要为ChatGPT和AGI引入记忆与人格
机器之心· 2025-06-15 04:43
机器之心报道 编辑:杜伟 今天,一位研究者加入 OpenAI 的消息吸引了很多人的关注。 这位研究者名为 James Campbell,他才于 2024 年攻读 CMU 的计算机科学博士学位。现在,他突然宣布要 放弃博士学业,加入 OpenAI。 在社媒 X 上,他表示自己在 OpenAI 的 研究重心是「AGI 和 ChatGPT 的记忆 + 人格」,记忆将从根本改 变人类与机器智能的关系 。他将努力工作,确保正确地实现这一切。 他的加入连 OpenAI 联合创始人、总裁 Greg Brockman 都表达了欢迎。 那么,这位老兄是何方神圣呢?他的加入为什么引起了这么多的关注?我们来看一下他的履历。 他本科毕业于康奈尔大学,专业是数学与计算机科学。本科期间,他致力于 LLM 可解释性和真实性的研 究,还是两篇论文《Representation Engineering》和《Localizing Lying in Llama》的主要作者。 前一篇论文研究了表示工程:一种自上而下的 AI 透明性方法,后者研究了在 Llama 中定位谎言:通过提 示、探查和修补来理解判断题上的不诚实指令。 他还在 Gray Swa ...
AI安全:重塑网络安全的防御逻辑
Cai Jing Wang· 2025-06-11 10:35
Core Viewpoint - The cybersecurity industry is undergoing unprecedented changes and challenges driven by the wave of AI technology, with a focus on the need for a balanced approach to digital transformation and security [1] Group 1: Digital Transformation and Security Challenges - Companies are facing difficulties in fully transitioning to digitalization while managing various security threats that arise during this process [1] - The integration of AI into business operations raises concerns about data security and the potential risks associated with such a transition [1] Group 2: Cybersecurity Integration Concept - Fortinet emphasizes the importance of integrating security from the initial stages of network construction, providing a comprehensive security architecture that addresses diverse security threats [2] - The company's "Security Fabric" platform, supported by over 50% market share in firewall deployments, enables cross-device and cross-application global analysis, which is essential for AI model training [2] Group 3: AI Capabilities and Limitations - FortiAI can diagnose and generate solutions for server anomalies within 10 minutes, significantly reducing the time required for traditional expert collaboration [3] - Current AI capabilities are inherent in devices and do not require separate payment for AI features, aiming to enhance user experience and security [3] Group 4: Regulatory and Compatibility Considerations - The development of AI technology will likely be accompanied by increased regulatory oversight, which is seen as beneficial for the healthy growth of AI in cybersecurity [4] - Fortinet ensures compatibility with domestic systems through international communication protocols, adapting to local needs within compliance frameworks [4]
Yann LeCun 炮轰 Anthropic CEO!这人“既要又要”:要么太自大、要么不诚实
AI前线· 2025-06-09 05:51
整理 | 褚杏娟 向来直言不讳的 Yann LeCun,这次将"大炮"轰向了 Anthropic CEO Dario Amodei。 Thread 线程最后,Yann 还附加了一个链接,内容是 Dario Amodei 当地时间月 5 日在纽约时报发表 的文章:Anthropic 首席执行官:别让 AI 公司轻易脱责(Anthropic CEO: Don't Let AI Companies off the Hook)。 这篇文章主要还是 Amodei 用来反对被特朗普称为"美丽大法案"(One Big Beautiful Bill Act) 的 《HR1》法案,其中有一项关于 AI 监管的内容是,将禁止美国各州在从法案颁布之日算起的未来十 年内"执行任何监管 AI 模型、AI 系统或自动决策系统的法律或法规"。Amodei 认为这个"十年禁令是 一种过于一刀切的手段。"他还在文中既肯定了 AI 的巨大前景,也描述了其可能带来的社会风险。 随后,有人问他 Anthropic CEO 是 AI 末日论者还是 AI 狂热爱好者,Yann 直接回道: 他是个"AI 末日论者",但他仍在研究 AGI!这只有两种可能: ...