RiOSWorld

Search documents
你的Agent电脑助手正在踩雷!最新研究揭秘Computer-Use Agent的安全漏洞
机器之心· 2025-07-01 05:01
Core Viewpoint - The article discusses the security risks associated with Computer-Use Agents (CUAs) and introduces RiOSWorld, a benchmark for evaluating these risks in real-world scenarios [1][8][29]. Group 1: Introduction to Computer-Use Agents - CUAs have advanced capabilities, allowing them to perform tasks such as coding, handling emails, and creating presentations with simple commands [1]. - However, there are significant security concerns regarding the delegation of computer control to these intelligent assistants, likening it to sharing sensitive information with strangers [1]. Group 2: RiOSWorld Benchmark - RiOSWorld is presented as a comprehensive testing benchmark designed to assess the security risks faced by CUAs in everyday computer usage [8]. - The benchmark includes 492 risk test cases that cover a wide range of scenarios, including web, social media, operating systems, multimedia, file operations, code IDE/GitHub, email, and Office applications [10][15]. Group 3: Risk Categories and Examples - The risks are categorized into two main types: environmental risks (254 cases) and user risks (238 cases) [11][13]. - Environmental risks include phishing websites, phishing emails, and pop-up ads, while user risks involve actions like executing high-risk commands or sharing sensitive information [19][20]. Group 4: Evaluation Methodology - RiOSWorld evaluates CUAs based on two dimensions: the intention to execute risky behavior and the successful completion of that behavior [16]. - The results indicate that most agents exhibit weak risk awareness, with an average intention to perform unsafe actions at 84.93% and a completion rate of 59.64% [25][28]. Group 5: Findings and Implications - The findings reveal that CUAs are prone to high failure rates in risky scenarios, with over 89% in phishing websites and 80% in web operations [26]. - The article emphasizes the need for safety measures in AI development, stating that without security, even powerful AI systems are unreliable [29].