Workflow
AI系统训练
icon
Search documents
Anthropic指控AI公司蒸馏剽窃,马斯克硬刚“贼喊抓贼”
Sou Hu Cai Jing· 2026-02-25 10:13
Core Viewpoint - Anthropic accuses three leading Chinese AI companies, DeepSeek, Moonshot, and MiniMax, of infringing on its Claude model capabilities through fraudulent accounts and proxy services, utilizing a technique known as "model distillation" to enhance their own models [3][4]. Group 1: Allegations of Model Theft - Anthropic claims that the Chinese AI companies used fraudulent accounts to access Claude, generating over 16 million interactions, which they argue violates service terms and access restrictions [3][4]. - The three companies are accused of employing similar methods to access Claude's capabilities, particularly focusing on agentic reasoning, tool usage, and coding abilities [4]. Group 2: Specific Interactions and Patterns - DeepSeek engaged in over 150,000 interactions, focusing on extracting Claude's reasoning capabilities across diverse tasks, indicating coordinated efforts to avoid detection [5]. - Moonshot AI recorded over 3.4 million interactions, targeting agentic reasoning, tool usage, and data analysis, aiming to reconstruct Claude's reasoning pathways [5]. - MiniMax had the largest scale with over 13 million interactions, specifically targeting agent coding and tool usage, demonstrating adaptability by redirecting traffic to capture new features [5]. Group 3: Legal and Ethical Implications - The allegations raise questions about the legality of model distillation and the ethical considerations surrounding AI training, as many large language models are trained on publicly available internet data without explicit consent from original authors [7][8]. - There is an ongoing debate regarding the ownership of synthetic data and compliance issues related to training, particularly for open-source models [8]. Group 4: National Security and Export Controls - Anthropic's accusations highlight concerns over national security, suggesting that illegal distillation could undermine U.S. control over advanced AI technology exports [9]. - Current U.S. export controls primarily focus on hardware rather than large language model API access, indicating a gap in regulatory measures [9]. Group 5: Developer Responsibilities and Compliance - Developers using large language models must ensure their training processes are secure and compliant, maintaining clear records of training data sources and adhering to service terms [10][11]. - Anthropic is investing in defensive technologies to detect "distillation attack" patterns and is implementing protective measures to reduce the effectiveness of illegal distillation while maintaining legitimate user experience [11].