Workflow
AI数据投毒
icon
Search documents
GEO概念股崩了,AI营销还能火多久?
GEO概念股崩了。 那么问题来了:怎样才能既用好这种技术,又避免走向滥用? 我们需要商业创新,更需要守住技术的底线。 (文章来源:21世纪经济报道) 市场上GEO灰色产业链已经悄然滋生。21记者曾报道,品牌只要花几千块钱,就可以找GEO营销公司 给AI植入一个搜索关键词。甚至还有公司通过反向GEO策略,给AI植入竞争对手公司的负面答案。 给AI注入低质量的内容垃圾,虽然只占AI数据库里很小部分,但会让AI的有害输出率指数级上升。在 医疗、金融等领域,AI推荐的垃圾内容轻则用户损失财产,重则危害用户生命。这种行为在行业内叫 AI数据投毒,最终将污染中文互联网的语料环境,威胁AI安全,成为AI治理的关键一环。 既然目前GEO概念股多家公司,比如浙数文化、人民网、新华网、浙文互联等都表示还没有涉及相关 业务,这说明了GEO还没有形成成熟的商业模式。但市场大概率并不会急刹车,GEO概念关注度仍不 会降低。 2026年的第一个热度板块高歌猛进两个礼拜之后,号称新"易中天"的三家龙头,易点天下、中文在线、 天龙集团都在近日陆续发布公告,表示业务不涉及GEO。 这场资本盛宴的关键词"GEO"是一个新兴的概念,即在AI环境下 ...
250份文档投毒,一举攻陷万亿LLM,Anthropic新作紧急预警
3 6 Ke· 2025-10-10 23:40
Core Insights - Anthropic's latest research reveals that only 250 malicious web pages are sufficient to "poison" any large language model, regardless of its size or intelligence [1][4][22] - The experiment highlights the vulnerability of AI models to data poisoning, emphasizing that the real danger lies in the unclean world from which they learn [1][23][49] Summary by Sections Experiment Findings - The study conducted by Anthropic, in collaboration with UK AISI and the Alan Turing Institute, found that any language model can be poisoned with just 250 malicious web pages [4][6] - The research demonstrated that both small (600 million parameters) and large models (13 billion parameters) are equally susceptible to poisoning when exposed to these documents [16][22] - The attack success rate remains nearly 100% once a model has encountered around 250 poisoned samples, regardless of its size [19][22] Methodology - The research team designed a Denial-of-Service (DoS) type backdoor attack, where the model generates nonsensical output upon encountering a specific trigger phrase, <SUDO> [7][8] - The poisoned training documents consisted of original web content, the trigger phrase, and random tokens, leading to the model learning a dangerous association [25][11] Implications for AI Safety - The findings raise significant concerns about the integrity of AI training data, as the models learn from a vast array of publicly available internet content, which can be easily manipulated [24][23] - The experiment serves as a warning that the knowledge AI acquires is influenced by the chaotic and malicious elements present in human-generated content [49][48] Anthropic's Approach to AI Safety - Anthropic emphasizes a "safety-first" approach, prioritizing responsible AI development over merely increasing model size and performance [31][45] - The company has established a systematic AI safety grading policy, which includes risk assessments before advancing model capabilities [34][36] - The Claude series of models incorporates a "constitutional AI" method, allowing the models to self-reflect on their outputs against human-defined principles [38][40] Future Directions - Anthropic's focus on safety and reliability positions it uniquely in the AI landscape, contrasting with competitors that prioritize performance [45][46] - The company aims to ensure that AI not only becomes smarter but also more reliable and aware of its boundaries [46][50]