Workflow
LLM脑损伤假说
icon
Search documents
短视频刷多了AI也会变蠢!“年度最令人不安的论文”
量子位· 2025-11-16 07:20
Core Insights - The article discusses the phenomenon of "Brain Rot" in AI, indicating that exposure to low-quality data can lead to irreversible cognitive decline in large language models (LLMs) [2][13][26] - The research highlights that even after retraining with high-quality data, the damage caused by low-quality data cannot be fully repaired, suggesting a permanent cognitive shift [4][26][27] Research Findings - The study introduces the "LLM Brain Rot Hypothesis," exploring whether LLMs experience cognitive decline similar to humans when exposed to low-quality data [8][13] - Two dimensions were used to define "garbage data": M1 focuses on engagement metrics (short, high-traffic content), while M2 assesses semantic quality (clickbait and conspiracy theories) [11][12] - The models tested showed a 23% decline in reasoning ability and a 30% decrease in long-context memory after exposure to garbage data [6][14] Cognitive Impact - The study found that LLMs exhibit cognitive decline akin to "Brain Rot," with significant negative effects on safety and personality traits, particularly from M1 data [14][19] - A dose-effect relationship was observed, where increased exposure to garbage data correlates with greater cognitive damage [15] Repair Attempts - Attempts to repair the cognitive damage through external feedback and large-scale fine-tuning were unsuccessful, with models failing to regain baseline performance [25][26] - The research indicates that LLMs lack the ability to self-correct effectively, unlike humans who can mitigate cognitive decline through various means [24][27] Industry Implications - The findings emphasize the importance of data quality during the pre-training phase, suggesting that the industry should focus on data selection as a safety issue [28] - Implementing cognitive assessments for LLMs, such as ARC and RULER benchmarks, is recommended to prevent long-term exposure to low-quality data [29] - The study suggests prioritizing the exclusion of short, high-engagement content from training datasets to enhance model performance [29]