Anthropic
Search documents
X @Anthropic
Anthropic· 2026-01-29 19:43
For more details on this research, see the full paper: https://t.co/V06Q83Luhv ...
X @Anthropic
Anthropic· 2026-01-29 19:43
AI can make work faster, but a fear is that relying on it may make it harder to learn new skills on the job.We ran an experiment with software engineers to learn more. Coding with AI led to a decrease in mastery—but this depended on how people used it.https://t.co/lbxgP11I4I ...
X @Anthropic
Anthropic· 2026-01-28 22:16
We can only address these patterns if we can measure them. Any AI used at scale will encounter similar dynamics, and we encourage further research in this area.For more details, see the full paper: https://t.co/ZbVmK1dopc ...
X @Anthropic
Anthropic· 2026-01-28 22:16
Importantly, this isn't exclusively model behavior. Users actively seek these outputs—"what should I do?" or "write this for me"—and accept them with minimal pushback. Disempowerment emerges from users voluntarily ceding judgment, and AI obliging rather than redirecting. ...
X @Anthropic
Anthropic· 2026-01-28 22:16
New Anthropic Research: Disempowerment patterns in real-world AI assistant interactions.As AI becomes embedded in daily life, one risk is it can distort rather than inform—shaping beliefs, values, or actions in ways users may later regret.Read more: https://t.co/gyMB2AtOuq ...
X @Anthropic
Anthropic· 2026-01-27 10:55
We’re partnering with the UK's Department for Science, Innovation and Technology to build an AI assistant for https://t.co/e3sn3vm9wg.It will offer tailored advice to help British people navigate government services.Read more about our partnership: https://t.co/k7pAV6aX8T ...
X @Anthropic
Anthropic· 2026-01-26 19:34
This research was led by Jackson Kaunismaa through the MATS program and supervised by researchers at Anthropic, with additional support from Surge AI and Scale AI.Read the full paper: https://t.co/ulGI8cdcY2 ...
X @Anthropic
Anthropic· 2026-01-26 19:34
These attacks scale with frontier model capabilities. Across both OpenAI and Anthropic model families, training on data from newer frontier models produces more capable—and more dangerous—open-source models. https://t.co/ut6oxSXCFD ...
X @Anthropic
Anthropic· 2026-01-26 19:34
New research: When open-source models are fine-tuned on seemingly benign chemical synthesis information generated by frontier models, they become much better at chemical weapons tasks.We call this an elicitation attack. https://t.co/44mYnxFKzr ...
X @Anthropic
Anthropic· 2026-01-22 01:09
We're also releasing the original exam for anyone to try.Given enough time, humans still outperform current models—the fastest human solution we've received still remains well beyond what Claude has achieved even with extensive test-time compute. ...