Workflow
Anthropic
icon
Search documents
X @Anthropic
Anthropic· 2025-08-22 16:19
If you’re interested in joining us to work on these and related issues, you can apply for our Research Engineer/Scientist role (https://t.co/x3G4F5qVWv) on the Alignment Science team. ...
X @Anthropic
Anthropic· 2025-08-22 16:19
There’s plenty of work to be done to make the classifiers even more accurate and effective. In the future, they might even be able to remove data relevant to misalignment risks (scheming, deception, and so on), as well as CBRN risks. ...
X @Anthropic
Anthropic· 2025-08-22 16:19
New Anthropic research: filtering out dangerous information at pretraining.We’re experimenting with ways to remove information about chemical, biological, radiological and nuclear (CBRN) weapons from our models’ training data without affecting performance on harmless tasks. https://t.co/YUBlLKIL2c ...
X @Anthropic
Anthropic· 2025-08-21 16:33
We’re also announcing a new Higher Education Advisory Board, which helps guide how Claude is used in teaching, learning, and research.Read more about the courses and the Board: https://t.co/TorRcYMHnd ...
X @Anthropic
Anthropic· 2025-08-21 16:33
We’ve made three new AI fluency courses, co-created with educators, to help teachers and students build practical, responsible AI skills.They’re available for free to any institution. https://t.co/nK2D3W5YcU ...
X @Anthropic
Anthropic· 2025-08-21 10:36
This demonstrates what's possible when government expertise meets industry capability. NNSA understands nuclear risks better than any company could; we have the technical capacity to build the safeguards. ...
X @Anthropic
Anthropic· 2025-08-21 10:36
We don't need to choose between innovation and safety. With the right public-private partnerships, we can have both. We’re sharing our approach with @fmf_org members so any AI company can implement similar protections.Read more: https://t.co/HxgrIwK8n9 ...
X @Anthropic
Anthropic· 2025-08-21 10:36
We partnered with @NNSANews to build first-of-their-kind nuclear weapons safeguards for AI.We've developed a classifier that detects nuclear weapons queries while preserving legitimate uses for students, doctors, and researchers. https://t.co/PlZ55ot74l ...
X @Anthropic
Anthropic· 2025-08-20 18:14
RT Claude (@claudeai)Claude Code is now available on Team and Enterprise plans.Flexible pricing lets you mix standard and premium Claude Code seats across your organization and scale with usage. https://t.co/co3UT5PcP3 ...
X @Anthropic
Anthropic· 2025-08-15 20:41
AI Model Research - Anthropic interpretability researchers are discussing looking into the mind of an AI model [1] Interpretability - The discussion highlights the importance of understanding AI model's decision-making processes [1]