X @Anthropic
Anthropic·2025-07-08 22:12
Recent LLMs don't fake alignment in the situation we studied. We’re investigating whether this is the same in more realistic settings (like when models aren’t directly told they’re in a training scenario). ...