Bookmark Article
Anthropic’s Claude Opus 4 and 4.1 models can now end conversations in extreme cases of harmful user interactions, such as soliciting sexual content involving minors or promoting violence. This feature aims to manage AI welfare and will only be used as a last resort after failed redirection attempts, allowing users to start new conversations immediately without affecting others.