Anthropic has introduced a new feature for its Claude Opus 4 and 4.1 AI models that allows them to end conversations with users in rare, extreme cases of harmful interactions, such as requests for illicit content or solicitations for violence. This measure aims to enhance AI welfare by managing risks and will only be used as a last resort after multiple attempts to redirect the conversation have failed.
This is an ainewsarticles.com news flash; the original news article can be found here: Read the Full Article…