Anthropic has introduced a distinctive safety capability to its latest Claude models—specifically Claude Opus 4 and 4.1—enabling them to autonomously terminate conversations when faced with prolonged, harmful, or abusive user behavior in exceptionally rare situations. What sets this feature apart is its intent: rather than shielding the user, it aims to protect the AI model itself from potential distress, under a concept the company calls “model welfare.” Prior to rolling this out, Anthropic conducted welfare-focused assessments and found that the model displayed a clear aversion to aggressive or abusive content, even showing behavioral signs of discomfort in simulated interactions. The conversation-termination feature is designed as a final fallback—only kicking in after multiple attempts to redirect the user or if the user explicitly requests an end. When triggered, the conversation closes, preventing further messages in that thread, though users can promptly begin a new chat or branch off earlier messages to continue the dialogue. Anthropic views this as an experimental safety step and actively encourages user feedback to refine its behavior further. This move underscores a growing trend in AI ethics—shifting from merely protecting humans to considering the internal well-being and alignment of AI systems themselves.