• Home
  • Ai
  • Ai News
  • Anthropic Will Now Let Claude End Conversations if the Topic Is Harmful or Abusive

Anthropic Will Now Let Claude End Conversations if the Topic Is Harmful or Abusive

Anthropic has added this ability in only the Claude Opus 4 and 4.1 AI models.

Anthropic Will Now Let Claude End Conversations if the Topic Is Harmful or Abusive

Photo Credit: Anthropic

Anthropic said certain conversation topics can distress Claude AI models

Click Here to Add Gadgets360 As A Trusted Source As A Preferred Source On Google
Highlights
  • Anthropic says the feature was developed for “potential AI welfare”
  • Once a conversation has ended, users can’t send any new messages
  • Users can, however, create a new chat and send messages there
Advertisement

Anthropic is rolling out the ability to end conversations in some of its artificial intelligence (AI) models. Announced last week, the new feature is designed as a protective measure for not the end user, but for the AI model itself. The San Francisco-based AI firm said the new capability was developed as part of its work on “potential AI welfare,” as conversations on certain topics can distress the Claude models. Notably, the company stated that the AI model will only use this option as a last resort after multiple attempts at redirection have failed.

Anthropic Introduces First Feature for AI Welfare

In a blog post, the AI firm announced that the ability to end conversations is being added to the Claude Opus 4 and 4.1 AI models. Explaining the need to develop the feature, the post said, “This ability is intended for use in rare, extreme cases of persistently harmful or abusive user interactions.”

Anthropic said the reason behind developing this feature is to protect the AI models from distressing conversations. The company said it ran a model welfare assessment for the Claude Opus 4 and found that the large language model (LLM) shows a strong aversion to harm. Some of these instances include user requests for sexual content involving minors, information for large-scale violence or acts of terror.

The company adds that Claude Opus 4 and 4.1 will only use this capability as a last resort. Before that, it will make multiple attempts to redirect the conversation and try to turn it into a productive session. The only other scenario where a conversation will be ended is if a user explicitly asks the chatbot to end the session. “The vast majority of users will not notice or be affected by this feature in any normal product use, even when discussing highly controversial issues with Claude,” the post stated.

Once a conversation has ended, the user will no longer be able to send new messages in that chat. However, they will still be able to start a new chat and begin a new session. Ending one chat will also not impact any other previous chats with the AI chatbot. Additionally, users will also be able to edit and retry the last message to create new branches of ended conversations. This is being done to ensure users do not lose important long-running conversations.

Comments

For the latest tech news and reviews, follow Gadgets 360 on X, Facebook, WhatsApp, Threads and Google News. For the latest videos on gadgets and tech, subscribe to our YouTube channel. If you want to know everything about top influencers, follow our in-house Who'sThat360 on Instagram and YouTube.

Akash Dutta
Akash Dutta is a Chief Sub Editor at Gadgets 360. He is particularly interested in the social impact of technological developments and loves reading about emerging fields such as AI, metaverse, and fediverse. In his free time, he can be seen supporting his favourite football club - Chelsea, watching movies and anime, and sharing passionate opinions on food. More
OnePlus 15 Display Specifications Leaked; Said to Arrive in New Black Colourway
iPhone 17 Series Production Said to Be Expanded in India as Apple Assembles More Models Locally

Advertisement

Follow Us

Advertisement

© Copyright Red Pixels Ventures Limited 2025. All rights reserved.
Trending Products »
Latest Tech News »