• Home
  • Ai
  • Ai News
  • Anthropic Thwarts Hacker Attempts to Misuse Claude AI for Cybercrime

Anthropic Thwarts Hacker Attempts to Misuse Claude AI for Cybercrime

Anthropic's report said its internal systems had stopped the attacks and it was sharing the case studies to help others understand the risks.

Anthropic Thwarts Hacker Attempts to Misuse Claude AI for Cybercrime

Photo Credit: Anthropic

Anthropic said it follows strict safety practices, including regular testing and outside reviews

Click Here to Add Gadgets360 As A Trusted Source As A Preferred Source On Google
Highlights
  • Anthropic said it had banned the accounts involved, tightened its filters
  • AI tools are being increasingly exploited in cybercrime
  • Attackers attempted to use Claude to produce harmful content
Advertisement

Anthropic said on Wednesday it had detected and blocked hackers attempting to misuse its Claude AI system to write phishing emails, create malicious code and circumvent safety filters.

The company's findings, published in a report, highlight growing concerns that AI tools are increasingly exploited in cybercrime, intensifying calls for tech firms and regulators to strengthen safeguards as the technology spreads.

Anthropic's report said its internal systems had stopped the attacks and it was sharing the case studies - showing how attackers had attempted to use Claude to produce harmful content - to help others understand the risks.

The report cited attempts to use Claude to draft tailored phishing emails, write or fix snippets of malicious code and sidestep safeguards through repeated prompting.

It also described efforts to script influence campaigns by generating persuasive posts at scale and helping low-skill hackers with step-by-step instructions.

The company, backed by Amazon.com and Alphabet, did not publish technical indicators such as IPs or prompts, but said it had banned the accounts involved and tightened its filters after detecting the activity.

Experts say criminals are increasingly turning to AI to make scams more convincing and to speed up hacking attempts. These tools can help write realistic phishing messages, automate parts of malware development and even potentially assist in planning attacks.

Security researchers warn that as AI models become more powerful, the risk of misuse will grow unless companies and governments act quickly.

Anthropic said it follows strict safety practices, including regular testing and outside reviews, and plans to keep publishing reports when it finds major threats.

Microsoft and SoftBank-backed OpenAI and Google have faced similar scrutiny over fears their AI models could be exploited for hacking or scams, prompting calls for stronger safeguards.

Governments are also moving to regulate the technology, with the European Union moving forward with its Artificial Intelligence Act and the United States pushing for voluntary safety commitments from major developers.

© Thomson Reuters 2025

Comments

For the latest tech news and reviews, follow Gadgets 360 on X, Facebook, WhatsApp, Threads and Google News. For the latest videos on gadgets and tech, subscribe to our YouTube channel. If you want to know everything about top influencers, follow our in-house Who'sThat360 on Instagram and YouTube.

Further reading: Anthropic, Claude, AI, Cyberattack, Cybercrime
Resurrecting the Dead Using AI: What Are Deadbots and Why Are They Stirring Ethical Debate?

Advertisement

Follow Us

Advertisement

© Copyright Red Pixels Ventures Limited 2025. All rights reserved.
Trending Products »
Latest Tech News »