Apple Says Will Implement System That Will Check iPhone Handsets for Images of Child Sexual Abuse

Apple says the detection of child abuse image uploads sufficient to guard against false positives will trigger a human review.

Advertisement
By Reuters | Updated: 6 August 2021 09:18 IST
Highlights
  • The system could open the door to monitoring of political speech
  • Apple has implemented that database using a technology called NeuralHash
  • iPhone will create a hash of the image to be uploaded

Apple's new system seeks to address requests from law enforcement to help stem child sexual abuse

Apple on Thursday said it will implement a system that checks photos on iPhone devices in the United States before they are uploaded to its iCloud storage services to ensure the upload does not match known images of child sexual abuse.

Detection of child abuse image uploads sufficient to guard against false positives will trigger a human review of and report of the user to law enforcement, Apple said. It said the system is designed to reduce false positives to one in one trillion.

Apple's new system seeks to address requests from law enforcement to help stem child sexual abuse while also respecting privacy and security practices that are a core tenet of the company's brand. But some privacy advocates said the system could open the door to monitoring of political speech or other content on iPhone handsets.

Advertisement

Most other major technology providers - including Alphabet's Google, Facebook, and Microsoft - are already checking images against a database of known child sexual abuse imagery.

"With so many people using Apple products, these new safety measures have lifesaving potential for children who are being enticed online and whose horrific images are being circulated in child sexual abuse material," John Clark, chief executive of the National Center for Missing & Exploited Children, said in a statement. "The reality is that privacy and child protection can co-exist."

Here is how Apple's system works. Law enforcement officials maintain a database of known child sexual abuse images and translate those images into "hashes" - numerical codes that positively identify the image but cannot be used to reconstruct them.

Advertisement

Apple has implemented that database using a technology called "NeuralHash", designed to also catch edited images similar to the originals. That database will be stored on iPhone gadgets.

When a user uploads an image to Apple's iCloud storage service, the iPhone will create a hash of the image to be uploaded and compare it against the database.

Advertisement

Photos stored only on the phone are not checked, Apple said, and human review before reporting an account to law enforcement is meant to ensure any matches are genuine before suspending an account.

Apple said users who feel their account was improperly suspended can appeal to have it reinstated.

Advertisement

The Financial Times earlier reported some aspects of the programme.

One feature that sets Apple's system apart is that it checks photos stored on phones before they are uploaded, rather than checking the photos after they arrive on the company's servers.

On Twitter, some privacy and security experts expressed concerns the system could eventually be expanded to scan phones more generally for prohibited content or political speech.

Apple has "sent a very clear signal. In their (very influential) opinion, it is safe to build systems that scan users' phones for prohibited content," Matthew Green, a security researcher at Johns Hopkins University, warned.

"This will break the dam — governments will demand it from everyone."

Other privacy researchers such as India McKinney and Erica Portnoy of the Electronic Frontier Foundation wrote in a blog post that it may be impossible for outside researchers to double check whether Apple keeps its promises to check only a small set of on-device content.

The move is "a shocking about-face for users who have relied on the company's leadership in privacy and security," the pair wrote.

"At the end of the day, even a thoroughly documented, carefully thought-out, and narrowly-scoped backdoor is still a backdoor," McKinney and Portnoy wrote.


Is OnePlus cannibalising itself with the Nord 2 5G? We discussed this and more on Orbital, the Gadgets 360 podcast. Orbital is available on Apple Podcasts, Google Podcasts, Spotify, Amazon Music and wherever you get your podcasts.
Affiliate links may be automatically generated - see our ethics statement for details.
 

Catch the latest from the Consumer Electronics Show on Gadgets 360, at our CES 2026 hub.

Further reading: Apple, iPhone, iCloud
Advertisement

Related Stories

Popular Mobile Brands
  1. Ustaad Bhagat Singh OTT Release: When, Where to Watch the Telugu Action Drama
  2. Sirai OTT Release: When, Where to Watch the Tamil Courtroom Drama Online
  3. Amazon Great Republic Day Sale 2026: Know the Best Deals on Tablets
  1. Toxic Gas May Have Sparked Life on Earth Through Icy ‘Cobweb’ Crystals
  2. Is Space Sticky? New Study Challenges Standard Dark Energy Theory
  3. Sirai OTT Release: When, Where to Watch the Tamil Courtroom Drama Online
  4. Wheel of Fortune India OTT Release: When, Where to Watch Akshay Kumar-Hosted Global Game Show
  5. NASA Confirms Expedition 74 Will Continue ISS Work After Crew-11 Exit
  6. European Space Agency Hit by Cyberattacks, Hundreds of Gigabytes of Data Stolen by Hackers
  7. Ustaad Bhagat Singh OTT Release: When, Where to Watch Harish Shankar's Telugu Action Drama Film
  8. Bha Bha Ba is Now Streaming: All You Need to Know About This Malayalam Comedy Thriller Film
  9. World’s Biggest Alien Search Enters Final Stage With 100 Mystery Signals
  10. NASA Pulls Out Artemis II Rocket to Launch Pad Ahead of Historic Moon Mission
Gadgets 360 is available in
Download Our Apps
Available in Hindi
© Copyright Red Pixels Ventures Limited 2026. All rights reserved.