After a week of criticism over Apple Inc.'s planned new system for detecting images of child abuse, Apple said Friday that it would hunt only for pictures flagged by clearinghouses in multiple countries.
That shift and others intended to reassure privacy advocates were detailed to reporters in an unprecedented fourth background briefing since the initial announcement eight days prior of a plan to monitor customer devices.
On Friday, after earlier declining to say how many matching images on a phone or computer would it take before the operating system notifies Apple for a human review and possible reporting to authorities, executives said that it would start with 30, although the number could get lower over time as the system improves.
Apple also said it would be easy for researchers to make sure that the list of image identifiers being sought on one iPhone was the same as the lists on other phones, seeking to blunt concerns that the new mechanism could be used to target individuals. The company published a long paper explaining how it had reasoned to defend potential attacks on the system and reasoned against them.
Apple acknowledged that it had handled communications around the program poorly, triggering backlash from influential technology policy groups and even its own employees concerned that the company was jeopardizing its reputation for protecting consumer privacy.
It declined to say whether that criticism had changed any of the policies or software, but said that the project was still in development and changes were to be expected.
Asked why it had only announced that the U.S. -based National Center for Missing and Exploited Children would be a supplier of flagged image IDs when at least one other clearinghouse would need to have separately flagged the same picture, an Apple executive said that the company had only finalized its deal with NCMEC.
The rolling series of explanations, each giving more details that make the plan seem less hostile to privacy, convinced some of the company's critics that their voices were forcing real change.
Our pushing has an effect, tweeted Riana Pfefferkorn, an encryption and surveillance researcher at Stanford University.
Apple said last week that it will check photos stored on the iCloud back up system if they are about to be uploaded by the United States, adding later that it would begin with just the iPhone.
Similar technology companies perform comparable checks once photos are uploaded to their servers. Why Apple's decision to put key parts of the system on the phone prompted concerns that governments could force Apple to expand the system for other uses, such as scanning for prohibited political imagery.
The controversy has even moved into Apple's ranks, with employees debating the move in hundreds of posts in an internal chat channel, Reuters reported this week.