Apple announced Thursday is it planning to scan all iPhones in the United States for child abuse imagery, raising alarm among security experts who said the plan could allow the firm to surveil tens of millions of personal devices for unrelated reasons.
In a blog post, the company confirmed reports saying that new scanning technology is part of a suite of child protection programs that would āevolve and expand.ā It will be rolled out as part of iOS 15, which is scheduled for release sometime in August.
Apple, which has often touted itself as a company that promises to safeguard usersā right to privacy, appeared to try and preempt privacy concerns by saying that the software will enhance those protections by avoiding the need to carry out widespread image scanning on its cloud servers.
āThis innovative new technology allows Apple to provide valuable and actionable information to [the National Center for Missing and Exploited Children] and law enforcement regarding the proliferation of known CSAM,ā said the company, referring to an acronym for child sexual abuse material. āAnd it does so while providing significant privacy benefits over existing techniques since Apple only learns about usersā photos if they have a collection of known CSAM in their iCloud Photos account. Even in these cases, Apple only learns about images that match known CSAM.ā
The Cupertino-based tech giant said the system will utilize breakthrough cryptography technology and artificial intelligence to find abuse material when it is stored in iCloud Photos, said the firm in its blog post. The images will be matched to a known database of illegal images, the firm said, adding that if a certain number of those images are uploaded to iCloud Photos, the company will review them.
Those imagesāif theyāre deemed illegalāwill be reported to theĀ National Center for Missing and Exploited Children. The software wonāt be applied to videos, Apple added.
āAppleās expanded protection for children is a game-changer,ā John Clark, the president and CEO of the National Center for Missing and Exploited Children, said in a statement on Thursday about the initiative. āThe reality is that privacy and child protection can coexist.ā
But some security experts and researchers, who stressed they support efforts to combat child abuse, said the program could present significant privacy concerns.
Ross Anderson, professor of security engineering at the University of Cambridge, described Appleās proposed system as āan absolutely appalling idea,āĀ accordingĀ to the Financial Times. āIt is going to lead to distributed bulk surveillance of ā¦āour phones and laptops,ā he remarked.
When news of the proposal broke on Wednesday evening, John Hopkins University professor and cryptographer Matthew Green echoed those concerns.
āThis sort of tool can be a boon for finding child pornography in peopleās phones,ā GreenĀ wrote on Twitter. āBut imagine what it could do in the hands of an authoritarian government?ā
Green saidĀ that āif you believe Apple wonāt allow these tools to be misused [crossed fingers emoji] thereās still a lot to be concerned about,ā noting that such āsystems rely on a database of āproblematic media hashesā that you, as a consumer, canāt review.ā
The expert toldĀ The Associated Press that heās concerned Apple could be pressured by other, more authoritarian governments to scan for other types of information.
Microsoft created photoDNA to assist companies in identifying child sexual abuse images on the internet, while Facebook and Google have implemented systems to flag and review possibly illegal content.
The Epoch Times has contacted Apple for comment.