SAN FRANCISCO: A backlash over Apple’s move to scan US customer phones and computers for child sex abuse images has grown to include employees speaking out internally, a notable turn in a company famed for its secretive culture, as well as provoking intensified protests from leading technology policy groups.
Apple employees have flooded an Apple internal Slack channel with more than 800 messages on the plan announced a week ago, workers who asked not to be identified told Reuters. Many expressed worries that the feature could be exploited by repressive governments looking to find other material for censorship or arrests, according to workers who saw the days-long thread.
Past security changes at Apple have also prompted concern among employees, but the volume and duration of the new debate is surprising, the workers said. Some posters worried that Apple is damaging its leading reputation for protecting privacy.
Though coming mainly from employees outside of lead security and privacy roles, the pushback marks a shift for a company where a strict code of secrecy around new products colors other aspects of the corporate culture.
Slack rolled out a few years ago and has been more widely adopted by teams at Apple during the pandemic, two employees said. As workers used the app to maintain social ties during the work-from-home era by sharing recipes and other light-hearted content, more serious discussions have also taken root.
In the Slack thread devoted to the photo-scanning feature, some employees have pushed back against criticism, while others said Slack wasn’t the proper forum for such discussions.
Core security employees did not appear to be major complainants in the posts, and some of them said that they thought Apple’s solution was a reasonable response to pressure to crack down on illegal material.
Other employees said they hoped that the scanning is a step toward fully encrypting iCloud for customers who want it, which would reverse Apple’s direction on the issue a second time.
Last week’s announcement is drawing heavier criticism from past outside supporters who say Apple is rejecting a history of well-marketed privacy fights.
They say that while the US government can’t legally scan wide swaths of household equipment for contraband or make others do so, Apple is doing it voluntarily, with potentially dire consequences.
People familiar with the matter said a coalition of policy groups are finalizing a letter of protest to send to Apple within days demanding a suspension of the plan. Two groups, the Electronic Frontier Foundation (EFF)and Center for Democracy and Technology (CDT) both released newly detailed objections to Apple’s plan in the past 24 hours.
“What Apple is showing with their announcement last week is that there are technical weaknesses that they are willing to build in,” CDT project director Emma Llanso said in an interview. “It seems so out of step from everything that they had previously been saying and doing.”
Apple declined to comment for this story. It has said it will refuse requests from governments to use the system to check phones for anything other than illegal child sexual abuse material.
Outsiders and employees pointed to Apple’s stand against the FBI in 2016, when it successfully fought a court order to develop a new tool to crack into a terrorism suspect’s iPhone. Back then, the company said that such a tool would inevitably be used to break into other devices for other reasons.
But Apple was surprised its stance then was not more popular, and the global tide since then has been toward more monitoring of private communication.
With less publicity, Apple has made other technical decisions that help authorities, including dropping a plan to encrypt widely used iCloud backups and agreeing to store Chinese user data in that country.
A fundamental problem with Apple’s new plan on scanning child abuse images, critics said, is that the company is making cautious policy decisions that it can be forced to change, now that the capability is there, in exactly the same way it warned would happen if it broke into the terrorism suspect’s phone.
Apple says it will scan only in the United States and other countries to be added one by one, only when images are set to be uploaded to iCloud, and only for images that have been identified by the National Center for Exploited and Missing Children and a small number of other groups.
But any country’s legislature or courts could demand that any one of those elements be expanded, and some of those nations, such as China, represent enormous and hard to refuse markets, critics said.
Police and other agencies will cite recent laws requiring “technical assistance” in investigating crimes, including in the United Kingdom and Australia, to press Apple to expand this new capablity, the EFF said.
“The infrastructure needed to roll out Apple’s proposed changes makes it harder to say that additional surveillance is not technically feasible,” wrote EFF General Counsel Kurt Opsahl.
Lawmakers will build on it as well, said Neil Brown, a UK tech lawyer at decoded.legal: “If Apple demonstrates that, even in just one market, it can carry out on-device content filtering, I would expect regulators/lawmakers to consider it appropriate to demand its use in their own markets, and potentially for an expanded scope of things.”