icon bookmark-bicon bookmarkicon cameraicon checkicon chevron downicon chevron lefticon chevron righticon chevron upicon closeicon v-compressicon downloadicon editicon v-expandicon fbicon fileicon filtericon flag ruicon full chevron downicon full chevron lefticon full chevron righticon full chevron upicon gpicon insicon mailicon moveicon-musicicon mutedicon nomutedicon okicon v-pauseicon v-playicon searchicon shareicon sign inicon sign upicon stepbackicon stepforicon swipe downicon tagicon tagsicon tgicon trashicon twicon vkicon yticon wticon fm
27 Jul, 2019 03:53

Siri ‘regularly’ records sex encounters, sends ‘countless’ private moments to Apple contractors

Siri ‘regularly’ records sex encounters, sends ‘countless’ private moments to Apple contractors

Apple’s Siri AI assistant sends audio of sexual encounters, embarrassing medical information, drug deals, and other private moments recorded without users’ knowledge to human ‘graders’ for evaluation, a whistleblower has revealed.

Recordings from Apple’s Siri voice assistant are fed to human contractors around the world, who grade the AI based on the quality of its response and whether its activation was deliberate, according to an anonymous contractor who spoke to the Guardian. They claimed accidental activations are much more frequent than Apple lets on, especially with Apple Watch users – and wants the company to own up to the problem.

There have been countless instances of recordings featuring private discussions between doctors and patients, business deals, seemingly criminal dealings, sexual encounters and so on. These recordings are accompanied by user data showing location, contact details, and app data,” the whistleblower revealed.

Also on rt.com Apple co-founder Steve Wozniak wants everyone to quit Facebook

Apple is subcontracting out, there’s a high turnover. It’s not like people are being encouraged to have consideration for people’s privacy, or even consider it,” the whistleblower said, explaining that they are concerned these recordings, produced when Siri thinks it hears its “wake word,” could be used against the people who (accidentally) made them – especially given the “broad” amount of user data they claim contractors are “free to look through.” In what sounds like a sick joke on the part of some programmer, the sound of a zipper unzipping often triggers Siri to wake up.

If there were someone with nefarious intentions, it wouldn’t be hard to identify [people on the recordings].

While Apple does not explicitly mention any human involvement in Siri’s training in the AI’s documentation, it acknowledged when asked about its practices that “a small portion of Siri requests are analyzed to improve Siri and dictation.” The company insisted that this amounted to less than one percent of all daily activations of the AI and that the recordings were “typically only a few seconds long.”

Also on rt.com ANYONE can be re-identified from ‘anonymous data’, researchers claim & let you TEST it

While Apple emphasized that a user’s Apple ID and name are not attached to clips reviewed by contractors, it also took pains to explain that recordings are “analyzed in secure facilities and all reviewers are under the obligation to adhere to Apple’s strict confidentiality requirements” – suggesting the company is aware of how easily even a recording stripped of its user ID can be connected to the user who made it.

Also on rt.com Amazon admits it keeps some Alexa recordings even when users delete them

Siri isn’t the only voice assistant that transmits users’ private moments back to the mothership, of course – Amazon’s Alexa infamously has entire chat rooms for its human trainers to discuss difficult-to-understand audio clips (or mock funny recordings) and Google Home uses a similar system of outsourced “language experts” that allows the company to claim that no one at Google has access to the recordings its devices make.

Also on rt.com Outsourced spying: Google admits ‘language experts’ listen to ‘some’ assistant recordings

Like this story? Share it with a friend!