View Single Post
Old 07-27-2019, 01:08 PM  
GFED
Confirmed User
 
GFED's Avatar
 
Industry Role:
Join Date: May 2002
Posts: 8,098
Quote:
Originally Posted by wehateporn View Post
Apple’s Siri AI assistant sends audio of sexual encounters, embarrassing medical information, drug deals, and other private moments recorded without users’ knowledge to human ‘graders’ for evaluation, a whistleblower has revealed.

Recordings from Apple’s Siri voice assistant are fed to human contractors around the world, who grade the AI based on the quality of its response and whether its activation was deliberate, according to an anonymous contractor who spoke to the Guardian. They claimed accidental activations are much more frequent than Apple lets on, especially with Apple Watch users – and wants the company to own up to the problem.

“There have been countless instances of recordings featuring private discussions between doctors and patients, business deals, seemingly criminal dealings, sexual encounters and so on. These recordings are accompanied by user data showing location, contact details, and app data,” the whistleblower revealed.

Continued https://www.rt.com/news/465181-apple...n-contractors/
All of the wake-on voice command devices, phones, etc do this.
GFED is offline   Share thread on Digg Share thread on Twitter Share thread on Reddit Share thread on Facebook Reply With Quote