Amazon accidentally sent 1,700 Alexa recordings to the wrong person, including audio from in the shower
- Amazon accidentally sent 1,700 recordings of someone speaking to Alexa to the wrong person, according to a German magazine.
- The recordings had lots of personal information, and the magazine was easily able to find the person whose personal data was leaked.
- The episode underscores that Amazon stores audio files when you speak to Alexa.
- "This was an unfortunate case of human error and an isolated incident," Amazon said in a statement. "We have resolved the issue with the two customers involved and have taken steps to further improve our processes. We were also in touch on a precautionary basis with the relevant regulatory authorities."
Imagine if you had Amazon Alexa-enabled speakers all over your house. Perhaps you had one Echo in your living room and an Amazon Fire Stick connected to your TV. Maybe you talk to Alexa to set alarms, control your smart home, and play music in the shower.
Then one day, over 1,700 recordings of you speaking to Alexa are sent to a totally random person — and you don't even know about it until a magazine gets in touch.
Apparently, this actually happened, in a scary story coming out of Germany on Thursday. C't magazine in Germany is reporting that someone requested their personal data from Amazon and were shocked to discover 1,700 audio files of someone he didn't know talking to Alexa.
These files even had audio recordings of the person in the shower, according to the report.
He provided those recordings to the German magazine, which was able to get in touch with the person who owned the house full of Amazon Alexa devices. The audio files, in German, reveal a lot of personal data about the person, including where he lived, first and last names, who his partner is, and his taste in music.
It turns out, Amazon had not contacted him about the data breach, but after his story was about to become public, Amazon ended up giving him new Echo devices and a Prime membership by way of compensation, according to the report.
The story underscores that Amazon does record and store your voice when you speak to Alexa. In fact, you can check what you've said to Alexa at amazon.com/alexaprivacy and delete portions or the entirety of the stored audio files.
Amazon needs to store these recordings to improve its voice recognition systems, but people who frequently speak to their smart speaker should think twice before telling Alexa any secrets.
"This was an unfortunate case of human error and an isolated incident," an Amazon spokesperson told Business Insider. "We have resolved the issue with the two customers involved and have taken steps to further improve our processes. We were also in touch on a precautionary basis with the relevant regulatory authorities."
Amazon also confirmed that it has apologized to the two customers and has been in touch with regulatory authorities, including GDPR administrators.
- Facebook had a secret data deal with Amazon which flouted its own privacy rules
- There's a scary iPhone feature that erases all your data after too many password attempts — here's why you should turn it on anyway
- Facebook hosted a pop-up event in New York City for 1 day only to teach people about privacy. Here's what it was like to visit