Publicly, that is. They have no doubt been doing it in secret since they launched it.
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
Off-device processing has been the default from day one. The only thing changing is the removal for local processing on certain devices, likely because the new backing AI model will no longer be able to run on that hardware.
With on-device processing, they don’t need to send audio. They can just send the text, which is infinitely smaller and easier to encrypt as “telemetry”. They’ve probably got logs of conversations in every Alexa household.
If you look at the article, it was only ever possible to do local processing with certain devices and only in English. I assume that those are the ones with enough compute capacity to do local processing, which probably made them cost more, and that the hardware probably isn't capable of running whatever models Amazon's running remotely.
I think that there's a broader problem than Amazon and voice recognition for people who want self-hosted stuff. That is, throwing loads of parallel hardware at something isn't cheap. It's worse if you stick it on every device. Companies
even aside from not wanting someone to pirate their model running on the device
are going to have a hard time selling devices with big, costly, power-hungry parallel compute processors.
What they can take advantage of is that for a lot of tasks, the compute demand is only intermittent. So if you buy a parallel compute card, the cost can be spread over many users.
I have a fancy GPU that I got to run LLM stuff that ran about $1000. Say I'm doing AI image generation with it 3% of the time. It'd be possible to do that compute on a shared system off in the Internet, and my actual hardware costs would be about $33. That's a heckofa big improvement.
And the situation that they're dealing with is even larger, since there might be multiple devices in a household that want to do parallel-compute-requiring tasks. So now you're talking about maybe $1k in hardware for each of them, not to mention the supporting hardware like a beefy power supply.
This isn't specific to Amazon. Like, this is true of all devices that want to take advantage of heavyweight parallel compute.
I think that one thing that it might be worth considering for the self-hosted world is the creation of a hardened network parallel compute node that exposes its services over the network. So, in a scenario like that, you would have one (well, or more, but could just have one) device that provides generic parallel compute services. Then your smaller, weaker, lower-power devices
phones, Alexa-type speakers, whatever
make use of it over your network, using a generic API. There are some issues that come with this. It needs to be hardened, can't leak information from one device to another. Some tasks require storing a lot of state
like, AI image generation requires uploading a large model, and you want to cache that. If you have, say, two parallel compute cards/servers, you want to use them intelligently, keep the model loaded on one of them insofar as is reasonable, to avoid needing to reload it. Some devices are very latency-sensitive
like voice recognition
and some, like image generation, are amenable to batch use, so some kind of priority system is probably warranted. So there are some technical problems to solve.
But otherwise, the only real option for heavy parallel compute is going to be sending your data out to the cloud. And even if you don't care about the privacy implications or the possibility of a company going under, as I saw some home automation person once point out, you don't want your light switches to stop working just because your Internet connection is out.
Having per-household self-hosted parallel compute on one node is still probably more-costly than sharing parallel compute among users. But it's cheaper than putting parallel compute on every device.
Linux has some highly-isolated computing environments like seccomp that might be appropriate for implementing the compute portion of such a server, though I don't know whether it's too-restrictive to permit running parallel compute tasks.
In such a scenario, you'd have a "household parallel compute server", in much the way that one might have a "household music player" hooked up to a house-wide speaker system running something like mpd or a "household media server" providing storage of media, or suchlike.
I have always told people to avoid Amazon.
They have doorbells to watch who comes to your house and when.
Indoor and outdoor security cameras to monitor when you go outside, for how long, and why.
They acquired roomba, which not only maps out your house, but they have little cameras in them as well, another angle to monitor you through your house in more personal areas that indoor cameras might not see.
They have the Alexa products meant to record you at all times for their own use and intent.
Why do you think along with Amazon Prime subscriptions you get free cloud storage, free video streaming, free music? They are categorizing you in the most efficient and accurate way possible.
Boycott anything Amazon touches
I agree with your sentiment and despise Amazon but they do not own roomba the deal fell through.
Christ, finally a win
Amazon really got people to pay to be spied on. Wild world we live in bois
People are saying don't get an echo but this is the tip of an iceberg. My coworkers' cell phones are eavesdropping. My neighbors doorbells record every time I leave the house. Almost every new vehicle mines us for data. We can avoid some of the problem but we cannot avoid it all. We need a bigger, more aggressive solution if we are going to have a solution at all.
How about regulation? Let's start with saying data about me belongs to me, not to whoever collected the data, as is currently the case
Today: "...they will be deleted after Alexa processes your requests."
Some point in the not-so-distant future: "We are reaching out to let you know that your voice recordings will no longer be deleted. As we continue to expand Alexa's capabilities, we have decided to no longer support this feature."
“We lied and paid a $3M fine.”
Want to setup a more privacy friendly solution?
Have a look at Home Assistant! It’s a great open source smart home platform that recently released a local (so not processing requests in the cloud) voice assistant. It’s pretty neat!
If Corporations were people, they'd be disappeared in the night for stuff like this.
Which is why they're not people.
Why anyone would want some Tech company spybot sifting through their private experiences is beyond me, but that's definitely what they are doing.
They literally could just leave the feature on the device, but then you can't force your users to send you all their data, voices, thoughts and first borns
Fuck Amazon, fuck Bezos
If you do not want to set your voice recordings setting to 'Don't save recordings,' please follow these steps before March 28th:
Am I the only one curious to know what these steps are? The image cuts off the rest of the email.
- Unplug your amazon echo devices
- Hit it with a hammer
- Send it to an electronics recycler
There's no way they weren't doing this already.
I mean if they were doing this already there would be no point in sending this email out. They would have just happily continued letting people think it wasn’t happening while doing it anyway, while not having to deal with the backlash this will generate.
Easy fix: don't buy this garbage to begin with. It's terrible for the environment, terrible for your privacy, of dubious value to begin with.
If every man is an onion, one of my deeper layers is crumudgeon. So take that into account when I say fuck all portable speakers. I'm so tired of hearing everyone's shitty noise. Just fucking everywhere. It takes one person feeling entitled to blast the shittiest music available to ruin everyone in a 500yd radius's day. If this is you, I hope you stub your toe on every coffee table, hit your head on every door jam, miss every bus.
How the fuck does anyone even buy one of these
The same people who buy mobile phones; despite those being bugs/spy-devices.
True, but a mobile phone is basically a world brain, calculator, camera, flashlight, you can watch movies on it in hi def, hate it all you want, it's one of the most versatile tools on the planet. An echo dot, it just spy garbage and nothing else
be aware, everything you say around amazon, apple, alphabet, meta, and any other corporate trash products are being sold, trained on, and sent to your local alphabet agency. it's been this way for a while, but this is a nice reminder to know when to speak and when to listen
I can’t believe people are still voluntarily wire tapping themselves in 2025
everything you say to your echo/alexa has always been sent to amazon.
theres literally been leaks proving it.
It's always been this way for the cheap speakers. They've no processing power on-board and need the cloud just to tell you the time.
So... if you own an inexpensive Alexa device, it just doesn't have the horsepower to process your requests on-device. Your basic $35 device is just a microphone and a wifi streamer (ok, it also handles buttons and fun LED light effects). The Alexa device SDK can run on a $5 ESP-32. That's how little it needs to work on-site.
Everything you say is getting sent to the cloud where it is NLP processed, parsed, then turned into command intents and matched against the devices and services you've installed. It does a match against the phrase 'slots' and returns results which are then turned into voice and played back on the speaker.
With the new LLM-based Alexa+ services, it's all on the cloud. Very little of the processing can happen on-device. If you want to use the service, don't be surprised the voice commands end up on the cloud. In most cases, it already was.
If you don't like it, look into Home Assistant. But last I checked, to keep everything local and not too laggy, you'll need a super beefy (expensive) local home server. Otherwise, it's shipping your audio bits out to the cloud as well. There's no free lunch.
Now they can hear me scream “shut the fuck up Alexa!!!!” every time she says “…by the way…” when I just want to know what time it is.
If anyone remembers the Mycroft Mark II Voice Assistant Kickstarter and was disappointed when development challenges and patent trolls caused the company's untimely demise, know that hope is not lost for a FOSS/OSHW voice assistant insulated from Big Tech..
FAQ: OVOS, Neon, and the Future of the Mycroft Voice Assistant
Disclaimer: I do not represent any of these organizations in any way; I just believe in their mission and wish them all the success in getting there by spreading the word.
In the age of techno-fascism, the people willingly pay to install the listening devices into their own homes.
Just sold my 3 devices and shut down Amazon account. It's very liberating and I don't miss it one bit. Have Home Assistant and a couple of really good 2nd hand Sonos speakers.
My family has one in most rooms of our house...ugh
For anyone with existing Home Assistant setup, the Home Assistant Voice Preview is pretty good alternative, when it comes to voice control of HA. The setup is very easy. If you want conversational functionality, you could even hook it up to an LLM, cloud or local. It can also be used for media playback and it's got an aux out port.
I used to use Google Home Mini for voice control of Home Assistant. The Voice Preview replaced that rather nicely.