this post was submitted on 10 Aug 2025
20 points (76.3% liked)

DIY Electronics and Hardware

144 readers
1 users here now

founded 3 months ago
MODERATORS
top 9 comments
sorted by: hot top controversial new old
[–] rbn@sopuli.xyz 11 points 1 week ago* (last edited 1 week ago)

For converting your spoken words into text, it taps into OpenAI’s Whisper model, an automatic speech recognition system renowned for its accuracy and ability to handle various accents and background noise.

Have the hardware requirements of Whisper dropped significantly over the last few months? I played around with it in context of home assustant year of the voice. Despite using a (4 year old) ThinkPad with 32 GB of RAM and a 4 core (8 threads) i7 the accuracy and performance of Whisper was still not at a point that I'd use for productive use.

A rather simple sentence like 'turn the light in the living room on' worked maybe in 70% of the cases if I sat right next to the microphone and without any background noise. With music playing in the background or other people talking in parallel it dropped to ~25% accuracy.

If it now runs just fine on a Raspberry Pi Zero that would be a massive improvement!

[–] fubarx@lemmy.world 4 points 1 week ago

Nice! On-device AI makes so much sense. I bet the next version will have a camera.

[–] wizzor@sopuli.xyz 3 points 1 week ago (2 children)

My bad... But then... What's the point of having a dedicated piece of LLM hardware, isn't that like having a hardware client for email?

[–] Coelacanthus@infosec.pub 1 points 2 days ago

Raspberry Pi Zero 2 can run Linux. So you can run mail client on it of course.

[–] SaneMartigan@aussie.zone 1 points 1 week ago* (last edited 1 week ago)

Someone (not apple) made an iPhone in the 90s with a big lcd screen, it worked as an email reader (pop3/smtp)

[–] wizzor@sopuli.xyz 3 points 1 week ago (1 children)

The models an rpi zero can run are very limited though. 512 mb ram is very, very little for AI models.

[–] Redex68@lemmy.world 9 points 1 week ago (1 children)

The LLM isn't local

For the actual conversational responses, the project typically utilizes cloud-based large language models accessed via APIs

[–] rbn@sopuli.xyz 16 points 1 week ago (1 children)

Then, from my perspective, there's little to no value to have a dedicated piece of hardware for it. At least I'd guess that 99.999% the target audience for such a thing already has a smartphone with them. What - if not for the sake of privacy - is the added value of a special chatbot device?

[–] Redex68@lemmy.world 4 points 1 week ago

I guess it makes it a bit easier to access + it's a fun project to DIY, not much else.