Everything is currently available from hardware to opensource software but what we have available running is likely what creates your incorrect assumptions.
Could be, but GPUs are much more suited for this task today. Looking around at the massive number of users here and on Discord who are hopelessly frustrated with their streaming nodes and use of faster whisper, Willow is a solution that works today, is incredibly accurate, employs auto correct, supports HA natively, passes unrecognized commands to Alexa, and is cheap and easy to implement, both from a hardware and software standpoint.
As a huge fan of HA, I sincerely hope the HA naitive solutions mature to the point of being useable like Willow is today, but the vast majority of Piās and NUCs being used today simply wont cut it if folks expect local voice to work as a viable and reliable replacement for Alexa or Google devices. Iām more than happy to be proven incorrect, but my career in electronics and audio processing indicates otherwise.
As Iāve repeatedly stated, folks should look at the current HA voice stack as a fun proof of concept, not something they can deploy for everyday use. Unfortunately, most of the slick demo videos and such donāt make that clear, and frustration ensues. In fact, in the 8 years or so that Iāve been using HA, Iāve never seen so much frustration expressed by users trying to get something billed as a new feature, actually working. Myself included, and Iām an advanced user. Of course I expect all of this to be streamlined and improve over time. History shows us this is the case more often than not, especially with the massive interest in voice as a core function of HA.
So, respectfully, I offer a look and comments on an alternative solution that more and more folks are finding quite useable. And one they can experiment with today that gives excellent results.
Jeff
Edit: In my previous post,I should have stated āthe current crop of widely deployed Rpisā wonāt be good for local STT. The hamster wheel of hardware upgrades will undoubtedly change that in the future.