Future-proofing HA with local LLMs: Best compact, low-power hardware?

I don’t think you need HAOS to support an npu for this case. Honestly if you’re doing local inference it’s NOT running as an addon in haos (which is where it would HAVE to live) instead you’re putting haos As a guest os on the same iron running inference - not the other way around.