Nvidia has just recently come out with some affordable supercomputers. I would LOOOVE to be able to run HA on one of those, to then be able to include numerous cameras and Frigate, as well as locally housed large language models for AI processing…
I have a gaming machine with a Nvidia card on which I run a fully local voice assistant. The details are way over my head, but my understanding is that it was especially written to exploit the speed of the GPU - which HA and the rest aren’t.
It’s an Nvidia Grace (ARM) CPU married directly to a Blackwell GPU that drives in excess of 5000 TOPS. (100x what qualifies a machine to be a Copilot +PC)
Its more like a gigantic Jetson Nano. And no i don’t expect it to run HA directly. Although I DO expect it to run ollama and 30B models locally (Jensen says it runs Nvidias ai stack…) basically you could in theory run a foundational model at home and if it’s capable of a test time compute or test time training (reasoning) capable model… You almost have something that could theoretically run an o1 class model or comparable on your desk.
I know 3K is a lot of scratch but for what I think this box will do - it will be a steal.
I been looking at the latest CES reveals and agree. The computing power in such a small form factor is amazing. Considering I have several GPUs I paid 2K for, a price of 3K seems reasonable cheap.
HA has no use for such power. Any integration would need seperate service. It is the seperate service that would use processing power and HA integration simply connect to the service.
A 5090 will push appropriately 3300 TOPS at max throuput will cost nearly as much as a digits but you get to install it in your machine of choice and game on it. It’s not a direct comparison.
The digits is pure ai.all the time and there’s still a lot of questions. A 5090. Well if you like PUBG… It’s gonna be a banger.
HA addons do not support the nvidia-runtime. You would need to (and likely very much want to) run a standard linux installation and run everything in docker
Honestly I’d run the box purely as suggested by Nvidia and call to it from off box. It’s a different animal, purpose built for what it does. I don’t think it’s going to run non ai workload worth a crud. Ok yeah it can brute force compute but if you want this box to do what it does you’ll want it to be a purpose built machine.