Local LLM returning GGGGGGGGGGGGG

Hi, I have spent days with AI going round and round in circles with no cigar. The inbuilt HA assistant controls things fine, but when I try to use Local LLM runing on my service i get GGGGGGGGGGGGGGG as a response. I can use the LLM when I SSH in to the machine its on (GMKtec K8+) I am using rocm for gpu acceleration, 64gb ram and 16gb dedicated to vram. Everything functions fine in there respective containers the problem appears to be HA sending requests to the K8.
Can anyone help me with this?

I think it's trying to say

I am Groot

Seriously I have no idea, just trying to make you feel better.
Glad my Ollama is still working.

I have seen the same thing, i am using a Nvidia Jetson Orin Nano, from terminal its working fine. I do find that sometimes if I i delete the ollama integration and start over it works. However the model i was using was " thinking out loud" qwen3:4b and so that took it 3 minutes to explain to me how to turn off a light. I will play and try more models and see if i can find one that works well.I will report back if i have progress.

so , i have tried other models, and they all seem to either immediatly fail or do the GGGGGGG thing. So far only qwen3:4b working, but dang its slow for HA