How to connect Llama.cpp server to Home Assistant?

Hi, I was able to get Ollama to connect to home assistant to serve a local LLM, but I’ve moved to using more llama.cpp for fine grain control. However, when I go to try to connect things together again, I haven’t found a way to connect a home assistant to my local llama server.

Has anyone been able to get to work? I see a thread that seems to use a specific llama.cpp python docker container. Was wondering if there is a way by just linking into the the llama-server? I’m able to get open webui to connect that way via a local lan ip address and port, but have tried using extended openai and custom conversations without any success…

Any help is appreciated!

I was able to connect using the Custom Conversation integration, selecting OpenAI an using the llama.cpp server address: http://llamaserverip:8080/v1 and anything for Api key