Quick question, I want to use a siri shortcut to launch the Home Assistant assist feature. This doesn’t work with the custom Ollama as it tries to use the default Assist even though I removed it. Any ideas?
Other than that, it works great within the webui via text input
Thank you! I was already playing around with Ollama in my home lab and found your add-on from the Ollama repo. It was very easy to get running and works great.
The only downside is that conversation agents can’t (yet?) control anything.
Control of devices is possible with the right model and prompts, I will be adding functionality for this soon. It is a bit sketchy at the moment but new features coming to Ollama will make it easier…
Thanks for your work @ej52! I was able to install ollama using docker, and integrate it with home assistant. However, when I try to talk to it using home assistant it does not work:
Any ideas for what I might be doing wrong? In case it is relevant, I’m running home assistant and ollama on a AMD Ryzen 9 6900HX(Up to 4.9GHz), Radeon 680M Graphics,8C/16T Micro Computer, 32GB DDR5. This is the docker-compose that I use to run ollama:
I suspect your hardware is taking too long to run the model in the predefined timeout of 60 seconds. I will push a update soon that will allow you to set your own timeout for server response.
Thanks. I have a couple of questions that you might be able to answer.
Does the model have a memory? For example, would it be possible to tell it something “in the future if someone says X do Y instead of what you just did” could that have a permanent effect?
I see that in the configuration you can provide a “system prompt”. I imagine that if the prompt is complicated everything will take longer. Is that right? If so, is it possible to train a new model based on an existing one an a complicated prompt to avoid that overhead?
Similar to 2, is it possible to train a new model based on an existing one an additional information? For example, could I provide additional information to the new model in the form of PDF files?
Thanks for the great work. This is very interesting and fun to play with.
this should be the default new AI assistant because 1) ollama is the only local AI which is user friendly and 2) just works without any pain and 3) no need to be linux expert to set it up 4) usual models can be loaded into it. https://github.com/jmorganca/ollama/blob/2a2fa3c3298194f4f3790aade78df2f53d170d8e/docs/linux.md HARDWARE: You need a machine with a GPU that has at least 8GB VRAM but if you wanna buy something then buy such card which has 16GB VRAM, for example NVIDIA T4 GPU. (note this GPU need a small cooling fan as it is server GPU)
the 13B model is maxing out the T4 but only for a few seconds until it generates the answer:
I have seen localai models which can control home assistant. But i believe this relies on openai functions support and thats something localai wraps for llama llms.
I think this is the missing piece for ollama. Localai does not run as well as ollama so a combination of both of these seems to be the best solution
Curious if you have had the chance to report your findings on this yet? I see localai has some documentation on this but it runs poorly compared to ollama
I have Ollama on lxc, and chat with it locally. However, any request from HA with this integration leads to 100% CPU on lxc, which never goes down - and timeout on HA side.
Also, as i understand, there are functions that we can use with Ollama, if model supports them. Right?
It’s not that easy like “i will give it this data and train it on that data”. Model training is nothing to do with the context you’re using in conversation.