OpenAI just released the API model gpt-4o-mini-search-preview
Unfortunately I’m receiving an error message using OpenAI Conversation integration for my assist pipeline, when I try to use it:
“Error talking to OpenAI”
Was anybody successful with gpt-4o-mini-search-preview?
it just says ‘error’ in the assist logs, but I found out its even happening via the OpenAI playground:
“The requested model ‘gpt-4o-mini-search-preview’ does not exist.”
I’m able to use the model (I’m in the US), but the model token limit is 6000, and the requests coming from HA all seem to be just above that at ~6200.
Typical error: Rate limited by OpenAI: Error code: 429 - {‘error’: {‘message’: ‘Request too large for gpt-4o-mini-search-preview in organization org-[orgID] on tokens per min (TPM): Limit 6000, Requested 6680. The input or output tokens must be reduced in order to run successfully. Visit https://platform.openai.com/account/rate-limits to learn more.’, ‘type’: ‘tokens’, ‘param’: None, ‘code’: ‘rate_limit_exceeded’}}
It seems like the current integration (or the OpenAI Extended Conversation integration) does not support this model, since the model does not support temperature and top_p parameters, which the integration sends. Here’s the response:
And here’s the response with OpenAI Extended Conversation integration:
Error code: 404 - {'error': {'message': 'functions is not supported in this model. For a list of supported models, refer to https://platform.openai.com/docs/guides/function-calling#models-supporting-function-calling.', 'type': 'invalid_request_error', 'param': None, 'code': None}}
Since you’re already using Node-RED, you could replace the Home Assistant automation with a sentence node and handle the dynamic response directly in Node-RED.
That’ll need a small tweak to the openai conversation handler just like what they did for o1/o3. Unfortunately the model needs tweaks to what’s being sent and doesn’t just ‘slot in’ so we’ll have to wait for the conversation agent to catch up.
That said, I doubt they miss it… I know continued convo is scheduled for 2025.4 I’d love to see this hit too. Search is one of the few things I’m missing rn. Tbh.
I’ve been working on building it for half a year. For local you need an ollama installation that supports openai api calling then you need to install a tool to do the search and something to publish. Look what I’m saying I’d there’s an easy way and a hard way.
For cloud 4omini search is about to be a giant easy button.
To get the same on local your llm either needs to support search locally (you installed the tool) or you gave access to a tool through MCP.
Ive done all of them mcp is far and away the easiest and most light weight.
Someone could (ive almost nailed it I’m working on it but my docker-fu sucks) create a container that has brave search, and an MCP proxy. Fire it up and publish to HA. Boom search.
Want dB access. Container with a sql tool.
Want tool a, b, c or even an entirely different llm?
Its absolutely the way we’ll be doing this in a year or so… So I’m building to that.