This would be a solution to the problem that the agent is not aware what the last prompt was but one of the benefits of trying it first locally is that it reduces costs. The LLMs are not (completely) free. For example the Google LLM is free for some amount of requests but if it is exceeded, it will cost you money.