Hi,
i’m looking at a method to store long-term sensor data.
I have been looking at different options, but for now haven’t find the right solution.
What I would like to have is for short-term (e.g. several days) have all ‘high-res’ sensor data available, for some of my sensors this is 1 datapoint per second so this quickly becomes much data.
For the long term I would like to have a min/max/average per sensor per day, so I can see seasonal changes.
Thoughts that came to mind are:
filter-sensor: data is ‘by-default’ stored into the same database the high-res data is in, and don’t know if it’s possible to purge the default database only for the high-res data and leave the min/max/avg in place.
other possibility: some sort of automation that processes the high-res data into 1 datapoint per day and store that into another database
run some python script and export sensor data to e.g. google-sheets or some other online platform and do further processing there.
also looked at the jupyter addon by with data-detective by @robmarkcole@frenck , this really looks awesome, but haven’t figured out if I can use this to create my long-term sensor data.
But as I said I haven’t really figured out how to set this up correctly, any thoughts on how to do this ?
What I’m looking at is basically what Domoticz uses in their sensor graphs.
Regards,
Herman
ah, yes, I forgot to mention that I actually did give them a try
The problem is that I’m not a software guy and only have very limited knowledge of python, e.g. for the last option I have been trying to copy/paste some code together to make e.g. plots of the current data in the hass-database, that already cost me several hours.
I don’t have a clue where to start when I actually want to start doing calculations on the data like min/max/average and store that again into some sort of database.
More or less the same goes for the other option, the ideas are there, but I lack the proficiency to actually implement it
I use influxdb for long term metric storage. It can ‘downsample’ data after a certain point but I haven’t configured that yet (only been using influxdb for a couple weeks now). It’s a lot faster than storing all the metrics in SQL and by combining it with Grafana it allows you to easily make nice dashboards. It sounds like that is exactly what you want and it’s rather easy to setup. I made some dashboards in Grafana and include them as iframe in home-assistant using Grafana’s kiosk mode, so it integrates cleanly
This is the page that describes Influxdb downsampling, which I think is what you are after. I have to admit, that I have just kept the default retention policy and haven’t looked in to it much.
Thanks @mvdm and @gpbenton, I wasn’t aware of the influxdb addon, that adds yet another option to my list. Just spend a few minutes configuring it and I must say this seems to be the most easy setup till now.
Without any programming I could just make nice plots of my sensor data and will check also the downsampling option for the long-term data.