Updated Docker and suddenly most of my zwave network is inaccessable/non-functional

I am running the homeassistant docker in Unraid. Yesterday there was an update and it processed without issue, but I was alerted by my wife that the lights were acting funny. I logged in to HA and sure enough, 80% of my zwave network is now showing up as “Node:Undefined Undefined”. For these nodes, there are no entities associated with them with the zwave manager. I checked the entitty manager and it stills sees the zwave devices as well as the entities, but they are all marked as “unavailable”.

I’ve tried restarting the app and zwave multiple times but this has not resolved the issue. I’ve attempted a network heal and again, no changes. Tested the network and again, no changes.

When rebooting the log clearly shows all the impacted nodes as timing out

2020-03-23 21:45:06 WARNING (MainThread) [homeassistant.components.zwave] Z-Wave node 67 not ready after 30 seconds, continuing anyway
2020-03-23 21:45:06 WARNING (MainThread) [homeassistant.components.zwave] Z-Wave node 68 not ready after 30 seconds, continuing anyway
2020-03-23 21:45:06 WARNING (MainThread) [homeassistant.components.zwave] Z-Wave node 69 not ready after 30 seconds, continuing anyway
2020-03-23 21:45:06 WARNING (MainThread) [homeassistant.components.zwave] Z-Wave node 70 not ready after 30 seconds, continuing anyway

many of the devices are even still calling back to Home Assistant, but it seems confused. For example, here is the power switch/meter on my coffee pot:

2020-03-24 07:09:51.241 Info, Node062, ApplicationCommandHandler - Unhandled Command Class 0x32
2020-03-24 07:10:20.942 Detail, Node062, Received: 0x01, 0x10, 0x00, 0x04, 0x00, 0x3e, 0x0a, 0x32, 0x02, 0x21, 0x74, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xba
2020-03-24 07:10:20.942 Detail,
2020-03-24 07:10:20.942 Info, Node062, ApplicationCommandHandler - Unhandled Command Class 0x32
2020-03-24 07:10:50.942 Detail, Node062, Received: 0x01, 0x10, 0x00, 0x04, 0x00, 0x3e, 0x0a, 0x32, 0x02, 0xa1, 0x64, 0x00, 0x01, 0xe6, 0x0b, 0x00, 0x00, 0xc6
2020-03-24 07:10:50.942 Detail,
2020-03-24 07:10:50.942 Info, Node062, ApplicationCommandHandler - Unhandled Command Class 0x32
2020-03-24 07:10:51.041 Detail, Node062, Received: 0x01, 0x10, 0x00, 0x04, 0x00, 0x3e, 0x0a, 0x32, 0x02, 0xa1, 0x6c, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x22
2020-03-24 07:10:51.041 Detail,
2020-03-24 07:10:51.041 Info, Node062, ApplicationCommandHandler - Unhandled Command Class 0x32
2020-03-24 07:10:51.144 Detail, Node062, Received: 0x01, 0x14, 0x00, 0x04, 0x00, 0x3e, 0x0e, 0x32, 0x02, 0x21, 0x64, 0x00, 0x00, 0x66, 0x38, 0x00, 0x3c, 0x00, 0x00, 0x66, 0x38, 0x96
2020-03-24 07:10:51.144 Detail,
2020-03-24 07:10:51.144 Info, Node062, ApplicationCommandHandler - Unhandled Command Class 0x32
2020-03-24 07:10:51.241 Detail, Node062, Received: 0x01, 0x10, 0x00, 0x04, 0x00, 0x3e, 0x0a, 0x32, 0x02, 0x21, 0x74, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xba

So it appears the network still exists, but the config in HA is toast. Not sure if there is someway to repair this or if I just need to nuke the entities and have HA rediscover them (not even sure how to do that entirely). Looking for any assistance or guidance I can try. Not really wanting to have to redo my entire network again as I have a lot of devices.

Any help would be very much appreciated.

Also seems you cannot factory reset a Nortek HUSBZB-1, so i’m going to have a super fun time here :\

How was your docker container created? Might be a problem there.

Its the same one i’ve been using for years thats published out on Unraid. No other users that i’ve seen reporting this issue at the moment. Every other aspect of home assistant works, its literally just zwave thats broken at this point.

How does Unraid create the container? There are a ton of users of Unraid and Synology reporting the same exact problem. There was a change in 0.107 how the docker containers are created. Others have reported that re-creating the container fixes it. You cannot override the run command anymore or it will result in multiple copies of HA, which will definitely cause zwave to fail.

Just a suggestion for something to look at before going off and hard resetting everything.

Could you point me to where you are seeing this, I googled, checked both the unraid and these forums and reddit and didn’t see anything obvious. Lots of other issues, but nothing like mine. my config XML is missing nearly all the metadata for the broken devices, so I don’t think this is recoverable unfortunately, but being that basically nothing in my house is working as the Wife is not too happy, just looking to get it sorted and not have this happen again :expressionless:

Examples:

There was also a zwave fix in 0.107.6 that might help. If your zwcfg*.xml file is bad, you can restore from a backup, or delete it and it will be recreated. In the latter case you’ll need to wake up battery devices to get them detected again.