I installed a new iblind this evening and it is appearing in ZWaveJS. It is operational thru Home Assistant just fine. It is operational thru the HA ZwaveJS plugin just fine.
Where it is non-responsive is in MSR for some reason. The entities are there. Adding the node to a Reaction and then attempting to run said Reaction nets me this:
[latest-23242]2023-09-08T05:33:08.454Z <ZWaveJSController:5:ZWaveJSController.js:655> ZWaveJSController#zwavejs _apply_value motion_sensor.state=false [latest-23242]2023-09-08T05:33:08.455Z <ZWaveJSController:5:ZWaveJSController.js:722> ZWaveJSController#zwavejs setting Binary Sensor#zwavejs>11-0.x_zwave_values.Binary_Sensor_Motion to false [latest-23242]2023-09-08T05:33:08.692Z <ZWaveJSController:5:ZWaveJSController.js:360> ZWaveJSController#zwavejs handling node event statistics updated entity Binary Sensor#zwavejs>11-0 [latest-23242]2023-09-08T05:40:23.960Z <ZWaveJSController:INFO> ZWaveJSController#zwavejs performing cover.open on Cover#zwavejs>20-0 with [Object]{ } [latest-23242]2023-09-08T05:40:23.962Z <ZWaveJSController:5:ZWaveJSController.js:1843> ZWaveJSController#zwavejs no implementation mapped; attempting default [latest-23242]2023-09-08T05:40:25.062Z <ZWaveJSController:INFO> ZWaveJSController#zwavejs performing zwave_device.refresh on Cover#zwavejs>20-0 with [Object]{ } **[latest-23242]2023-09-08T05:40:25.063Z <ZWaveJSController:5:ZWaveJSController.js:1843> ZWaveJSController#zwavejs no implementation mapped; attempting default** [latest-23242]2023-09-08T05:40:25.065Z <ZWaveJSController:5:ZWaveJSController.js:294> ZWaveJSController#zwavejs sending #1694151625064<9/8/2023, 1:40:25 AM>: [Object]{ "command": "node.refresh_values", "nodeId": 20, "messageId": 1694151625064 } [latest-23242]2023-09-08T05:40:26.307Z <ZWaveJSController:5:ZWaveJSController.js:360> ZWaveJSController#zwavejs handling node event statistics updated entity Cover#zwavejs>20-0 [latest-23242]2023-09-08T05:40:26.317Z <ZWaveJSController:5:ZWaveJSController.js:360> ZWaveJSController#zwavejs handling node event value updated entity Cover#zwavejs>20-0 [latest-23242]2023-09-08T05:40:26.318Z <ZWaveJSController:5:ZWaveJSController.js:667> ZWaveJSController#zwavejs update node 20 value "0:128:level:" data [Object]{ "source": "node", "event": "value updated", "nodeId": 20, "args": { "commandClassName": "Battery", "commandClass": 128, "property": "level", "endpoint": 0, "newValue": 100, "prevValue": 100, "propertyName": "level" } } [latest-23242]2023-09-08T05:40:26.319Z <ZWaveJSController:5:ZWaveJSController.js:684> ZWaveJSController#zwavejs updating attributes for node 20 value "0:128:level:"=100: [Array][ "battery_power.level", "battery_power.since" ] [latest-23242]2023-09-08T05:40:26.320Z <ZWaveJSController:5:ZWaveJSController.js:698> ZWaveJSController#zwavejs updating attribute battery_power.level with [Object]{ "entity": "20-0", "impl": { "expr": "float( value ) / 100", "valueId": "128:level:" } } [latest-23242]2023-09-08T05:40:26.321Z <ZWaveJSController:5:ZWaveJSController.js:591> ZWaveJSController#zwavejs _apply_value entity Cover#zwavejs>20-0 battery_power.level [latest-23242]2023-09-08T05:40:26.322Z <ZWaveJSController:5:ZWaveJSController.js:655> ZWaveJSController#zwavejs _apply_value battery_power.level=1 [latest-23242]2023-09-08T05:40:26.323Z <ZWaveJSController:5:ZWaveJSController.js:698> ZWaveJSController#zwavejs updating attribute battery_power.since with [Object]{ "entity": "20-0", "impl": { "expr": "time()", "valueId": "128:level:" } } [latest-23242]2023-09-08T05:40:26.323Z <ZWaveJSController:5:ZWaveJSController.js:591> ZWaveJSController#zwavejs _apply_value entity Cover#zwavejs>20-0 battery_power.since [latest-23242]2023-09-08T05:40:26.324Z <ZWaveJSController:5:ZWaveJSController.js:655> ZWaveJSController#zwavejs _apply_value battery_power.since=1694151626324<9/8/2023, 1:40:26 AM> [latest-23242]2023-09-08T05:40:26.325Z <ZWaveJSController:5:ZWaveJSController.js:722> ZWaveJSController#zwavejs setting Cover#zwavejs>20-0.x_zwave_values.Battery_level to 100 [latest-23242]2023-09-08T05:40:26.333Z <ZWaveJSController:5:ZWaveJSController.js:360> ZWaveJSController#zwavejs handling node event value updated entity Cover#zwavejs>20-0 [latest-23242]2023-09-08T05:40:26.334Z <ZWaveJSController:5:ZWaveJSController.js:667> ZWaveJSController#zwavejs update node 20 value "0:128:isLow:" data [Object]{ "source": "node", "event": "value updated", "nodeId": 20, "args": { "commandClassName": "Battery", "commandClass": 128, "property": "isLow", "endpoint": 0, "newValue": false, "prevValue": false, "propertyName": "isLow" } } [latest-23242]2023-09-08T05:40:26.336Z <ZWaveJSController:5:ZWaveJSController.js:324> ZWaveJSController#zwavejs request 1694151625064<9/8/2023, 1:40:25 AM> (node.refresh_values) success notification [latest-23242]2023-09-08T05:40:26.337Z <ZWaveJSController:5:ZWaveJSController.js:684> ZWaveJSController#zwavejs updating attributes for node 20 value "0:128:isLow:"=false: [Array][ ] [latest-23242]2023-09-08T05:40:26.338Z <ZWaveJSController:5:ZWaveJSController.js:722> ZWaveJSController#zwavejs setting Cover#zwavejs>20-0.x_zwave_values.Battery_isLow to false [latest-23242]2023-09-08T05:40:26.558Z <ZWaveJSController:5:ZWaveJSController.js:360> ZWaveJSController#zwavejs handling node event statistics updated entity Cover#zwavejs>20-0Please note the highlighted logpart - no other iblind has this. In addition, in MSR>Entities several show as null rather than having a value as the others do.
69631fbc-db4e-4e8d-aa98-10565811cd93-image.png
I tried deleting all Entities associated with this device from MSR and then refreshing ZWaveJS in Home Assistant and restarting MSR - the Entities return but in the same state/s.
Admittedly, it's been a very long three weeks at work - I could have missed something obvious during setup but I sure can't figure what it is.
Hi @toggledbits
I'm trying to use OWM, apparently, I receive the correct information, but after a while, the MSR disconnects.
My configuration is as follows.
- id: weather enabled: true implementation: OWMWeatherController name: OWM Weather config: # Place your OWM API key here (remember to enable the controller after adding your appid) appid: "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" # How often weather is allowed to be refreshed. This helps limit OWN API use, # to keep you in their good graces (and on their free tier). This value is # in minutes. interval: 30 locations: - id: home name: Home Weather # Set the location by specifying ONE OF: latitude+longitude, OWN city # ID, or location (as postal,country). If none is set, the system # location will be used. latitude: 8.9936 longitude: -79.5197 city_id: 3703443 #location: "30269,us" # Enable "save_full_response" to save the full response (you may have # need to use parts of the response that are not part of the wx # capability in your dashboard widgets). #save_full_response: falseWhen I request a systemctl status reactor I get this message:
root@main:/home/wilson/reactor/logs# systemctl status reactor ● reactor.service - Multi System Reactor Loaded: loaded (/etc/systemd/system/reactor.service; enabled; vendor preset: enabled) Active: active (running) since Fri 2023-09-08 21:03:13 EST; 5s ago Main PID: 194711 (node) Tasks: 11 (limit: 9390) Memory: 82.5M CPU: 2.508s CGroup: /system.slice/reactor.service └─194711 /usr/bin/node app -p Sep 08 21:03:13 main node[194711]: Error: Incompatible serialization data; can't unserialize Sep 08 21:03:13 main node[194711]: at System.unserialize (/home/wilson/reactor/server/lib/Entity.js:624:19) Sep 08 21:03:13 main node[194711]: at /home/wilson/reactor/server/lib/Controller.js:458:70 Sep 08 21:03:13 main node[194711]: at Array.forEach (<anonymous>) Sep 08 21:03:13 main node[194711]: at OWMWeatherController._restoreEntities (/home/wilson/reactor/server/lib/Controller.js:446:36) Sep 08 21:03:13 main node[194711]: at new Controller (/home/wilson/reactor/server/lib/Controller.js:37:45) Sep 08 21:03:13 main node[194711]: at new OWMWeatherController (/home/wilson/reactor/server/lib/OWMWeatherController.js:327:9) Sep 08 21:03:13 main node[194711]: at /home/wilson/reactor/server/lib/Controller.js:93:37 Sep 08 21:03:13 main node[194711]: [latest-23242]2023-09-09T02:03:13.855Z <NUTController:null> Module NUTController v22305 Sep 08 21:03:13 main node[194711]: [latest-23242]2023-09-09T02:03:13.859Z <SystemController:null> Module SystemController v23214 root@main:/home/wilson/reactor/logs# root@main:/home/wilson/reactor# systemctl status reactor ● reactor.service - Multi System Reactor Loaded: loaded (/etc/systemd/system/reactor.service; enabled; vendor preset: enabled) Active: active (running) since Fri 2023-09-08 20:48:36 EST; 2min 58s ago Main PID: 194353 (node) Tasks: 11 (limit: 9390) Memory: 66.0M CPU: 9.493s CGroup: /system.slice/reactor.service └─194353 /usr/bin/node app -p Sep 08 20:48:37 main node[194353]: at System.unserialize (/home/wilson/reactor/server/lib/Entity.js:624:19) Sep 08 20:48:37 main node[194353]: at /home/wilson/reactor/server/lib/Controller.js:458:70 Sep 08 20:48:37 main node[194353]: at Array.forEach (<anonymous>) Sep 08 20:48:37 main node[194353]: at OWMWeatherController._restoreEntities (/home/wilson/reactor/server/lib/Controller.js:446:36) Sep 08 20:48:37 main node[194353]: at new Controller (/home/wilson/reactor/server/lib/Controller.js:37:45) Sep 08 20:48:37 main node[194353]: at new OWMWeatherController (/home/wilson/reactor/server/lib/OWMWeatherController.js:327:9) Sep 08 20:48:37 main node[194353]: at /home/wilson/reactor/server/lib/Controller.js:93:37 Sep 08 20:48:37 main node[194353]: [latest-23242]2023-09-09T01:48:37.017Z <NUTController:null> Module NUTController v22305 Sep 08 20:48:37 main node[194353]: [latest-23242]2023-09-09T01:48:37.024Z <SystemController:null> Module SystemController v23214 Sep 08 20:48:37 main node[194353]: [latest-23242]2023-09-09T01:48:37.475Z <Timer:null> Timer#rule-l7ujwva5 just a note: I'm setting a delay of > lines 1-20/20 (END)I can see this information in the log, but I don't understand what I should do to fix it.
[latest-23242]2023-09-09T02:03:13.838Z <DynamicGroupController:null> Module DynamicGroupController v22313 [latest-23242]2023-09-09T02:03:13.845Z <Structure:INFO> Structure#1 loading controller interface weather (OWMWeatherController) [latest-23242]2023-09-09T02:03:13.847Z <OWMWeatherController:null> Module OWMWeatherController v22294 [latest-23242]2023-09-09T02:03:13.850Z <Controller:WARN> OWMWeatherController#weather failed (1) to restore entity controller_all: [Error] Inco> [latest-23242]2023-09-09T02:03:13.850Z <Controller:CRIT> Error: Incompatible serialization data; can't unserialize [-] Error: Incompatible serialization data; can't unserialize at Group.unserialize (/home/wilson/reactor/server/lib/Entity.js:624:19) at /home/wilson/reactor/server/lib/Controller.js:458:70 at Array.forEach (<anonymous>) at OWMWeatherController._restoreEntities (/home/wilson/reactor/server/lib/Controller.js:446:36) at new Controller (/home/wilson/reactor/server/lib/Controller.js:37:45) at new OWMWeatherController (/home/wilson/reactor/server/lib/OWMWeatherController.js:327:9) at /home/wilson/reactor/server/lib/Controller.js:93:37 [latest-23242]2023-09-09T02:03:13.851Z <Controller:WARN> OWMWeatherController#weather failed (1) to restore entity default: [Error] Incompatibl> [latest-23242]2023-09-09T02:03:13.851Z <Controller:CRIT> Error: Incompatible serialization data; can't unserialize [-] Error: Incompatible serialization data; can't unserialize at Entity.unserialize (/home/wilson/reactor/server/lib/Entity.js:624:19) at /home/wilson/reactor/server/lib/Controller.js:458:70 at Array.forEach (<anonymous>) at OWMWeatherController._restoreEntities (/home/wilson/reactor/server/lib/Controller.js:446:36) at new Controller (/home/wilson/reactor/server/lib/Controller.js:37:45) at new OWMWeatherController (/home/wilson/reactor/server/lib/OWMWeatherController.js:327:9) at /home/wilson/reactor/server/lib/Controller.js:93:37 [latest-23242]2023-09-09T02:03:13.852Z <Controller:WARN> OWMWeatherController#weather failed (1) to restore entity system: [Error] Incompatible> [latest-23242]2023-09-09T02:03:13.852Z <Controller:CRIT> Error: Incompatible serialization data; can't unserialize [-] Error: Incompatible serialization data; can't unserialize at System.unserialize (/home/wilson/reactor/server/lib/Entity.js:624:19) at /home/wilson/reactor/server/lib/Controller.js:458:70 at Array.forEach (<anonymous>) at OWMWeatherController._restoreEntities (/home/wilson/reactor/server/lib/Controller.js:446:36) at new Controller (/home/wilson/reactor/server/lib/Controller.js:37:45) at new OWMWeatherController (/home/wilson/reactor/server/lib/OWMWeatherController.js:327:9) at /home/wilson/reactor/server/lib/Controller.js:93:37 [latest-23242]2023-09-09T02:03:13.853Z <Structure:INFO> Structure#1 loading controller interface nut (NUTController) [latest-23242]2023-09-09T02:03:13.855Z <NUTController:null> Module NUTController v22305 [latest-23242]2023-09-09T02:03:13.855Z <Controller:INFO> Loaded NUTController version "0.1.22305"; Patrick Rigney/Kedron Holdings LLC <patrick@> [latest-23242]2023-09-09T02:03:13.857Z <Structure:INFO> Structure#1 loading controller interface reactor_system (SystemController) [And finally, I have a second house, with the same OWM configuration I made today, and it doesn't show any errors.
Please, your traditional help.
Thanks.
After a prolonged power cut, the UPS was unable to cope with the weather and shut down the entire system.
When I returned, I had the problem that the host time displayed in the browser and automatically used in the rules was not correct.
I searched for a topic here on this forum that talked about it, but it referred to a container installation. There was a recommendation to go into the storage/states directory and delete the files and restart MSR, I did it and it didn't solve the problem.
There was also a recommendation to go directly to the browser to check the time, and I did so and I see that the time is correct.
8ad429c8-ea5f-45b4-8670-7d0353a50a0e-image.png
As you can see on the screen above:
MSR browser shows the correct browser time 21:23, but host time 13:02; In the browser below the direct query and the host time is correct 9:22:54 PM; Setting the VM date command on the third screen, the time is 9:23:00 PM.I've opened the entire reactor.log file and searched for any errors, and there aren't any.
^L [latest-23242]2023-09-15T02:11:12.928Z <app:null> Reactor build latest-23242-5ee8e1d4 starting on v18.12.0 [latest-23242]2023-09-15T02:11:12.929Z <app:null> Process ID 1807 user/group 1000/1000; bare-metal; platform linux/x64 #179-Ubuntu SMP Mo> [latest-23242]2023-09-15T02:11:12.929Z <app:null> Basedir /home/leonardo/reactor; data in /home/leonardo/reactor/storage [latest-23242]2023-09-15T02:11:12.929Z <app:null> NODE_PATH=/home/leonardo/reactor [latest-23242]2023-09-15T02:11:12.933Z <app:INFO> Configured locale (undefined); selected locale(s) en_US.UTF-8 [latest-23242]2023-09-15T02:11:12.955Z <app:INFO> Loaded locale en-US for en-US [latest-23242]2023-09-15T02:11:12.958Z <Structure:null> Module Structure v23172 [latest-23242]2023-09-15T02:11:12.959Z <Capabilities:null> Module Capabilities v22356 [latest-23242]2023-09-15T02:11:12.969Z <Capabilities:NOTICE> System capabilities loaded from core distribution, data version 23914 revisi> [latest-23242]2023-09-15T02:11:12.978Z <Plugin:null> Module Plugin v22300 [latest-23242]2023-09-15T02:11:12.981Z <TimerBroker:null> Module TimerBroker v22283 [latest-23242]2023-09-15T02:11:12.983Z <Entity:null> Module Entity v22353 [latest-23242]2023-09-15T02:11:12.986Z <Controller:null> Module Controller v23069 [latest-23242]2023-09-15T02:11:12.992Z <default:null> Module Ruleset v22293 [latest-23242]2023-09-15T02:11:12.992Z <default:null> Module Rulesets v22146 [latest-23242]2023-09-15T02:11:12.996Z <GlobalExpression:null> Module GlobalExpression v23211 [latest-23242]2023-09-15T02:11:13.003Z <Predicate:null> Module Predicate v23093 [latest-23242]2023-09-15T02:11:13.005Z <AlertManager:null> Module AlertManager v22283 [latest-23242]2023-09-15T02:11:13.007Z <Rule:null> Module Rule v23107 [latest-23242]2023-09-15T02:11:13.009Z <GlobalReaction:null> Module GlobalReaction v22324 [latest-23242]2023-09-15T02:11:13.010Z <Engine:null> Module Engine v23231 [latest-23242]2023-09-15T02:11:13.012Z <httpapi:null> Module httpapi v23058 [latest-23242]2023-09-15T02:11:13.019Z <wsapi:null> Module wsapi v23172 [latest-23242]2023-09-15T02:11:13.019Z <app:NOTICE> Starting Structure...As you can see, the time in the log is 02:11, which is correct, because I'm in time zone -5, which means it's 09:11PM for me.
I've restarted the VM, the computer hosting the VM, and the MSR a few times and the same difference remains. What do you recommend for the next step in understanding and solving the problem?
Thanks.
Build 21228 has been released. Docker images available from DockerHub as usual, and bare-metal packages here.
Home Assistant up to version 2021.8.6 supported; the online version of the manual will now state the current supported versions; Fix an error in OWMWeatherController that could cause it to stop updating; Unify the approach to entity filtering on all hub interface classes (controllers); this works for device entities only; it may be extended to other entities later; Improve error detail in messages for EzloController during auth phase; Add isRuleSet() and isRuleEnabled() functions to expressions extensions; Implement set action for lock and passage capabilities (makes them more easily scriptable in some cases); Fix a place in the UI where 24-hour time was not being displayed.Good afternoon,
I'm running my MSR instance on an old NAS server running Fedora 37 Server. This is an older Core i5-4690K server that has served me well overall, but it's showing it's age, and will often reboot into maint mode if it applies updates overnight. I've been trying to troubleshoot, but it's also in my office (makes my office hot), and I'm looking at getting either a QNAP or Syslogy rack mount NAS to replace it.
What I'm wondering is what the process would be to migrate my current MSR config and move it to a different platform. I have quite a large ruleset that I don't want to have to recreate. I know I can run it in a container on either platform, but I'm not sure how I would migrate all my rules over. Is there one config file that has all of the rules, or could I simply copy my /reactor directory over to a enw platform?
I am also considering moving the MSR instance to a stand alone RPI 4 with a SSD. This way, while I'm migrating all of my files off of my NAS, I'm not loosing Reactor's functionality until I have whichever rack mount NAS I go with up and running, with all of the data migrated.
Thoughts?
Hello,
I've installed MSR in a Docker container on my Synology NAS. I'm able to access my Vera, but I'm having trouble using the Reactor import within the Docker environment. I'm using the Docker terminal, and I've navigated to "/opt/reactor". However, when I try to run the command, it's telling me I'm not in the installation directory, even though I can see the "tools" directory and the "apps.js" file there. I've also tried changing the file permissions, but nothing seems to work. How do you manage this in a Synology Docker container? Do you use the terminal inside Docker, or SSH into the Synology itself? I'm not finding the "/opt/reactor" path, but I see a "/var/reactor" directory, which I placed in my storage folder.
ERROR: This script must only be run from the Reactor app directory.
Maybe I'm missing something, but this appears to be a bug.
I'm troubleshooting a rule to see if one of the system participants is arriving (presence changes from not home to home). I want this arriving state to be less than 5 minutes (300 seconds), to differentiate between being home for a longer period of time, and just getting home.
In the restrictions sections of the Trigger, I set "Condition must be sustained for less than 300 seconds", and I save the trigger.
After I save and exit, I'm looking at the rule, and it's changed itself to x_hass.state == home; for at least 300 secs. I go back into edit the rule, and it in fact shows "at least" from the drop down.
This logic is of course failing and I can't seem to get it to work.
Changed to less than, and saved:
10a57ff9-4d8c-46b3-b594-399ef2a6d320-image.png
Exited:
a81587d6-f0fa-4518-ad00-a9f601f14509-image.png
Back in rule:
b8e27dd2-e6e7-4f93-844e-83e2b11e976d-image.png
I'm running latest-23218-ebab7a1b.
I added the Proximity integration to HA to create automations based on if a person is on vacation or not. The integration creates an entity and its x_hass.state is the distance to the zone you have setup in HA. For some reason MSR sees that value as a string so I am not able to use the<, >, etc operators as a trigger. Is there anyway this could be fixed without needing to convert the string in a local expression?
MSR 23218
HA 2023.8.3
Entity Attributes if needed
x_hass.domain="proximity" x_hass.entity_id="proximity.arman_proximity" x_hass.services=["proximity"] x_hass.state="0" x_hass_attr.dir_of_travel="unknown" x_hass_attr.friendly_name="arman_proximity" x_hass_attr.nearest="Arman’s iPhone" x_hass_attr.unit_of_measurement="km"MSR Version: latest-23218
Running as Docker Container : Docker version 24.0.5, build ced0996
OS: Linux 15BPMSR1 6.1.26-05272-g26c406245a2c #1 SMP PREEMPT_DYNAMIC Thu Apr 27 10:15:40 UTC 2023 aarch64 aarch64 aarch64 GNU/Linux
SMTP notifications had been working fine and I am unsure exactly what action(s) stopped it from working.
I have verified that the SMTP server configured in notifications.yml is working properly.
To duplicate the problem I am forcing an SMTP email notification using the set reaction play button, the reactor.log shows the following errors:
[latest-23218]2023-08-21T14:23:22.462Z <wsapi:ERR> [WSAPI]wsapi#1 can't queue reaction rule-lgxmrba4:S: [Error] Engine has not been started [-] [latest-23218]2023-08-21T14:23:22.463Z <wsapi:CRIT> Error: Engine has not been started [-] Error: Engine has not been started at Engine.queueReaction (/opt/reactor/server/lib/Engine.js:1494:46) at WSAPI.clientMessage (/opt/reactor/server/wsapi/wsapi.js:782:64) at WebSocket.<anonymous> (/opt/reactor/server/wsapi/wsapi.js:301:157) at WebSocket.emit (node:events:527:28) at WebSocket.emit (node:domain:475:12) at Receiver.receiverOnMessage (/opt/reactor/node_modules/ws/lib/websocket.js:1068:20) at Receiver.emit (node:events:527:28) at Receiver.emit (node:domain:475:12) at Receiver.dataMessage (/opt/reactor/node_modules/ws/lib/receiver.js:517:14) at Receiver.getData (/opt/reactor/node_modules/ws/lib/receiver.js:435:17)I have tried:
OS level: apt udate and apt upgrade and reboot
docker pull toggledbits/reactor:latest-aarch64 latest-aarch64: Pulling from toggledbits/reactor Digest: sha256:d8a019d457c9fef61f7c3e6908db1f1d0a2d14b6f7c5e9d4fc536c791503d152 Status: Image is up to date for toggledbits/reactor:latest-aarch64 docker.io/toggledbits/reactor:latest-aarch64docker stop reactor
docker start reactor
docker compose down
docker compose up -d
To no avail.
I feel like I'm missing something obvious but can't put my finger on it.
Thank you in advance for any guidance.
-bh
I have a rule that's driving my pool heater to match the solar excess. In the previous iteration, I checked for the status every minute, but I wanted something different and smarter. The logic is basically:
run continuously from 9:15 to 18:15 check for conditions, and if there's solar excess, use the power to turn on the heater, then wait 1 minute to check again turn off the heater, then wait 5 minutes to prevent frequent on/off cyclesI thought that this could be very similar to Reactor-Ex machina's logic, so I crafted a rule that's basically doing the same: an infinite loop from 9:15 to 18:15 that's checking for status and has a delay based on the heater's condition.
All went well, but after a couple of hours, I got this:
Reaction task "49683" failed due to an exception. Please see the log. rule-liy7z3gf:S (Pool-Heater-Solar) step 0: Error: Task has run too much! Runaway?and the rule was reset (and the heater was left on). Is that by design and I should adopt another strategy? What are others doing when they have a state machine?
With the EV charger I did two groups of condition, each triggering every x minutes based on the charger status, but this seems less smarter and very rudimental. Any suggestion is appreciated. Thanks.
I have a Tripp-Lite UPS (https://www.tripplite.com/smartpro-lcd-120v-1500va-900w-line-interactive-ups-avr-tower-lcd-usb-10-outlets~smart1500lcdt) that connected by a USB cable to my computer, reports various power supply and failure information, as shown in the panel below.
b9f0ea34-d654-4b26-b968-4f8cf0a4e0c1-image.png
afa73913-5318-42c6-af9a-02d36e2e6ea1-image.png
The application I installed runs Java and opens the console to send the information.
Has anyone tried to bring the information into the MSR? Or even a VeraPlus?
Thanks.
Hi
I am stuck trying to extract the "SignalLevel" value which is the WIFI Signal Strength off my Foscam via a http request. The request returns this, which isn't Json formatted data I don't think.
<CGI_Result> <result>0</result> <SignalLevel>52</SignalLevel> </CGI_Result>I created a rule that saves the http response into a Global Expression
872f534e-5899-43e4-ade5-f2b8ab54a412-image.png
But I am struggling to work out what is the correct expression code to extract the value?
Can anyone more knowledgeable please help?
Thank you.
Hi
Today in MSR alerts I see 100 alerts saying that the devices on my Ezlo controllers no longer exist.
They do still exist. I checked some of these devices in the Ezlo Online API Tool doing a hub.devices.list query.
I then checked that the ID numbers were the same for the devices and they are the same device ID's still.
So my devices do still exist with the same device ID's it seems.
So not sure why MSR has over 100 alerts saying otherwise.
9d7c4732-40f3-4551-b8d4-e24e0e8dda10-image.png
I then looked in Entities selected my Ezlo Plus controller and it listed all my devices OK I think and nothing looked out of place.
Thanks
I just discovered there is a secret webpage I didn't know about for my Asus Merlin router that when I run it in a browser gives me all the routers current temperatures.
http://192.168.0.1/ajax_coretmp.asp
249c6e47-38ae-4812-b149-da1fd22072cb-image.png
I have to be logged into the router to access it however.
I tried to create an MSR rule with a HTTP request to store all this data in to a local expression.
But it looks like its not passing or accepting the username and password for my router, I tried Basic and Digest.
I just get this back in the local expression in the rule.
(string) "<HTML><HEAD><script>window.top.location.href='/Main_Login.asp';</script>\n</HEAD></HTML>\n"The data returned in the browser looked like this:
curr_coreTmp_wl0_raw = "43°C"; curr_coreTmp_wl0 = (curr_coreTmp_wl0_raw.indexOf("disabled") > 0 ? 0 : curr_coreTmp_wl0_raw.replace("°C", "")); curr_coreTmp_wl1_raw = "49°C"; curr_coreTmp_wl1 = (curr_coreTmp_wl1_raw.indexOf("disabled") > 0 ? 0 : curr_coreTmp_wl1_raw.replace("°C", "")); curr_coreTmp_wl2_raw = "disabled"; curr_coreTmp_wl2 = (curr_coreTmp_wl2_raw.indexOf("disabled") > 0 ? 0 : curr_coreTmp_wl2_raw.replace("°C", "")); curr_coreTmp_wl3_raw = "disabled"; curr_coreTmp_wl3 = (curr_coreTmp_wl3_raw.indexOf("disabled") > 0 ? 0 : curr_coreTmp_wl3_raw.replace("°C", "")); curr_cpuTemp = "66.495"; fanctrl_info = "";I know how to do this for webpages and sites that give back JSON data, however this data I am not sure what format it is, Javascript ? The page is a .asp ajax page.
So not sure if this is possible ?
I really wanted to monitor the CPU temps on my router as its been getting very hot, even with extra USB fans now on the back of it. Think its a firmware bug maybe in the Asus firmware as others have been talking about it online.
Thanks.
So, I've just installed two venetian blinds and I want to control their lamellas with Fibaro FGRM222.
According to the old place, this could be accomplished with some Zwave raw messages and I'm OK with it.
Since I've moved (mostly) everything virtual in MSR, I'm now ready to add a virtual device with actions.
Looking at the docs, it seems possible, but it's not documented. @toggledbits any hints? I want to define a cover and send Zwave commands to the existing x_vera_svc_micasaverde_com_ZWaveNetwork1.SendData action. Thanks.
@toggledbits,
I'm not sure if this is something you've looked into (or something I'm completely missing in the installation documentation), but would it be possible to add user authentication to the MSR frontend? On my Home Assistant instance, I have external access set up using Cloudflare, and I can add other instances within my network to that Cloudflare configuration, but without any sort of authentication to access MSR, I obviously can't open it to the internet, unless I want random people screwing around with my home automation.
I've just updated to 23218 and now I'm getting this:
27a959bd-486b-41ae-8137-c5f456044a82-image.png
There seems to be related to global variables. The first and last ones are global reactions, while the second is a normal one. All are setting global variables.
Hey @toggledbits was wondering if we could have the latest reactor version number as an attribute for the Reactor System entity. It would be useful for me so I can use it for my wget command to automatically download the latest .tar.gz file. Ideally the complete version number like this 23171-85463e02.
@toggledbits ,
I was just taking a look at some of my scheduling to adjust for a new work schedule, and I noticed that the checkboxes for days of the week appear to be off.
75e1b936-1693-4db1-8ed5-ecfae57d3abf-image.png
if you'll notice in the screenshot, Sun - Tuesday are good, but then Sat comes next, then Wed, then Sat again, then Wed again. No TGIF in this week.
I know this isn't how I set up the rules (which would explain the behavior I was investigating).
Edit: BTW, I checked all my Week Day entries, and they are all like this.
I'm running latest-23196-40ef07f4 on Fedora 37 Server w/o Docker.
Quality of Life Request: Update Button
-
Massive upvote on all 3. Just like back in the old Vera days
(Unsure how easy update would be with the various ways of implementing MSR inside/outside docker etc) - unless a standard install method is specified, and one of the features of that is 'update' capability.
Docker-Compose please -
@Cadwizzard said:
Just like back in the old Vera days
How soon we forget the tales of bricked Veras. Who among us didn't have a little sense that they were playing Russian roulette every time we hit that button?
Unsure how easy update would be with the various ways of implementing MSR inside/outside docker etc
OK. He hits it on the head here. Let me explain some of the complications and my reservations around this.
The biggest pitfall is for docker users, IMO; that's the majority of you. The first thing you need to understand about docker is that the image and the container are separate objects in the system. The container is created from the image, but it's basically a copy, not linked in any meaningful way. The container can change, so that's good — I can download a release package and apply it to the container, restart, and the container will now be running the new release files. Unfortunately, that has no bearing on what happens to the image. Changing files in the container does nothing to the image. So let's take a scenario... @tunnus (Docker, Synology) downloads the image for Reactor 22274 and creates a container for it, so he's now running 22274. A little later, 22291 is released, so he hits the handy, flashy new "Upgrade" button and the container is upgraded in place. Perfect. Except, not... his image is still 22274. Stay with me now... In all likelihood, because of the "ease" of the automated upgrade, tunnus never needs to download a new image again (so he thinks), so he never bothers (it's a pain anyway on Synology, I'll agree). So build 22293 comes along, and then 22302, and then 22305, and then 22308, and he upgrades to all of them using the automated process, but the image is still sitting there on his NAS at 22274. The problem strikes if, for any reason (DSM major upgrade?), he decides to reset and rebuild the container, or delete it. He will get.... 22274. Because that's the image he has.
Can I make docker download the newer image as part of the upgrade process? No. Reactor is running inside the container, and the container, by definition, contains Reactor and keeps it from doing anything external to the container (except the limited data volume that's specifically created for the single purpose). So the running Reactor instance has no ability whatsoever to cause docker/DSM to pull later images. Pulling a new image and rebuilding the container is the real "right" way to upgrade, but it's not possible to automate it from within the container itself (and it's darned clunky in Synology's UI, unfortunately).
It's not hard to imagine that this problem would not bite him for months or years. But when it bites, it has the potential to bite hard. Imagine along Reactor's evolutionary path from 22274 to a future 24107 (released in 2024, all automated updates between, no image refreshes), there are changes that needed to be made to the data structures of rules, reactions, stored states, etc. (not at all hard to imagine, it actually happens all the time). It is easy, although sometimes a bit cumbersome, for me to provide forward compatibility: to make sure that newer versions of code read the old data and upgrade the structures, and the mechanisms for those upgrades remain in the code for some time. But there is no way under the sun for 22274, now running once again unexpectedly, to know what to do with data from the future 24107 build, and there's a chance it could do something really bad to it. Now tunnus has an old version running in his container with corrupt data. I hope he has a backup.
I'll take the opportunity to say that this is a cautionary tale for all of you who stay on older builds. I keep the code that reads and upgrades the data, when needed, for a while, so that people who skip a few upgrades can safely do so and "jump in time" when they are ready to apply a new build, but I don't and won't keep that code forever; it becomes a maintenance nightmare and it's beyond my available time and sensibilities to test every possible combination of upgrades between versions. If you're running on a Reactor that's more than a year out (21307 or earlier), you're playing with fire as far as I'm concerned, and you should not expect a smooth upgrade when you get around to it. You may need to upgrade to an interim build still available, which works for bare-metal, but isn't an option for docker users. And before the "I can't have something like that in my home" people start in here, please know that I'm sorry that the free software I offer you and for which I provide ready, quick, and free ongoing support (and upgrades) isn't perfectly to your liking. If you don't like the way it works, you have alternatives, and I fully support your freedom to choose them.
To continue with @Cadwizzard's point: this is equally or more egregious, unfortunately, for docker-compose users, because up to this point, the recommended way for stopping Reactor when using docker-compose is to run
docker-compose down
. This causes Reactor to stop, but also deletes the container. Any upgrades applied to the container are lost in that instant, because the container is discarded. When you later rundocker-compose up -d
, the container is re-created from the (old) image, and will be whatever version that image is. Maybe not a disaster, maybe it is. This could be addressed by retraining docker-compose users to usedocker-compose stop
rather thandocker-compose down
, but the distinction would need to be taught (and learned) as both are useful, and the infrequency of use of these commands would likely suffer from brain-drain over time (i.e. when to use which and what their side effects could be/will be lost on the user a few months from now). But it's such a subtle distinction that people will shoot themselves in the foot easily and regularly, I fear.Bare-metal is somewhat easier, because at least the process can be assured it's writing on the one and only (relevant) image, in the install directory, so that's a bit of relief. Unfortunately, a lot of people really don't understand Linux file permissions, their relations to users and groups, etc., and routinely goof up the permissions of files all over their system, including in the Reactor install directory. This isn't a problem for them after the first "fix," because thereafter they do the manual upgrades the same way, logged in as the same user (in some cases, even as root, which is a serious no-no), and so it works for them as that user in that case, good enough. But for an automated process running in an unprivileged environment, it can mean that some files aren't writable, and the upgrade only half-happens... the upgrade process crashes, some files are new and some are old, and the Reactor install is basically dead and broken. I can't fix the permissions from the running instance, because it's running as an unprivileged user (well, hopefully; woe unto those who run anything as root). The user then manually applies the upgrade to recover the system, which goes fine because of course he's running the privileged user with the right permissions. A bug/post for the upgrade process then gets reported, and I then spend hours or days going back and forth, digging through the user's 3,000+ files in a typical Reactor install, looking for the broken ones and teaching the user how to fix them. (Permissions and their potential brokenness is also an issue for doing automated backups/restores, since that was mentioned as well.)
Oh, and then there's Windows. I won't even start. I've already written a book (again).
With regard to the suggestion of a standard install method: (a) there is no "one size fits all" — what works for Ubuntu doesn't work for Synology DSM or QNAP, and certainly not for Windows; and (b) the install methods that are recommended are all carefully documented; experience shows that I can write out every detail I can think of, and what actually happens on the user's system is 100% of that or some amount less, or the user has some condition in their system/environment that I could not/did not anticipate that causes problems. My preferred method for most users is docker (and specifically, with docker-compose), because the container strategy removes some of these risks, but that's not always the easiest for their environment (e.g. Synology has docker but no -compose), and the accepted mechanisms for upgrading containers in the docker world in general are ironically exactly the subject of complaints by OP and others here, despite the relative ubiquity and ease of these mechanisms.
The point is, there is no panacaea here. You run these systems, not me. You do things I have no knowledge of, and sometimes those things bite back. The majority of my time supporting this product is troubleshooting your environments, not my code (I'm not saying I'm perfect — I make mistakes and bugs are a reality, but they're not the majority of support issues here in terms of time spent). Anything and everything I do in the system is looked at not just through the lens of whether it's convenient for the user, but very much through the lens of supportability. There are lots of features I get asked to do, and as you've seen (even recently), there are some that I refuse to consider simply because it would make the system less supportable, in my view. As features get added, not only is the usability of the system required to improve, but its quality is expected to improve as well (fewer bugs, better support, etc.) — those are my expectations, which I'm sure you share. If one doesn't consider supportability (and that means both in user support and code maintenance/reliability/scalability), one ends up with a lot of features that nobody asked for, don't work, and aren't usable (I can point you in the direction of such a product as an alternative if you're really interested in that).
There is a running, hidden upgrade process in the current build. I've been experimenting with this for a bit, getting to learn it, and discovering these issues. It's not that I won't consider making it available; I'm still studying it, and pondering the wisdom of it. Maybe sometimes I worry too much about things like this, I don't know. But when it goes wrong, there's nobody but you and me to fix it, and there's a lot of you and only one of me, so as I said in another recent conversation, handing out something that feels like a grenade with no pin sometimes doesn't seem like the best idea to me, and there are probably other things this system needs to do that I can better spend my time on. Maybe this is one of those things.
I'll leave this one up to you guys. If you can tolerate these side effects, I'll release the feature. But know that if you break your install because (docker) you somehow delete the container and recreate it from an old image, or (bare-metal) your install has broken permissions or other issues that the upgrade process can't work through, my answer will be short: that's a risk you accepted, do a clean reinstall from a current image, restore your config/state backup, and start over.
-
One more class of knowledge! Really the desire to have an automatic process would be very good, but your explanation makes clear the difficulty and risks, and I do not want to have it. As you said, the errors we generate are enough, I don't want to implement more risks. I think almost everyone here has a wife, so better to stay in the safety of the system working.
Well let's remove this from wishlist request, and could you share this list so we know everything that's coming in the future? Also put an item, display the status widget, in a window/iframe inside the HE dashboard
-
@toggledbits this makes tons of sense why anyone should want an update button mainly Docker users.
In terms of bare metal users, say if a user messed with their files permissions enough that it would cause issues when updating Reactor, wouldn't they run into the same errors even if they manually updated or used the update button? I wouldn't mind an update button for bare metal users, since from your explanation seems like a possible issue with won't come up with the actually update process itself, it can come with something else (like file permissions etc). Meaning that they'd run into these errors even if they manually updated Reactor like we do now.
Not arguing though, its a fairly low level request from me. I can clearly see why an update button for Docker users could be a slow and silent death. As @CatmanV2 for bare metal the update process really only takes 90 seconds ahah.
-
@pabla said in Quality of Life Request: Update Button:
wouldn't they run into the same errors even if they manually updated or used the update button?
Not necessarily... some users... I've seen it... will run into permission problems and their answer, not understanding the problem or how to fix it, is to use
sudo tar xvf
to just lay tracks over everything. This would eliminate the permissions problem unpackaging the archive, but new files may become root-owned, which isn't right but the code doesn't care as long as its readable. If their umask allows world-readable files (and 022 is a common default that does exactly that), the Reactor runtime will never know permissions are broken, because every file it needs is readable without consideration of ownership. The un-tar'ing doesn't touchlogs
,config
, etc. so any permissions there aren't relevant and aren't changed. And because some of the files are now root-owned that shouldn't be, the permissions problem has been made worse and again, unless they are truly fixed the right way, thensudo
will continue to be the only way upgrades will succeed. It perpetuates and exacerbates.I really get how painful the docker upgrades are on Synology. I'm guessing QNAP is probably not much different, and I think several people have been bitten by Portainer oddities regardless of platform.
The process just needs more thought. I could, for example, from the next build onward, prevent the system from starting if the config and data are from a newer version. The problem there is that it needs to be detected early in startup, and if the system can't use the data, it has to exit hard, because it can't run without any data at all, and it can't touch what it has. There would be no UI feedback other than "DISCONNECTED" (i.e. the behavior when Reactor can't start). A "click-to-upgrade" to fix it wouldn't be an option because the UI would not be running, so a manual upgrade would be required at that point. And maybe that's OK? Maybe that's such an extreme/infrequent circumstance that it should be that way? A manual upgrade once in a blue moon may not be so bad... I don't know... looking for feedback... trying to figure it out...
-
I personally do not think the update process on Synology docker is that bad. A few more clicks than an easy button but not horrible. All my other docker containers are updated the same way. I like the docker image though. I am not familiar with the other platforms so I can’t comment on those update processes.
-
C CatmanV2 referenced this topic on
-
I too am happy with the current process. Super fast for me. I run everything under Synology/Docker. I no longer have issues upgrading containers such as Reactor, HA, etc. since I switched to Portainer several months ago. So not sure what those "Portainer" oddities are/were. Something I should keep an eye on? Or I have just been lucky?
-
Some input from a Windows user.
An update button would of course be a nice to have feature, but I also agree with several other here. A "normal" update, aka don't need new dependencies, just take a short moment to install.
Were I usually stumble is when an update of dependencies are needed. That have taken me hours of search-try--error-tryagain before getting that to work sometimes.My dream would be to have a "windows installer" for MSR, that checks dependencies, install a systemservice etc.
Over time I think that would be a safer/more stable way, with fewer user errors.With this said, I can understand really understand that @toggledbits need to handle this "his way" to be able to support differen't enviroments (and users
)
-
I don't know if this helps for other Docker users, but not long after I got started with Docker I found Portainer, and I've been running it alongside Reactor and my other containers on my Raspberry Pi 4. With Portainer, there may not be a one-step update button, but I find it makes updates much easier.
I just updated Reactor to the latest. All I had to do was go to the Portainer URL in my browser, then
- Click on the Reactor container in the Containers list
-
Click 'Recreate"
-
Toggle "Always pull new image" on the window that pops up
- Click "Recreate"
It isn't one click, but it can be done in a browser tab from any machine with network access to the Docker host. No VNC/SSH into the machine, no Docker commands to run from the command line.
Portainer also has links to view the container logs and to open a command window in the container, which I use all the time. You can also use the "duplicate/edit" button to change or add environment variables while updating, which is how I added the NODE_PATH a few updates back.
-
Thanks for the Portainer explanation, I'm certain I've had a spell cast on me. I'll try again once the Pi400 become available once more.
Lastly, I think the point has been lost, it's about QOL, not about how easy it is to do in another way.
From my perspective if it isn't easy to use by 98% of the public then it's too much trouble and they might look at it then discard it for another solution.
The comments so far are from users who are in the 2% and are happy to tinker. I'm happy for you.
If anyone wants to see how Consumer friendly software should be to set up, then have a look at Homeseer4. Update ...no problem with 1 click. -
@black-cat Isn't homeseer a walled garden like Hubitat, Ezlo, etc.? You buy their hub and live within their infrastructure.
That's not MSR. MSR works on various OS/hardware and communicates with multiple hubs.
Whilst I appreciate your POV, it's not apples>apples comparison you're making here.
-
@gwp1 said in Quality of Life Request: Update Button:
You buy their hub and live within their infrastructure.
Nup, you can use any old Laptop or RasPi. Runs on Windows or Lynx. i'd love to promote MSR to Homeseer users but it lacks the simplicity hence the backing of the request.
Realistically, I'm not going to see it happen which is a shame as Patrick has put a lot of time into development for the 2%. -
@sweetgenius I agree, Synology Docker container upgrade process is not too bad.
I frequently keep both MSR and Synology UI open on separate browser tabs and either do a quick upgrade using "reset" or a bit careful upgrade using "duplicate settings" and retaining old container as a backup/rollback option.
Originally I favored a simple update button for MSR, but after Patrick's explanations I realized it's not that simple after all.