Morning, experts. Hard on learning about the internet check script in MSR tools, I was wondering what suggestions anyone has about a local (i.e. non-internet dependent) notification method.
This was prompted by yesterday's fun and games with my ISP.
I've got the script Cronned and working properly but short of flashing a light on and off, I'm struggling to think of a way of alerting me (ideally to my phone)
I guess I could set up a Discord server at home, but that feels like overkill for a rare occasion. Any other suggestions?
TIA
C
Hi,
I'm trying to integrate the sonos-mqtt (https://sonos2mqtt.svrooij.io/) with the MSR and it's coming along nicely so far.
But cannot wrap my head around how to define custom capabilities in MQTT templates. I need this for the TTS announcements and similarly for the notification sounds where I would pass the sound file as parameter.
So this is what I have in the local_mqtt_devices.yaml
capabilities: x_sonos_announcement: attributes: actions: speak: arguments: text: type: string volume: type: int delay: type: intAnd this is the template:
templates: sonos-announcement: capabilities: - x_sonos_announcement actions: x_sonos_announcement: speak: topic: "sonos/cmd/speak" payload: expr: > { "text": parameters.text, "volume": parameters.volume, "delayMs": parameters.delay, "onlyWhenPlaying": false, "engine": "neural" } type: jsonSo the speak action should send something like this to topic sonos/cmd/speak
{ "text": "message goes here", "volume": 50, "delayMs": 100, "onlyWhenPlaying": false, "engine": "neural" }At startup the MSR seems to be quite unhappy with my configuration:
reactor | [latest-25016]2025-02-09T08:19:59.029Z <MQTTController:WARN> MQTTController#mqtt entity Entity#mqtt>sonos-announcement unable to configure capabilities [Array][ "x_sonos_announcement" ] reactor | i18n: missing fi-FI language string: Configuration for {0:q} is incomplete because the following requested capabilities are undefined: {1} reactor | i18n: missing fi-FI language string: Configuration for {0:q} has unrecognized capability {1:q} in actions reactor | Trace: Configuration for {0:q} is incomplete because the following requested capabilities are undefined: {1} reactor | at _T (/opt/reactor/server/lib/i18n.js:611:28) reactor | at AlertManager.addAlert (/opt/reactor/server/lib/AlertManager.js:125:25) reactor | at MQTTController.sendWarning (/opt/reactor/server/lib/Controller.js:627:30) reactor | at MQTTController.start (/var/reactor/ext/MQTTController/MQTTController.js:268:26) reactor | at async Promise.allSettled (index 0) Configuration for "sonos-announcement" has unrecognized capability "x_sonos_announcement" in actions Controller: MQTTController#mqtt Last 10:21:37 AM Configuration for "sonos-announcement" is incomplete because the following requested capabilities are undefined: x_sonos_announcement Controller: MQTTController#mqtt Last 10:21:37 AMThis is probably a pretty stupid question and the approach may not even work at all, but maybe someone or @toggledbits for sure, could point me to the right direction.
Basically the idea is to be able to send TTS messages from reactions using entity actions. I've previously used HTTP requests to Sonos HTTP API (https://hub.docker.com/r/chrisns/docker-node-sonos-http-api/) for the same functionality, but since moving to sonos-mqtt, I need a way to send the TTS notifications using MQTTController. Along with the actual message, volume and delay must also be parameterizable.
br,
mgvra
MSR latest-25016-d47fea38 / MQTTController [0.2.24293]
Hi, @toggledbits
I just noticed that following a reboot of my raspberry pi, some of the rules, that I was expecting to recover, are not catching up following a reboot. I have made a simple test rule (rule-m6rz6ol1) with only "after Date/time" as trigger and "turn on a lamp" as a set reaction. All my infrastructure is on the same board so Reactor, Hass, Zwavejs, ... are all rebooting.
Here is the sequence of the test case (All time converted to Zulu to match logs):
Rule "after Date/Time" set to 14:05:00z Shutdown on Raspberry Pi at 14:04:00z Power back up at 14:08:00z Rule overview shows true as of 14:08:14z waiting for 00:00:00 in GUIFrom the log I can see that MSR is picking up the rule and knows that the state of the rule has changed from false to true and tries to send the update to HASS but failed with websocket error.
Here is what I see from the log:
14:04:04z shutdown complete 14:08:08z Power up 14:08:13.111z websocket connection 14:08:15:323z Reaction to the light failed, Websocket not opened After there is a series of websocket connection attempt until 14:08:51z where it seemed to be really ready.Back in 2021 we had a discussion (https://smarthome.community/topic/700/solved-start-up?_=1738766986566) and you proposed to add a startup_delay:xxxx and startup_wait:xxxx parameter in the engine section of "reactor.yaml". When I try the startup_delay (this used to be a hard delay), the engine failed to start (I think). I then try the startup_wait:xxxx without any success. Since it wait for the connection status to be up to cancel the delay, it does not do anyting since Hass is reporting the socket up without really being up ( I think...).
Questions:
Did I figured it all wrong? should the startup_delay:xxxxx have worked? Any ideas?Here is the log:
OK now I am stuck. I did add the log but when I submit the editor complained saying that I am limited to 32767 characters. The log from the shutdown to the time the websocket is stable is about 300000 character long. What are my options?
Not a big issue simply a request if easily doable.
The MSR logs files inside the container are owned by root witch is fine however, the permissions are very restrictive. I do not know if there is something wrong with my installation but the logs permission are set to 222 (write only). Even if the docker volume is set for Read/Write the log files are retaining these values.
I go around the problem by doing a chmod 777 on all reactor logs but every time there is an MSR log rotation the permissions are set back to 222. So unless the permission are implemented in the container there is no permanent solution to this (that I know of).
I do not know much about Docker container so I do not know what is involved here.
Can the logfiles permission be simply chaged in the container to at least allow "other" read permission?
Could the MSR log rotation routine implement a chmod to set the permission?
Just a small anoyance
Thanks
@toggledbits In the MSR documentation, under Standard Capabilities, I noticed that the
button.since attribute was deprecated as of version 22256 and the metadata is the preferred way to access the last-modified time of an attribute.
Am I reading this right? Should I stop using it in my rules?
Thanks
When on my bare metal RPi with MSR I had a rule that ran every minute to check Internet status via a script in MSR called reactor_inet_check.sh
I've moved to containerized MSR and see in the instructions that this cannot be run from the container.
The script cannot run within the Reactor docker container. If you are using Reactor in a docker container, the script needs to be run by cron or an equivalent facility on the host system (e.g. some systems, like Synology NAS, have separate task managers that may be used to schedule the repeated execution of tasks such as this).
I've put a script on my container host that calls the reator_inet_check.sh script and it isn't erroring... but I still see the internet status within MSR as null.
Before I go diving down the rabbit hole... should this work?
My cronjob on the proxmox host:
909fe6f0-77fd-4734-80a4-c9e354c910b6-image.png
The contents of msr_internet_check_caller.sh
16337528-cf31-4968-bffe-af1149f7103e-image.png
Background: this is a Windows MSR install I've done for our local pool/amenity center just to run some fans and lights (not my daily driver at home). Install went perfectly fine.
Scenario: I want the lights to go on when it's dark enough (even if during a storm, not just after sunset) so I'm using solarRadiation from my weather station to drive that Trigger. Easy stuff.
Issue: sometimes, someone goes in the office and just starts flipping switches and the result can be lights turned on in the daytime or off at night. I'm trying to create a "catch-all" wherein if it is daytime and the lights somehow find their way ON, they will turn themselves back OFF.
I have the following Reaction built:
b30eab5b-5a14-4a3a-8c9a-47e3e7e53dc3-image.png
I also have this Reaction for opposite, ie the lights find themselves turned off after dark and they will turn themselves back on:
5c6946b1-297c-4eb1-9618-74820979df29-image.png
Here are my two rules:
288cba86-f941-4157-86d9-d8e7487905f7-image.png *NOTE that in my manual testing, ie I turn on the light switch at the incorrect time, when the solarRadiation level changes the Lights ON rule flags and shows as SET. On the next change of solarRadiation it goes back to reset again.
My expectation is that Lights OFF rule should see the lights are on, the solarRadiation is above the set limit, and turn them off. Instead, every other run, the ON rule moves to SET and then reset again on the following run.
Logs appear angry:
[latest-25016]2025-01-26T22:03:31.696Z <Engine:INFO> Enqueueing "Lights ON<RESET>" (rule-m6e4ajh7:R) [latest-25016]2025-01-26T22:03:31.712Z <Engine:NOTICE> Starting reaction Lights ON<RESET> (rule-m6e4ajh7:R) [latest-25016]2025-01-26T22:03:31.713Z <Engine:INFO> Lights ON<RESET> all actions completed. [latest-25016]2025-01-26T22:03:42.565Z <wsapi:INFO> client "127.0.0.1#3" closed, code=1001, reason= [latest-25016]2025-01-26T22:03:42.753Z <httpapi:INFO> [HTTPAPI]#1 API request from ::ffff:127.0.0.1: GET /api/v1/lang [latest-25016]2025-01-26T22:03:42.754Z <httpapi:INFO> [HTTPAPI]#1 request for /api/v1/lang from ::ffff:127.0.0.1 user anonymous auth none matches /api/v1/lang ACL (#7): [Object]{ "url": "/api/v1/lang", "allow": true, "index": 7 } [latest-25016]2025-01-26T22:03:42.790Z <wsapi:INFO> wsapi: connection from ::ffff:127.0.0.1 [latest-25016]2025-01-26T22:03:42.839Z <wsapi:INFO> [WSAPI]wsapi#1 client "127.0.0.1#6" authorized [latest-25016]2025-01-26T22:03:43.353Z <httpapi:INFO> [HTTPAPI]#1 API request from ::ffff:127.0.0.1: GET /api/v1/systime [latest-25016]2025-01-26T22:03:43.353Z <httpapi:INFO> [HTTPAPI]#1 request for /api/v1/systime from ::ffff:127.0.0.1 user anonymous auth none matches /api/v1/systime ACL (#5): [Object]{ "url": "/api/v1/systime", "allow": true, "index": 5 } [latest-25016]2025-01-26T22:03:48.146Z <wsapi:INFO> client "127.0.0.1#6" closed, code=1001, reason= [latest-25016]2025-01-26T22:03:48.308Z <httpapi:INFO> [HTTPAPI]#1 API request from ::ffff:127.0.0.1: GET /api/v1/lang [latest-25016]2025-01-26T22:03:48.309Z <httpapi:INFO> [HTTPAPI]#1 request for /api/v1/lang from ::ffff:127.0.0.1 user anonymous auth none matches /api/v1/lang ACL (#7): [Object]{ "url": "/api/v1/lang", "allow": true, "index": 7 } [latest-25016]2025-01-26T22:03:48.346Z <wsapi:INFO> wsapi: connection from ::ffff:127.0.0.1 [latest-25016]2025-01-26T22:03:48.390Z <wsapi:INFO> [WSAPI]wsapi#1 client "127.0.0.1#7" authorized [latest-25016]2025-01-26T22:03:49.412Z <httpapi:INFO> [HTTPAPI]#1 API request from ::ffff:127.0.0.1: GET /api/v1/systime [latest-25016]2025-01-26T22:03:49.413Z <httpapi:INFO> [HTTPAPI]#1 request for /api/v1/systime from ::ffff:127.0.0.1 user anonymous auth none matches /api/v1/systime ACL (#5): [Object]{ "url": "/api/v1/systime", "allow": true, "index": 5 } [latest-25016]2025-01-26T22:03:52.734Z <wsapi:INFO> client "127.0.0.1#7" closed, code=1001, reason= [latest-25016]2025-01-26T22:03:52.891Z <httpapi:INFO> [HTTPAPI]#1 API request from ::ffff:127.0.0.1: GET /api/v1/lang [latest-25016]2025-01-26T22:03:52.892Z <httpapi:INFO> [HTTPAPI]#1 request for /api/v1/lang from ::ffff:127.0.0.1 user anonymous auth none matches /api/v1/lang ACL (#7): [Object]{ "url": "/api/v1/lang", "allow": true, "index": 7 } [latest-25016]2025-01-26T22:03:52.925Z <wsapi:INFO> wsapi: connection from ::ffff:127.0.0.1 [latest-25016]2025-01-26T22:03:52.965Z <wsapi:INFO> [WSAPI]wsapi#1 client "127.0.0.1#8" authorized [latest-25016]2025-01-26T22:03:54.383Z <httpapi:INFO> [HTTPAPI]#1 API request from ::ffff:127.0.0.1: GET /api/v1/systime [latest-25016]2025-01-26T22:03:54.384Z <httpapi:INFO> [HTTPAPI]#1 request for /api/v1/systime from ::ffff:127.0.0.1 user anonymous auth none matches /api/v1/systime ACL (#5): [Object]{ "url": "/api/v1/systime", "allow": true, "index": 5 } [latest-25016]2025-01-26T22:04:01.590Z <wsapi:INFO> [WSAPI]wsapi#1 rpc_echo [Object]{ "comment": "UI activity" } [latest-25016]2025-01-26T22:04:39.646Z <Rule:INFO> Lights OFF (rule-m6e33ja3 in Atrium Lights) evaluated; rule state transition from RESET to SET! [latest-25016]2025-01-26T22:04:39.656Z <Rule:INFO> Lights ON (rule-m6e4ajh7 in Atrium Lights) evaluated; rule state transition from RESET to SET! [latest-25016]2025-01-26T22:04:39.663Z <Engine:INFO> Enqueueing "Lights OFF<SET>" (rule-m6e33ja3:S) [latest-25016]2025-01-26T22:04:39.665Z <Engine:INFO> Enqueueing "Lights ON<SET>" (rule-m6e4ajh7:S) [latest-25016]2025-01-26T22:04:39.668Z <Engine:NOTICE> Starting reaction Lights OFF<SET> (rule-m6e33ja3:S) [latest-25016]2025-01-26T22:04:39.669Z <Engine:NOTICE> Starting reaction Lights ON<SET> (rule-m6e4ajh7:S) [latest-25016]2025-01-26T22:04:39.669Z <Engine:INFO> Lights ON<SET> all actions completed. [latest-25016]2025-01-26T22:04:39.675Z <Rule:INFO> Lights OFF (rule-m6e33ja3 in Atrium Lights) evaluated; rule state transition from SET to RESET! [latest-25016]2025-01-26T22:04:39.680Z <Engine:NOTICE> ReactionHistory: no entry for [latest-25016]2025-01-26T22:04:39.683Z <Engine:NOTICE> [Engine]Engine#1 entry 256 reaction rule-m6e33ja3:S-1q2f1j0p: [Error] terminated [parent terminating] [latest-25016]2025-01-26T22:04:39.683Z <Engine:CRIT> Error: terminated [parent terminating] Error: terminated at Engine._process_reaction_queue (C:\Users\Jalan\msr\reactor\server\lib\Engine.js:1644:47) [latest-25016]2025-01-26T22:04:39.699Z <Engine:NOTICE> [Engine]Engine#1 entry 254 reaction rule-m6e33ja3:S: [Error] terminated [preempted by rule state change] [latest-25016]2025-01-26T22:04:39.699Z <Engine:CRIT> Error: terminated [preempted by rule state change] Error: terminated at Engine._process_reaction_queue (C:\Users\Jalan\msr\reactor\server\lib\Engine.js:1644:47) [latest-25016]2025-01-26T22:04:39.700Z <Engine:INFO> Enqueueing "Lights OFF<RESET>" (rule-m6e33ja3:R) [latest-25016]2025-01-26T22:04:39.704Z <Engine:NOTICE> Starting reaction Lights OFF<RESET> (rule-m6e33ja3:R) [latest-25016]2025-01-26T22:04:39.705Z <Engine:INFO> Lights OFF<RESET> all actions completed. [latest-25016]2025-01-26T22:05:48.822Z <Rule:INFO> Lights ON (rule-m6e4ajh7 in Atrium Lights) evaluated; rule state transition from SET to RESET! [latest-25016]2025-01-26T22:05:48.831Z <Engine:INFO> Enqueueing "Lights ON<RESET>" (rule-m6e4ajh7:R) [latest-25016]2025-01-26T22:05:48.847Z <Engine:NOTICE> Starting reaction Lights ON<RESET> (rule-m6e4ajh7:R) [latest-25016]2025-01-26T22:05:48.847Z <Engine:INFO> Lights ON<RESET> all actions completed.Hi @toggledbits
I found this very old post that talked about a way to limit device reading to avoid the throttled problem, because it's not a question of logic, it's that the device actually sends a lot of information, in my case the NUT ups installed in HE.
https://smarthome.community/topic/687/flapping-device?_=1737652139854
It mentions engine section of reactor.yaml by setting update_rate_limit, but I looked in the current MSR documentation and I can't find this information, so I don't know if it's still valid, its effect and parameters.
My situation is simple, when I have a UPS problem the NUT is sending dozens of reports per second and then I have the throttled problem. The same rule applies when the power is normal.
This is the rule, and the parameter that fails is the Tripp Lite UPS status.
cf9ddabf-3144-4e5a-80a4-0dc7664b9573-image.png
a813a077-974e-4737-897c-e383085b3d8f-image.png
All error is the same scenario.
[latest-25016]2025-01-23T12:01:32.753Z <Rule:WARN> (13) NUT Disconected (rule-l4djr0p7 in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Throttl> [latest-25016]2025-01-23T12:01:32.756Z <Rule:WARN> (27) Falta de Energia (rule-l4h9ceod in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Thrott> [latest-25016]2025-01-23T12:01:32.769Z <Rule:WARN> (73) UPS Battery Low (rule-l4hj850o in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Throttl> [latest-25016]2025-01-23T12:01:32.772Z <Rule:WARN> (74) UPS Comm Fail (rule-l4kbs5cp in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Throttlin> [latest-25016]2025-01-23T12:01:32.776Z <Rule:WARN> (76) UPS Utility Back (rule-l4hjhs6m in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Thrott> [latest-25016]2025-01-23T12:01:32.780Z <Rule:WARN> UPS On Battery (rule-l4hjuka5 in Datacenter) update rate 121/min exceeds limit (120/min)! Logic loop? Throttling> [latest-25016]2025-01-23T12:01:32.781Z <Rule:WARN> UPS Info (rule-l4gheo63 in Datacenter) update rate 121/min exceeds limit (120/min)! Logic loop? Throttling... [latest-25016]2025-01-23T12:01:40.757Z <Rule:WARN> (13) NUT Disconected (rule-l4djr0p7 in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Throttl> [latest-25016]2025-01-23T12:01:40.759Z <Rule:WARN> (27) Falta de Energia (rule-l4h9ceod in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Thrott> [latest-25016]2025-01-23T12:01:40.776Z <Rule:WARN> (73) UPS Battery Low (rule-l4hj850o in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Throttl> [latest-25016]2025-01-23T12:01:40.777Z <Rule:WARN> (74) UPS Comm Fail (rule-l4kbs5cp in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Throttlin> [latest-25016]2025-01-23T12:01:40.778Z <Rule:WARN> (76) UPS Utility Back (rule-l4hjhs6m in Warning) update rate 121/min exceeds limit (120/min)! Logic loop? Thrott>Thanks.
Hello -
Long time. Hope everyone is good.
I have a rule that looks at a number of temperature sensors around the house. It simply sends a general alert if any of them fall below their threshold. (A basic “House is too cold” alert for when we’re away)
Generally, this has worked well. But I was wondering if there’s a way to make the message somewhat dynamic without creating separate rules for each sensor.
E.g. “House is too cold due to Sump Temperature below 45 degrees.”
I thought I remember reading about someone doing this in the past but couldn’t find it.
Thanks for any ideas!
After upgrade from 24366 to 25016 toggle on Vera (openLuup) lock entity doesn't seem to work.
VeraController log level 5
24366:
25016:
[latest-25016]2025-01-20T20:30:25.457Z <VeraController:INFO> VeraController#vera perform action toggle.toggle on Lock#vera>device_10016 with [Object]{ } [latest-25016]2025-01-20T20:30:25.458Z <VeraController:5:VeraController.js:1210> VeraController#vera no configured implementation for toggle.toggle on Lock#vera>device_10016; calling super [latest-25016]2025-01-20T20:30:25.458Z <VeraController:INFO> VeraController#vera perform action lock.set on Lock#vera>device_10016 with [Object]{ "state": false } [latest-25016]2025-01-20T20:30:25.461Z <VeraController:5:VeraController.js:1255> VeraController#vera sending action request [Object]{ "newTargetValue": "0", "DeviceNum": 10016, "id": "action", "serviceId": "urn:upnp-org:serviceId:DoorLock1", "action": "SetTarget" } [latest-25016]2025-01-20T20:30:25.461Z <VeraController:5:VeraController.js:515> [VeraController:verarequest] requesting http://127.0.0.1:3480/data_request?newTargetValue=0&DeviceNum=10016&id=action&serviceId=urn%3Aupnp-org%3AserviceId%3ADoorLock1&action=SetTarget&output_format=json&_r=1737405025461 ([Object]{ }) [latest-25016]2025-01-20T20:30:25.463Z <VeraController:NOTICE> VeraController#vera action lock.set([Object]{ "state": false }) on Lock#vera>device_10016 succeededI tried to log an issue in Mantis Bug Tracker but I can't log in anymore?
Trying to understand what cause a local expresssion to be evaluated. I have read the manual but I am still not clear about it. Using the test rule below, I can see in the log that the rule is being automatically evaluated every time the temperature entity is changing. That is great...
What I am trying to understand is why the expression is not evaluated based on time as well since the "case" statement has time dependencies.
Any help would be appreciated
I have the following test rule:
eba6a3ea-ff61-4610-88c9-9b9864f11ff8-Screenshot 2025-01-21 095244.png
2d9c1ff5-7b73-4005-b324-9029c2709db9-Screenshot 2025-01-21 095302.png
Here is the expressioncode:
vFrom1 = "09:25:00", vFrom2 = "09:30:00", vFrom3 = "09:41:00", vTo = "10:55:00", # Get current time (format HH:MM:SS) vToDay = strftime("%H:%M:%S"), #Get current house temperature CurrentHouseTemp = getEntity( "hass>Thermostat2 " ).attributes.temperature_sensor.value, case when CurrentHouseTemp <= 19 and vToDay >= vFrom1 && vToDay <= vTo: "true1" # From1 when CurrentHouseTemp <= 20 and vToDay >= vFrom2 && vToDay <= vTo: "true2" # From2 when CurrentHouseTemp < 26 and vToDay >= vFrom3 && vToDay <= vTo: "true3" # From3 else "false" endAnd then MSR...
My first issue: I'm logged into the msr CT as reactor (I used the suggested username just to keep things simple as this is new space for me and I was high off my success of migrating HA over).
When I run
docker pull toggledbits/reactor:latest-amd64... it assigns the \reactor\ subdirectory where installed root ownership. I am absolutely logged in with the correct non-root user.
1c58aead-85ca-4b2c-8f48-c3d1f57d7fe3-image.png
Second issue: I copied over the following folders:
67e7e4a5-cee8-4de1-90c7-1df35f1070b9-image.png
When MSR loads, all of my Global Expressions are missing.
Third issue:
All controllers connect wonderfully (Hubitat, etc)... except HA.
After changing ownership of the logs to reactor again I can see this when MSR calls HA:
Yes, I created a fresh new long-lived access token for the MSR containerized install and updated the reactor.yaml config file correctly.
Honestly, all-in-all, for my total lack of expertise here I'm very pleased that I only have these three issues. But they are def blockers atm.
My RPi bare metal install of MSR hooked right up to the new HA and is humming along just fine (I used hostnames were possible and shuffled some IPs in other places so I wouldn't run into things later that were mapped incorrectly that I'd forgotten about.)
Proxmox 8.3.2 MSR lives in an Ubuntu 24.04 Proxmox container MSR is latest docker versionWhat else can I provide to those smarter than me here?
Reactor (Multi-hub) latest-24366-3de60836
Running on Proxmox 8 VM
Ubuntu 22.04.5 LTS
Docker version 27.5.0, build a187fa5
Docker Compose version v2.32.3
Browsers being used on Mac OS Sequoia: Safari, Firefox also occurs with Safari on iPhone 16 Pro 18.2.1
This occurs on two different instances of MSR running at two different locations having the same environment detailed above.
When I select "Reactions->Create Reaction" I get an error window with a red “Runtime Error:” banner. Note that I can edit and save existing Reactions
—-------------------<SNIP>————————————
Runtime Error:
@http://192.168.119.137:8111/reactor/en-US/lib/js/reactor-ui-reactions.js:445:34
You may report this error, but do not screen shot it. Copy-paste the complete text. Remember to include a description of the operation you were performing in as much detail as possible. Report using the Reactor Bug Tracker (in your left navigation) or at the SmartHome Community.
---------------------</SNIP>————————
apt update, apt upgrade, reboot have been performed as well as
docker system prune -a
docker compose down
docker compose up -d
Many thanks in advance,
-bh
Build 21228 has been released. Docker images available from DockerHub as usual, and bare-metal packages here.
Home Assistant up to version 2021.8.6 supported; the online version of the manual will now state the current supported versions; Fix an error in OWMWeatherController that could cause it to stop updating; Unify the approach to entity filtering on all hub interface classes (controllers); this works for device entities only; it may be extended to other entities later; Improve error detail in messages for EzloController during auth phase; Add isRuleSet() and isRuleEnabled() functions to expressions extensions; Implement set action for lock and passage capabilities (makes them more easily scriptable in some cases); Fix a place in the UI where 24-hour time was not being displayed.I may have posted this in the wrong section. MSR running on Bare metal Debian bullseye. Both Openluup and MSR are on the same device (an Intel NUC) at IP 192.168.70.249. Any suggestions as to where I go to resolve?
TIA
Happy new year, everyone! Hope all are well!
Looking for some pointers troubleshooting a slightly puzzling to me issue. When digging around on a different issue I noticed this happening regularly in the MSR logs:
[latest-24366]2025-01-10T19:50:07.630Z <Engine:NOTICE> Starting reaction Garden lights on when the doors are open<SET> (rule-lb2h69nb:S) [latest-24366]2025-01-10T19:50:07.630Z <VeraController:INFO> VeraController#vera perform action power_switch.on on Switch#vera>device_20060 with [Object]{ } [latest-24366]2025-01-10T19:50:07.630Z <VeraController:INFO> VeraController#vera perform action power_switch.set on Switch#vera>device_20060 with [Object]{ "state": true } [latest-24366]2025-01-10T19:50:07.670Z <VeraController:NOTICE> VeraController#vera action power_switch.set([Object]{ "state": true }) on Switch#vera>device_20060 succeeded [latest-24366]2025-01-10T19:50:07.671Z <Engine:INFO> Resuming reaction Garden lights on when the doors are open<SET> (rule-lb2h69nb:S) from step 1 [latest-24366]2025-01-10T19:50:07.672Z <Engine:NOTICE> Garden lights on when the doors are open<SET> delaying until 1736538787672<10/01/2025, 19:53:07> [latest-24366]2025-01-10T19:50:19.595Z <Rule:INFO> Garden lights on when the doors are open (rule-lb2h69nb in Outside Lights) evaluated; rule state transition from SET to RESET! [latest-24366]2025-01-10T19:52:16.506Z <Rule:INFO> Garden lights on when the doors are open (rule-lb2h69nb in Outside Lights) evaluated; rule state transition from RESET to SET! [latest-24366]2025-01-10T19:52:16.515Z <Engine:INFO> [Engine]Engine#1 not enqueueing rule-lb2h69nb:S: already in queue with status 2 [latest-24366]2025-01-10T19:52:20.823Z <Rule:INFO> Garden lights on when the doors are open (rule-lb2h69nb in Outside Lights) evaluated; rule state transition from SET to RESET! [latest-24366]2025-01-10T19:53:07.676Z <Engine:INFO> Resuming reaction Garden lights on when the doors are open<SET> (rule-lb2h69nb:S) from step 2 [latest-24366]2025-01-10T19:53:07.677Z <VeraController:INFO> VeraController#vera perform action power_switch.off on Switch#vera>device_20060 with [Object]{ } [latest-24366]2025-01-10T19:53:07.678Z <VeraController:INFO> VeraController#vera perform action power_switch.set on Switch#vera>device_20060 with [Object]{ "state": false } [latest-24366]2025-01-10T19:53:07.719Z <VeraController:NOTICE> VeraController#vera action power_switch.set([Object]{ "state": false }) on Switch#vera>device_20060 succeeded [latest-24366]2025-01-10T19:53:07.720Z <Engine:INFO> Resuming reaction Garden lights on when the doors are open<SET> (rule-lb2h69nb:S) from step 3 [latest-24366]2025-01-10T19:53:07.721Z <Engine:INFO> Garden lights on when the doors are open<SET> all actions completed. [latest-24366]2025-01-10T19:55:04.468Z <VeraController:ERR> VeraController#vera update request failed: [FetchError] network timeout at: http://192.168.70.249:3480/data_request?id=status&Timeout=15&DataVersion=416912953&MinimumDelay=50&output_format=json&_r=1736538886459 [-] [latest-24366]2025-01-10T19:55:09.646Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20050: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20050) [-] [latest-24366]2025-01-10T19:55:09.646Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.646Z <VeraController:CRIT> *Entity#vera>device_20050 [latest-24366]2025-01-10T19:55:09.656Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20570: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20570) [-] [latest-24366]2025-01-10T19:55:09.656Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.656Z <VeraController:CRIT> *Entity#vera>device_20570 [latest-24366]2025-01-10T19:55:09.678Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20610: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20610) [-] [latest-24366]2025-01-10T19:55:09.679Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.679Z <VeraController:CRIT> *Entity#vera>device_20610 [latest-24366]2025-01-10T19:55:09.744Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20631: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20631) [-] [latest-24366]2025-01-10T19:55:09.744Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.744Z <VeraController:CRIT> *Entity#vera>device_20631 [latest-24366]2025-01-10T19:55:09.889Z <VeraController:NOTICE> VeraController#vera reload detected! [latest-24366]2025-01-10T19:55:09.910Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20050: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20050) [-] [latest-24366]2025-01-10T19:55:09.910Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.910Z <VeraController:CRIT> *Entity#vera>device_20050 [latest-24366]2025-01-10T19:55:09.935Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20570: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20570) [-] [latest-24366]2025-01-10T19:55:09.936Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.936Z <VeraController:CRIT> *Entity#vera>device_20570 [latest-24366]2025-01-10T19:55:09.937Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20610: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20610) [-] [latest-24366]2025-01-10T19:55:09.937Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.937Z <VeraController:CRIT> *Entity#vera>device_20610 [latest-24366]2025-01-10T19:55:09.939Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20631: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20631) [-] [latest-24366]2025-01-10T19:55:09.939Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.939Z <VeraController:CRIT> *Entity#vera>device_20631 [latest-24366]2025-01-10T19:55:09.968Z <Controller:INFO> VeraController#vera 0 dead entities older than 86400000s purged [latest-24366]2025-01-10T19:55:10.037Z <VeraController:NOTICE> VeraController#vera reload detected!That repeats until something like this:
[latest-24366]2025-01-10T19:55:10.049Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20050: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20050) [-] [latest-24366]2025-01-10T19:55:10.049Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:10.049Z <VeraController:CRIT> *Entity#vera>device_20050 [latest-24366]2025-01-10T19:55:10.053Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20570: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20570) [-] [latest-24366]2025-01-10T19:55:10.053Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:10.053Z <VeraController:CRIT> *Entity#vera>device_20570 [latest-24366]2025-01-10T19:55:10.062Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20610: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20610) [-] [latest-24366]2025-01-10T19:55:10.062Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:10.062Z <VeraController:CRIT> *Entity#vera>device_20610 [latest-24366]2025-01-10T19:55:10.112Z <VeraController:WARN> VeraController#vera failed to apply attribute scene_activation.scene_id to Entity#vera>device_20631: [TypeError] Can't set NaN on attribute scene_activation.scene_id (vera>device_20631) [-] [latest-24366]2025-01-10T19:55:10.112Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:10.113Z <VeraController:CRIT> *Entity#vera>device_20631 [latest-24366]2025-01-10T20:00:05.003Z <Engine:INFO> [Engine]Engine#1 master timer tick, local time "10/01/2025 20:00:05" (TZ offset 0 mins from UTC) [latest-24366]2025-01-10T20:13:51.872Z <Rule:INFO> No motion in Cinema (rule-m4ocglke in Cinema Environment) evaluated; rule state transition from SET to RESET! [latest-24366]2025-01-10T20:13:51.882Z <Rule:INFO> Cinema Heater On (rule-m4ocf1di in Cinema Environment) evaluated; rule state transition from RESET to SET! [latest-24366]2025-01-10T20:13:51.888Z <Engine:INFO> Enqueueing "Cinema Heater On<SET>" (rule-m4ocf1di:S)And the errors / reloads just stop.
From Openluup:
2025-01-10 19:49:56.379 luup_log:63: BroadLink_Mk2 debug: RM3 Mini - IR 1: urn:schemas-micasaverde-com:device:IrTransmitter:1 2025-01-10 19:50:00.085 luup_log:0: 14Mb, 1.7%cpu, 36.1days 2025-01-10 19:50:07.591 luup.variable_set:: 20160.urn:micasaverde-com:serviceId:EnergyMetering1.KWH was: 18.6793008 now: 18.6805008 #hooks:0 2025-01-10 19:50:07.591 luup.variable_set:: 20160.urn:micasaverde-com:serviceId:EnergyMetering1.KWHReading was: 1736538000 now: 1736538600 #hooks:0 2025-01-10 19:50:07.591 luup.variable_set:: 20160.urn:micasaverde-com:serviceId:EnergyMetering1.Watts was: 7.4 now: 7.3 #hooks:0 2025-01-10 19:50:07.591 luup.variable_set:: 20170.urn:micasaverde-com:serviceId:EnergyMetering1.KWH was: 32.2417984 now: 32.2470016 #hooks:0 2025-01-10 19:50:07.591 luup.variable_set:: 20170.urn:micasaverde-com:serviceId:EnergyMetering1.KWHReading was: 1736538000 now: 1736538600 #hooks:0 2025-01-10 19:50:07.591 luup.variable_set:: 20330.urn:micasaverde-com:serviceId:EnergyMetering1.KWHReading was: 1736538000 now: 1736538600 #hooks:0 2025-01-10 19:50:07.592 luup.variable_set:: 20770.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped was: 0 now: 1 #hooks:0 2025-01-10 19:50:07.592 luup.variable_set:: 20770.urn:micasaverde-com:serviceId:SecuritySensor1.LastTrip was: 1736534850 now: 1736538607 #hooks:0 2025-01-10 19:50:07.593 openLuup.server:: request completed (3392 bytes, 1 chunks, 12875 ms) tcp{client}: 0x55c3299a9cf8 2025-01-10 19:50:07.618 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c3299a9cf8 2025-01-10 19:50:07.624 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c329d0a5b8 2025-01-10 19:50:07.624 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912906&MinimumDelay=50&output_format=json&_r=1736538607623 HTTP/1.1 tcp{client}: 0x55c329d0a5b8 2025-01-10 19:50:07.632 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c3292ed678 2025-01-10 19:50:07.633 openLuup.server:: GET /data_request?newTargetValue=1&DeviceNum=20060&id=action&serviceId=urn%3Aupnp-org%3AserviceId%3ASwitchPower1&action=SetTarget&output_format=json&_r=1736538607631 HTTP/1.1 tcp{client}: 0x55c3 292ed678 2025-01-10 19:50:07.633 luup.call_action:: 20060.urn:upnp-org:serviceId:SwitchPower1.SetTarget 2025-01-10 19:50:07.633 luup.call_action:: action will be handled by parent: 37 2025-01-10 19:50:07.633 luup.variable_set:: 20060.urn:upnp-org:serviceId:SwitchPower1.Target was: 0 now: 1 #hooks:0 2025-01-10 19:50:07.669 openLuup.server:: request completed (35 bytes, 1 chunks, 35 ms) tcp{client}: 0x55c3292ed678 2025-01-10 19:50:07.673 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c3292ed678 2025-01-10 19:50:07.776 openLuup.server:: request completed (821 bytes, 1 chunks, 151 ms) tcp{client}: 0x55c329d0a5b8 2025-01-10 19:50:07.784 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c329d0a5b8 2025-01-10 19:50:07.795 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c3287bc8f8 2025-01-10 19:50:07.796 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912907&MinimumDelay=50&output_format=json&_r=1736538607794 HTTP/1.1 tcp{client}: 0x55c3287bc8f8 2025-01-10 19:50:08.644 luup.variable_set:: 20060.urn:upnp-org:serviceId:SwitchPower1.Status was: 0 now: 1 #hooks:0 2025-01-10 19:50:08.950 openLuup.server:: request completed (821 bytes, 1 chunks, 1154 ms) tcp{client}: 0x55c3287bc8f8 2025-01-10 19:50:08.958 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c3287bc8f8 2025-01-10 19:50:08.969 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c3297e95a8 2025-01-10 19:50:08.970 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912908&MinimumDelay=50&output_format=json&_r=1736538608969 HTTP/1.1 tcp{client}: 0x55c3297e95a8 2025-01-10 19:50:19.181 luup.variable_set:: 20770.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped was: 1 now: 0 #hooks:0 2025-01-10 19:50:19.585 openLuup.server:: request completed (832 bytes, 1 chunks, 10615 ms) tcp{client}: 0x55c3297e95a8 2025-01-10 19:50:19.602 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c3297e95a8 2025-01-10 19:50:19.605 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c328d298a8 2025-01-10 19:50:19.605 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912909&MinimumDelay=50&output_format=json&_r=1736538619604 HTTP/1.1 tcp{client}: 0x55c328d298a8 2025-01-10 19:50:34.950 openLuup.server:: request completed (593 bytes, 1 chunks, 15344 ms) tcp{client}: 0x55c328d298a8 2025-01-10 19:50:34.953 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c328d298a8 2025-01-10 19:50:34.965 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c328c48a58 2025-01-10 19:50:34.966 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912909&MinimumDelay=50&output_format=json&_r=1736538634964 HTTP/1.1 tcp{client}: 0x55c328c48a58 2025-01-10 19:50:34.989 luup.variable_set:: 25019.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped was: 0 now: 1 #hooks:0 2025-01-10 19:50:34.990 luup.variable_set:: 25019.urn:micasaverde-com:serviceId:SecuritySensor1.LastTrip was: 1736534437 now: 1736538634 #hooks:0 2025-01-10 19:50:35.094 openLuup.server:: request completed (975 bytes, 1 chunks, 127 ms) tcp{client}: 0x55c328c48a58 2025-01-10 19:50:35.101 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c328c48a58 2025-01-10 19:50:35.113 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c32985e298 2025-01-10 19:50:35.113 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912911&MinimumDelay=50&output_format=json&_r=1736538635111 HTTP/1.1 tcp{client}: 0x55c32985e298 2025-01-10 19:50:40.255 luup.variable_set:: 25021.urn:micasaverde-com:serviceId:LightSensor1.CurrentLevel was: 0 now: 30 #hooks:1 2025-01-10 19:50:40.256 scheduler.watch_callback:: 25021.urn:micasaverde-com:serviceId:LightSensor1.CurrentLevel called [20]DataWatcherCallback() function: 0x55c3288a8d20 2025-01-10 19:50:40.460 openLuup.server:: request completed (835 bytes, 1 chunks, 5346 ms) tcp{client}: 0x55c32985e298 2025-01-10 19:50:40.472 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c32985e298 2025-01-10 19:50:40.478 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c329b28238 2025-01-10 19:50:40.479 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912912&MinimumDelay=50&output_format=json&_r=1736538640478 HTTP/1.1 tcp{client}: 0x55c329b28238 2025-01-10 19:50:44.471 luup.variable_set:: 25007.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped was: 0 now: 1 #hooks:1 2025-01-10 19:50:44.472 luup.variable_set:: 25007.urn:micasaverde-com:serviceId:SecuritySensor1.LastTrip was: 1736538400 now: 1736538644 #hooks:0 2025-01-10 19:50:44.472 scheduler.watch_callback:: 25007.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped called [20]DataWatcherCallback() function: 0x55c3288a8d20 2025-01-10 19:50:44.775 openLuup.server:: request completed (975 bytes, 1 chunks, 4296 ms) tcp{client}: 0x55c329b28238 2025-01-10 19:50:44.775 openLuup.server:: request completed (975 bytes, 1 chunks, 4296 ms) tcp{client}: 0x55c329b28238 2025-01-10 19:50:44.782 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c329b28238 2025-01-10 19:50:44.793 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c328f1e968 2025-01-10 19:50:44.793 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912914&MinimumDelay=50&output_format=json&_r=1736538644791 HTTP/1.1 tcp{client}: 0x55c328f1e968 2025-01-10 19:51:00.122 openLuup.server:: request completed (593 bytes, 1 chunks, 15328 ms) tcp{client}: 0x55c328f1e968 2025-01-10 19:51:00.125 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c328f1e968 2025-01-10 19:51:00.136 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c32995b318 2025-01-10 19:51:00.136 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912914&MinimumDelay=50&output_format=json&_r=1736538660134 HTTP/1.1 tcp{client}: 0x55c32995b318 2025-01-10 19:51:15.481 openLuup.server:: request completed (593 bytes, 1 chunks, 15344 ms) tcp{client}: 0x55c32995b318 2025-01-10 19:51:15.484 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c32995b318 2025-01-10 19:51:15.495 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c32998b068 2025-01-10 19:51:15.497 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912914&MinimumDelay=50&output_format=json&_r=1736538675493 HTTP/1.1 tcp{client}: 0x55c32998b068 2025-01-10 19:51:30.869 openLuup.server:: request completed (593 bytes, 1 chunks, 15371 ms) tcp{client}: 0x55c32998b068 2025-01-10 19:51:30.872 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c32998b068 2025-01-10 19:51:30.884 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c32905bda8 2025-01-10 19:51:30.885 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912914&MinimumDelay=50&output_format=json&_r=1736538690882 HTTP/1.1 tcp{client}: 0x55c32905bda8 2025-01-10 19:51:32.886 luup.variable_set:: 20380.urn:upnp-org:serviceId:TemperatureSensor1.CurrentTemperature was: 21 now: 22 #hooks:0 2025-01-10 19:51:33.090 openLuup.server:: request completed (841 bytes, 1 chunks, 2205 ms) tcp{client}: 0x55c32905bda8 2025-01-10 19:51:33.100 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c32905bda8 2025-01-10 19:51:33.112 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c328de0d58 2025-01-10 19:51:33.112 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912915&MinimumDelay=50&output_format=json&_r=1736538693111 HTTP/1.1 tcp{client}: 0x55c328de0d58 2025-01-10 19:51:36.064 luup.variable_set:: 25007.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped was: 1 now: 0 #hooks:1 2025-01-10 19:51:36.065 scheduler.watch_callback:: 25007.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped called [20]DataWatcherCallback() function: 0x55c3288a8d20 2025-01-10 19:51:36.369 openLuup.server:: request completed (832 bytes, 1 chunks, 3256 ms) tcp{client}: 0x55c328de0d58 2025-01-10 19:51:36.377 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c328de0d58 2025-01-10 19:51:36.387 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c329054188 2025-01-10 19:51:36.388 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912916&MinimumDelay=50&output_format=json&_r=1736538696386 HTTP/1.1 tcp{client}: 0x55c329054188 2025-01-10 19:51:37.134 luup.variable_set:: 20380.urn:upnp-org:serviceId:TemperatureSensor1.CurrentTemperature was: 22 now: 21 #hooks:0 2025-01-10 19:51:37.540 openLuup.server:: request completed (841 bytes, 1 chunks, 1152 ms) tcp{client}: 0x55c329054188 2025-01-10 19:51:37.553 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c329054188 2025-01-10 19:51:37.566 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c328d97568 2025-01-10 19:51:37.566 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912917&MinimumDelay=50&output_format=json&_r=1736538697564 HTTP/1.1 tcp{client}: 0x55c328d97568 2025-01-10 19:51:41.367 luup.variable_set:: 20380.urn:upnp-org:serviceId:TemperatureSensor1.CurrentTemperature was: 21 now: 22 #hooks:0 2025-01-10 19:51:41.874 openLuup.server:: request completed (841 bytes, 1 chunks, 4307 ms) tcp{client}: 0x55c328d97568 2025-01-10 19:51:41.884 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c328d97568 2025-01-10 19:51:41.895 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c329385678 2025-01-10 19:51:41.896 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912918&MinimumDelay=50&output_format=json&_r=1736538701894 HTTP/1.1 tcp{client}: 0x55c329385678 2025-01-10 19:51:57.168 openLuup.server:: request completed (593 bytes, 1 chunks, 15272 ms) tcp{client}: 0x55c329385678 2025-01-10 19:51:57.171 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c329385678 2025-01-10 19:51:57.183 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c329b092b8 2025-01-10 19:51:57.184 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912918&MinimumDelay=50&output_format=json&_r=1736538717182 HTTP/1.1 tcp{client}: 0x55c329b092b8 2025-01-10 19:52:00.124 luup_log:0: 14Mb, 1.6%cpu, 36.1days 2025-01-10 19:52:00.476 openLuup.server:: request completed (1841 bytes, 1 chunks, 3292 ms) tcp{client}: 0x55c329b092b8 2025-01-10 19:52:00.483 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c329b092b8 2025-01-10 19:52:00.495 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c3297be088 2025-01-10 19:52:00.495 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912929&MinimumDelay=50&output_format=json&_r=1736538720494 HTTP/1.1 tcp{client}: 0x55c3297be088 2025-01-10 19:52:09.867 luup.variable_set:: 25021.urn:micasaverde-com:serviceId:LightSensor1.CurrentLevel was: 30 now: 0 #hooks:1 2025-01-10 19:52:09.868 scheduler.watch_callback:: 25021.urn:micasaverde-com:serviceId:LightSensor1.CurrentLevel called [20]DataWatcherCallback() function: 0x55c3288a8d20 2025-01-10 19:52:10.071 openLuup.server:: request completed (834 bytes, 1 chunks, 9575 ms) tcp{client}: 0x55c3297be088 2025-01-10 19:52:10.079 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c3297be088 2025-01-10 19:52:10.088 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c329c16a08 2025-01-10 19:52:10.089 openLuup.server:: GET /data_request?id=status&Timeout=15&DataVersion=416912930&MinimumDelay=50&output_format=json&_r=1736538730087 HTTP/1.1 tcp{client}: 0x55c329c16a08 2025-01-10 19:52:16.194 luup.variable_set:: 20770.urn:micasaverde-com:serviceId:SecuritySensor1.Tripped was: 0 now: 1 #hooks:0 2025-01-10 19:52:16.195 luup.variable_set:: 20770.urn:micasaverde-com:serviceId:SecuritySensor1.LastTrip was: 1736538607 now: 1736538736 #hooks:0 2025-01-10 19:52:16.498 openLuup.server:: request completed (976 bytes, 1 chunks, 6409 ms) tcp{client}: 0x55c329c16a08 2025-01-10 19:52:16.515 openLuup.io.server:: HTTP:3480 connection closed openLuup.server.receive closed tcp{client}: 0x55c329c16a08 2025-01-10 19:52:16.516 openLuup.io.server:: HTTP:3480 connection from 192.168.70.249 tcp{client}: 0x55c328dbad18Nothing I can see indicating that Openluup is reloading?
.249 IP address is the internal IP of the NUC that hosts both Openluup and MSR.
Any thoughts as to how I can troubleshoot this? It's not a big deal, but would like to get to the bottom of it.
I should add that all the devices listed in entries like this:
[latest-24366]2025-01-10T19:55:09.744Z <VeraController:INFO> VeraController#vera class scene_controller meta [Object]{ "source": "urn:micasaverde-com:serviceId:SceneController1/sl_SceneActivated", "expr": "int(value)" } orig final NaN [latest-24366]2025-01-10T19:55:09.744Z <VeraController:CRIT> *Entity#vera>device_20631Are the tamper switches on Fibaro FGMS001 multifunction detectors, of which I have 4, and they correspond exactly to the devices listed.
TIA
C
Hi
I was looking at an old rule and I wanted to edit it, to add another Constraint, however I cannot seem to do it.
On this screen shot you can see an existing entry in the Constraints and on its pull down menu the "Changes" option is available.
09735de3-8e92-4e12-bfa2-5191f48924a7-image.png
However on the new line I just added I have no changes option in its pull down menu.
d85c4067-880e-4281-a12b-dac4d316a4da-image.png
Here is the original now locked post about this topic.
https://smarthome.community/topic/395/contact-sensor-opened-1-minute-ago-how?_=1736354690742
If you look on the old screen shots on that post, I was using the "changes" operator. Like this:
a1262e01-d3fd-4723-872f-872f1f6d9899-image.png
However today when I edited this rule the operators are showing as == and not as changes on all the entries in the Constraints area.
Also the old entries now say -- and the value is blank. But on the new line I just added it says that is not valid, so not sure how the old lines are like that.
a458d52d-214d-4862-a2b7-d31009f89cde-image.png
So I am a bit confused what happened.
Thanks
@toggledbits I understand that you do not perform testing on Mac computers but thought I'd share the following with you in case something can be done.
I started seeing these errors with version 24302. I thought that upgrading to 24343 would have fixed the issue but unfortunately not. I either have to close the browser or clear the cache for the errors to stop popping-up but they slowly come back.
I see these errors on the following browsers:
Safari 16.6.1 on macOS Big Sur Safari 18.1.1 on MacOS Sonoma DuckDuckGo 1.118.0 on macOS Big Sur and Sonoma Firefox 133.0.3 on macOS Big Sur Chrome 131.0.6778 on macOS Big SurHere are the errors
Safari while creating/updating an expression
@http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:543:91 makeExprMenu@http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:537:28 @http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:92:64 @http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:89:68 each@http://192.168.0.13:8111/node_modules/jquery/dist/jquery.min.js:2:3133 @http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:89:35 @http://192.168.0.13:8111/client/MessageBus.js:98:44 forEach@[native code] @http://192.168.0.13:8111/client/MessageBus.js:95:54 @http://192.168.0.13:8111/client/MessageBus.js:106:44 @http://192.168.0.13:8111/client/Observable.js:78:28 signalModified@http://192.168.0.13:8111/reactor/en-ca/lib/js/ee.js:146:21 signalModified@http://192.168.0.13:8111/reactor/en-ca/lib/js/expression-editor.js:40:29 reindexExpressions@http://192.168.0.13:8111/reactor/en-ca/lib/js/expression-editor.js:71:32 @http://192.168.0.13:8111/reactor/en-ca/lib/js/expression-editor.js:608:40 dispatch@http://192.168.0.13:8111/node_modules/jquery/dist/jquery.min.js:2:40040DuckDuckGo while clicking on status
http://192.168.0.13:8111/reactor/en-ca/lib/js/reactor-ui-status.js:789:44 asyncFunctionResume@[native code] saveGridLayout@[native code] dispatchEvent@[native code] _triggerEvent@http://192.168.0.13:8111/node_modules/gridstack/dist/gridstack.js:1401:30 _triggerAddEvent@http://192.168.0.13:8111/node_modules/gridstack/dist/gridstack.js:1383:31 makeWidget@http://192.168.0.13:8111/node_modules/gridstack/dist/gridstack.js:968:30 addWidget@http://192.168.0.13:8111/node_modules/gridstack/dist/gridstack.js:388:24 placeWidgetAdder@http://192.168.0.13:8111/reactor/en-ca/lib/js/reactor-ui-status.js:183:44Firefox while updating a rule
@http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:543:91 makeExprMenu@http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:537:28 @http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:92:64 @http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:89:68 each@http://192.168.0.13:8111/node_modules/jquery/dist/jquery.min.js:2:3133 @http://192.168.0.13:8111/reactor/en-ca/lib/js/reaction-editor.js:89:35 @http://192.168.0.13:8111/client/MessageBus.js:98:44 forEach@[native code] @http://192.168.0.13:8111/client/MessageBus.js:95:54 @http://192.168.0.13:8111/client/MessageBus.js:106:44 @http://192.168.0.13:8111/client/Observable.js:78:28 notifySaved@http://192.168.0.13:8111/reactor/en-ca/lib/js/ee.js:82:21 notifySaved@http://192.168.0.13:8111/reactor/en-ca/lib/js/expression-editor.js:47:26 @http://192.168.0.13:8111/reactor/en-ca/lib/js/reactor-ui-rules.js:1460:39 forEach@[native code] @http://192.168.0.13:8111/reactor/en-ca/lib/js/reactor-ui-rules.js:1459:58Chrome while creating/updating an expression
TypeError: Cannot read properties of undefined (reading 'getEditor') at RuleEditor.makeExprMenu (http://192.168.0.13:8111/reactor/en-ca/lib/js/rule-editor.js:1788:86) at Object.handler (http://192.168.0.13:8111/reactor/en-ca/lib/js/rule-editor.js:2174:54) at http://192.168.0.13:8111/client/MessageBus.js:98:44 at Array.forEach (<anonymous>) at MessageBus._sendToBus (http://192.168.0.13:8111/client/MessageBus.js:95:54) at MessageBus.send (http://192.168.0.13:8111/client/MessageBus.js:106:44) at ExpressionEditor.publish (http://192.168.0.13:8111/client/Observable.js:78:28) at ExpressionEditor.signalModified (http://192.168.0.13:8111/reactor/en-ca/lib/js/ee.js:146:14) at ExpressionEditor.signalModified (http://192.168.0.13:8111/reactor/en-ca/lib/js/expression-editor.js:40:15) at ExpressionEditor.reindexExpressions (http://192.168.0.13:8111/reactor/en-ca/lib/js/expression-editor.js:71:18) ``Not sure that it is the same issue but just got this on built 24302 when running a reaction for testing purpose. Despite the error message, the reaction ran properly.
Error: Command timeout (195 start_reaction)
at _ClientAPI._commandTimeout (http://192.168.2.163:8111/client/ClientAPI.js:552:136)
1a3422eb-d760-4609-a740-a40d04a6bab2-Screenshot 2024-12-29 231851.png
Thanks to @toggledbits for adding a custom CSS. I've started doing a darker Reactor style.
Here's the file: https://gist.github.com/dbochicchio/825098ac13b7f8cac22012eae37ff7ce
A couple of things are still too bright and I'll eventually catch-up. Just place it under your /config directory, naming the file as customstyles.css. Hard refresh your browser.
Need Testers
-
@toggledbits thanks for this feature! I sort of had to volunteer as a tester as I originally did a feature request for authentication support...
Anyway, seems to work nicely, although only about an hour of testing behind. Will keep you posted if anything unexpected comes up. You could update docs though:
"1. Copy the template users.yaml-dist from the dist-config distribution directory to your config directory."
--> "1. Copy the template users.yaml-dist from the /opt/reactor/dist-config distribution directory to your config directory. Rename it as users.yaml file."
(at least for me as a Docker user it wasn't immediately clear where to find this template)
-
Hi, I used Portainer to change the image from latest to userauth and it was easy and thanks for this update!
current configuration:
Synology NAS/Docker/Reactor (Multi-hub) userauth-24120-7745fb8dunexpected:
when I try to log in on http://10.0.0.1:8111/dashboard/ as admin:admin, I'm redirect to reactor http://10.0.0.1:8111/reactor/en-US/ and not to dashboard
This happens on Chrome and EdgeFeatures/suggestion that I image:
Guest user has access to:
all dashboard set to public (homescreen - all public group tile - all public single tile)where group tile are rooms/hubs/virtual/weather on homescreen dashboard
and single tile are all the entities: devices/virtual devices/etcAdmin can use option on tile VISIBLE/HIDDEN icon by clicking on it.
If a tile is set to HIDDEN, all entities associated with it will also be hidden. Setting VISIBLE on specific entities will not be respected. Setting hidden on specific entities it will be not visible to guests.
So Public means VISIBLE.imaging it:
-
Not sure if others are experiencing the same, but with this build (userauth-24120-7745fb8d) if I go to Manual (i.e. http://reactor-ip:8111/docs/), it takes a very long time before search can be used as it stays in "initializing search" state. At least I haven't noticed the same with earlier releases. Browser is Chrome.
-
@toggledbits another thing noticed with this build is that deleting global reactions does not work.
How to reproduce:
- create reaction
- delete it
- refresh browser
- reaction is back...
-
@tunnus said in Need Testers:
noticed with this build is that deleting global reactions does not work.
I see it. It's also happening to latest-24115. I've got it fixed on my end and will roll a build on Sunday.
@noelab said in Need Testers:
when I try to log in on http://10.0.0.1:8111/dashboard/ as admin:admin, I'm redirect to reactor http://10.0.0.1:8111/reactor/en-US/ and not to dashboard
I think I've got that nailed down as well. Look for Sunday's build.
@Pabla said in Need Testers:
Yes this would be a good improvement to the docs
I think better still, have Reactor docker containers copy
dist-config
to your data directory (asdist-config
, not overwriting any existing real config), so no special steps are needed to find those files, they are just there. Sunday's build. -
@tunnus said in Need Testers:
Not sure if others are experiencing the same, but with this build (userauth-24120-7745fb8d) if I go to Manual (i.e. http://reactor-ip:8111/docs/), it takes a very long time before search can be used as it stays in "initializing search" state. At least I haven't noticed the same with earlier releases. Browser is Chrome.
Anyone else noticed this? @noelab, @Pabla? Other thing I found out today was that there's some kind of a log problem going on, Docker logs claim that reactor.log file cannot be found:
2024/05/11 21:48:36 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/11 21:48:32 stdout [Reactor Log Panic] FileLogStream("/var/reactor/logs/reactor.log") error: Error [ERR_STREAM_WRITE_AFTER_END]: write after end 2024/05/11 21:48:32 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/11 21:48:32 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:32 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:32 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:32 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:32 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:32 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/11 21:48:09 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/11 21:48:08 stdout [Reactor Log Panic] FileLogStream("/var/reactor/logs/reactor.log") error: Error [ERR_STREAM_WRITE_AFTER_END]: write after end 2024/05/11 21:48:08 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/11 21:48:08 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:08 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:08 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:48:08 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/11 21:46:50 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/11 21:46:45 stdout [Reactor Log Panic] FileLogStream("/var/reactor/logs/reactor.log") error: Error [ERR_STREAM_WRITE_AFTER_END]: write after end 2024/05/11 21:46:45 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/11 21:46:45 stdout /var/reactor/logs/reactor.log not found; reopening stream
But reactor.log file seems to in place when checked via terminal:
-
@tunnus yes, some problem here too from the reactor log in container manager. Hope this may help too:
2024/05/08 21:20:46 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/08 21:20:46 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:20:46 stdout [Reactor Log Panic] FileLogStream("/var/reactor/logs/reactor.log") error: Error [ERR_STREAM_WRITE_AFTER_END]: write after end 2024/05/08 21:20:46 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/08 21:20:46 stdout [userauth-24120]2024-05-08T19:20:46.401Z <Timer:null> Timer#rule-lq5rhax4 just a note: I'm setting a delay of only 7ms (from 1715196046401<5/8/2024, 9:20:46 PM> to 1715196046408<5/8/2024, 9:20:46 PM>) 2024/05/08 21:20:45 stdout [userauth-24120]2024-05-08T19:20:45.347Z <Timer:null> Timer#rule-lq1iq5i3 just a note: I'm setting a delay of only 8ms (from 1715196045347<5/8/2024, 9:20:45 PM> to 1715196045355<5/8/2024, 9:20:45 PM>) 2024/05/08 21:20:42 stdout [userauth-24120]2024-05-08T19:20:42.093Z <Timer:null> Timer#rule-lq5ri09p just a note: I'm setting a delay of only 8ms (from 1715196042093<5/8/2024, 9:20:42 PM> to 1715196042101<5/8/2024, 9:20:42 PM>) 2024/05/08 21:19:04 stdout [userauth-24120]2024-05-08T19:19:04.547Z <Timer:null> Timer#rule-lq5rgx5t just a note: I'm setting a delay of only 8ms (from 1715195944547<5/8/2024, 9:19:04 PM> to 1715195944555<5/8/2024, 9:19:04 PM>) 2024/05/08 21:19:02 stdout [userauth-24120]2024-05-08T19:19:02.765Z <Timer:null> Timer#rule-lq5rhvua just a note: I'm setting a delay of only 8ms (from 1715195942765<5/8/2024, 9:19:02 PM> to 1715195942773<5/8/2024, 9:19:02 PM>) 2024/05/08 21:19:02 stdout [userauth-24120]2024-05-08T19:19:02.153Z <EzloController:null> remapped item action to [Object]{ "item": "switch", "value": true, "method": "hub.item.value.set", "parameters": { "value": { "value": true } } } 2024/05/08 21:19:01 stdout [userauth-24120]2024-05-08T19:19:01.110Z <Timer:null> Timer#rule-lq5qgdau just a note: I'm setting a delay of only 8ms (from 1715195941110<5/8/2024, 9:19:01 PM> to 1715195941118<5/8/2024, 9:19:01 PM>) 2024/05/08 21:18:59 stdout [userauth-24120]2024-05-08T19:18:59.226Z <Timer:null> Timer#rule-lq5rh750 just a note: I'm setting a delay of only 8ms (from 1715195939226<5/8/2024, 9:18:59 PM> to 1715195939234<5/8/2024, 9:18:59 PM>) 2024/05/08 21:17:10 stdout [userauth-24120]2024-05-08T19:17:10.678Z <Timer:null> Timer#rule-lq5rhax4 just a note: I'm setting a delay of only 8ms (from 1715195830678<5/8/2024, 9:17:10 PM> to 1715195830686<5/8/2024, 9:17:10 PM>) 2024/05/08 21:14:44 stdout [userauth-24120]2024-05-08T19:14:44.540Z <Timer:null> Timer#rule-lq5rh750 just a note: I'm setting a delay of only 8ms (from 1715195684540<5/8/2024, 9:14:44 PM> to 1715195684548<5/8/2024, 9:14:44 PM>) 2024/05/08 21:09:18 stdout [userauth-24120]2024-05-08T19:09:18.150Z <Timer:null> Timer#rule-lq5rhax4 just a note: I'm setting a delay of only 8ms (from 1715195358150<5/8/2024, 9:09:18 PM> to 1715195358158<5/8/2024, 9:09:18 PM>) 2024/05/08 21:05:44 stdout [userauth-24120]2024-05-08T19:05:44.738Z <Timer:null> Timer#rule-ks56ms0b just a note: I'm setting a delay of only 8ms (from 1715195144738<5/8/2024, 9:05:44 PM> to 1715195144746<5/8/2024, 9:05:44 PM>) 2024/05/08 21:05:44 stdout [userauth-24120]2024-05-08T19:05:44.719Z <Timer:null> Timer#rule-lq5rh750 just a note: I'm setting a delay of only 8ms (from 1715195144719<5/8/2024, 9:05:44 PM> to 1715195144727<5/8/2024, 9:05:44 PM>) 2024/05/08 21:00:16 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/08 21:00:14 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout [Reactor Log Panic] FileLogStream("/var/reactor/logs/reactor.log") error: Error [ERR_STREAM_WRITE_AFTER_END]: write after end 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 21:00:12 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/08 20:59:52 stdout [userauth-24120]2024-05-08T18:59:52.886Z <Timer:null> Timer#rule-lq5rhax4 just a note: I'm setting a delay of only 8ms (from 1715194792886<5/8/2024, 8:59:52 PM> to 1715194792894<5/8/2024, 8:59:52 PM>) 2024/05/08 20:58:22 stdout [userauth-24120]2024-05-08T18:58:22.585Z <Timer:null> Timer#rule-lq5rh750 just a note: I'm setting a delay of only 8ms (from 1715194702585<5/8/2024, 8:58:22 PM> to 1715194702593<5/8/2024, 8:58:22 PM>) 2024/05/08 20:56:26 stdout [userauth-24120]2024-05-08T18:56:26.917Z <Timer:null> Timer#rule-lq5rh1ns just a note: I'm setting a delay of only 8ms (from 1715194586917<5/8/2024, 8:56:26 PM> to 1715194586925<5/8/2024, 8:56:26 PM>) 2024/05/08 20:56:06 stdout [userauth-24120]2024-05-08T18:56:06.581Z <Timer:null> Timer#rule-lq1iq5i3 just a note: I'm setting a delay of only 8ms (from 1715194566581<5/8/2024, 8:56:06 PM> to 1715194566589<5/8/2024, 8:56:06 PM>) 2024/05/08 20:54:49 stdout [userauth-24120]2024-05-08T18:54:49.514Z <Timer:null> Timer#rule-lq5rgx5t just a note: I'm setting a delay of only 8ms (from 1715194489514<5/8/2024, 8:54:49 PM> to 1715194489522<5/8/2024, 8:54:49 PM>) 2024/05/08 20:54:49 stdout [userauth-24120]2024-05-08T18:54:49.058Z <Timer:null> Timer#rule-lq5qgdau just a note: I'm setting a delay of only 8ms (from 1715194489058<5/8/2024, 8:54:49 PM> to 1715194489066<5/8/2024, 8:54:49 PM>) 2024/05/08 20:54:43 stdout [userauth-24120]2024-05-08T18:54:43.027Z <Timer:null> Timer#rule-lq5qgdau just a note: I'm setting a delay of only 8ms (from 1715194483027<5/8/2024, 8:54:43 PM> to 1715194483035<5/8/2024, 8:54:43 PM>) 2024/05/08 20:50:44 stdout [userauth-24120]2024-05-08T18:50:44.225Z <Timer:null> Timer#rule-lq5rh750 just a note: I'm setting a delay of only 8ms (from 1715194244225<5/8/2024, 8:50:44 PM> to 1715194244233<5/8/2024, 8:50:44 PM>) 2024/05/08 20:50:43 stdout [userauth-24120]2024-05-08T18:50:43.256Z <EzloController:null> remapped item action to [Object]{ "item": "dimmer", "value_expr": "min(100,max(0,floor(parameters.level*100+0.5)))", "method": "hub.item.value.set", "parameters": { "value": { "value_expr": "min(100,max(0,floor(parameters.level*100+0.5)))" } } } 2024/05/08 20:47:23 stdout [userauth-24120]2024-05-08T18:47:23.713Z <Timer:null> Timer#rule-ks56ms0b just a note: I'm setting a delay of only 8ms (from 1715194043713<5/8/2024, 8:47:23 PM> to 1715194043721<5/8/2024, 8:47:23 PM>) 2024/05/08 20:47:23 stdout [userauth-24120]2024-05-08T18:47:23.693Z <Timer:null> Timer#rule-lq5rhax4 just a note: I'm setting a delay of only 8ms (from 1715194043693<5/8/2024, 8:47:23 PM> to 1715194043701<5/8/2024, 8:47:23 PM>) 2024/05/08 20:47:22 stdout [userauth-24120]2024-05-08T18:47:22.392Z <EzloController:null> remapped item action to [Object]{ "item": "switch", "value": false, "method": "hub.item.value.set", "parameters": { "value": { "value": false } } } 2024/05/08 20:46:10 stdout [userauth-24120]2024-05-08T18:46:09.716Z <default:null> Module NotifyAlert v21092 2024/05/08 20:46:10 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/08 20:46:10 stdout [userauth-24120]2024-05-08T18:46:09.704Z <default:null> Module NotifyPushover v21195 2024/05/08 20:46:10 stdout [Reactor Log Panic] FileLogStream("/var/reactor/logs/reactor.log") error: Error [ERR_STREAM_WRITE_AFTER_END]: write after end 2024/05/08 20:46:10 stdout [Reactor Log Panic] FileLogStream("/var/reactor/logs/reactor.log") error: Error [ERR_STREAM_WRITE_AFTER_END]: write after end 2024/05/08 20:46:10 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/08 20:46:10 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/08 20:46:10 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 20:46:10 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 20:46:10 stdout /var/reactor/logs/reactor.log not found; reopening stream 2024/05/08 20:46:10 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/08 20:46:09 stdout [userauth-24120]2024-05-08T18:46:09.014Z <Notifier:null> Module Notifier v22283 2024/05/08 20:46:06 stdout at /opt/reactor/server/lib/VirtualEntityController.js:357:319 2024/05/08 20:46:06 stdout at VirtualEntityController._load_devices (/opt/reactor/server/lib/VirtualEntityController.js:644:114) 2024/05/08 20:46:06 stdout at Entity.setPrimaryAttribute (/opt/reactor/server/lib/Entity.js:563:19) 2024/05/08 20:46:06 stdout Error: Invalid primary attribute assignment; capability not assigned (state_sensor.value) 2024/05/08 20:46:06 stdout [userauth-24120]2024-05-08T18:46:06.542Z <VirtualEntityController:CRIT> Error: Invalid primary attribute assignment; capability not assigned (state_sensor.value) [-] 2024/05/08 20:46:06 stdout at /opt/reactor/server/lib/VirtualEntityController.js:357:319 2024/05/08 20:46:06 stdout at VirtualEntityController._load_devices (/opt/reactor/server/lib/VirtualEntityController.js:644:114) 2024/05/08 20:46:06 stdout at Entity.setPrimaryAttribute (/opt/reactor/server/lib/Entity.js:563:19) 2024/05/08 20:46:06 stdout Error: Invalid primary attribute assignment; capability not assigned (string_sensor.value) 2024/05/08 20:46:06 stdout [userauth-24120]2024-05-08T18:46:06.541Z <VirtualEntityController:CRIT> Error: Invalid primary attribute assignment; capability not assigned (string_sensor.value) [-] 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:05.410Z <VirtualEntityController:null> Module VirtualEntityController v24117 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:05.391Z <SystemController:null> Module SystemController v24076 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:05.382Z <OWMWeatherController:null> Module OWMWeatherController v22294 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:05.161Z <EzloController:null> Module EzloController v23345 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:05.090Z <VeraController:null> Module VeraController v24050 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:05.089Z <TaskQueue:null> Module TaskQueue 24113 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.993Z <HassController:null> Module HassController v24115 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.726Z <InfluxFeed:null> Module InfluxFeed v23341 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.682Z <wsapi:null> Module wsapi v24115 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.439Z <httpapi:null> Module httpapi v24119 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.231Z <Engine:null> Module Engine v24113 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.229Z <GlobalReaction:null> Module GlobalReaction v24099 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.197Z <Rule:null> Module Rule v24115 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.176Z <AlertManager:null> Module AlertManager v24099 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.107Z <Predicate:null> Module Predicate v23093 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.045Z <GlobalExpression:null> Module GlobalExpression v24099 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.023Z <default:null> Module Rulesets v24099 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:04.022Z <default:null> Module Ruleset v24099 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:03.952Z <Controller:null> Module Controller v24099 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:03.929Z <Entity:null> Module Entity v24108 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:03.925Z <TimerBroker:null> Module TimerBroker v22283 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:03.460Z <Plugin:null> Module Plugin v22300 2024/05/08 20:46:05 stdout [userauth-24120]2024-05-08T18:46:03.311Z <Capabilities:null> Module Capabilities v23331 2024/05/08 20:46:03 stdout [userauth-24120]2024-05-08T18:46:03.269Z <Structure:null> Module Structure v24110 2024/05/08 20:46:03 stdout [userauth-24120]2024-05-08T18:46:03.059Z <app:null> Local date/time using configured timezone and locale formatting is "5/8/2024, 8:46:03 PM" 2024/05/08 20:46:03 stdout [userauth-24120]2024-05-08T18:46:03.058Z <app:null> Loaded locale en-US for en-US 2024/05/08 20:46:02 stdout [userauth-24120]2024-05-08T18:46:02.890Z <app:null> Configured locale (undefined); selected locale(s) en-US.UTF-8 2024/05/08 20:46:02 stdout [userauth-24120]2024-05-08T18:46:02.750Z <app:null> Resolved timezone=Europe/Rome, environment TZ=Europe/Rome; offset minutes from UTC=120 2024/05/08 20:46:02 stdout [userauth-24120]2024-05-08T18:46:02.656Z <app:null> NODE_PATH=/opt/reactor:/opt/reactor/node_modules 2024/05/08 20:46:02 stdout [userauth-24120]2024-05-08T18:46:02.656Z <app:null> Basedir /opt/reactor; data in /var/reactor/storage 2024/05/08 20:46:02 stdout [userauth-24120]2024-05-08T18:46:02.656Z <app:null> Process ID 1 user/group 0/0; docker; platform linux/x64 #69057 SMP Fri Jan 12 17:02:28 CST 2024; locale (undefined) 2024/05/08 20:46:02 stdout [userauth-24120]2024-05-08T18:46:02.584Z <app:null> Reactor build userauth-24120-7745fb8d starting on v20.10.0 /usr/local/bin/node 2024/05/08 20:46:02 stdout NODE_PATH /opt/reactor:/opt/reactor/node_modules 2024/05/08 20:46:02 stdout Reactor userauth-24120-7745fb8d app [33m24120[39m configuration from /var/reactor/config
-
It looks like you may have a misconfiguration in your VirtualEntityController config. Can you post the config section for that?
-
@toggledbits
some devices are created for test purpose so maybe there are errorscode_text
-
id: virtual
name: VEC
implementation: VirtualEntityController
enabled: true
config:
entities:HomeMode
- id: v00_Automation template: Binary Sensor name: v00 Automation Mode
Indici Temperatura Umidità
- id: v00_InsideHOT template: Binary Sensor name: v00 Temp Inside HOT - id: v00_OutsideHOT template: Value Sensor name: v00 Temp Outside HOT - id: v00_TempOutsideInside template: Binary Sensor name: v00 Temp Outside>Inside - id: v00_umidita_assoluta_interna name: v00 Umidita Assoluta Interna capabilities: value_sensor: attributes: value: expr: xWeather_Umidita_Assoluta_Interna primary_attribute: value_sensor.value type: ValueSensor - id: v00_umidita_assoluta_esterna name: v00 Umidita Assoluta Esterna capabilities: value_sensor: attributes: value: expr: xWeather_Umidita_Assoluta_Esterna primary_attribute: value_sensor.value type: ValueSensor
IndiceThom
- id: vB1_IndiceThom template: Value Sensor name: vB1 IndiceThom capabilities: value_sensor: attributes: value: expr: xB1_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vB2_IndiceThom name: vB2 IndiceThom capabilities: value_sensor: attributes: value: expr: xB2_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vCS_IndiceThom name: vCS IndiceThom capabilities: value_sensor: attributes: value: expr: xCS_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vH_IndiceThom name: vH IndiceThom capabilities: value_sensor: attributes: value: expr: xH_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vK_IndiceThom name: vK IndiceThom capabilities: value_sensor: attributes: value: expr: xK_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vL_IndiceThom name: vL IndiceThom capabilities: value_sensor: attributes: value: expr: xL_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR1_IndiceThom name: vR1 IndiceThom capabilities: value_sensor: attributes: value: expr: xR1_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR2_IndiceThom name: vR2 IndiceThom capabilities: value_sensor: attributes: value: expr: xR2_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR3_IndiceThom name: vR3 IndiceThom capabilities: value_sensor: attributes: value: expr: xR3_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR4_IndiceThom name: vR4 IndiceThom capabilities: value_sensor: attributes: value: expr: xR4_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor - id: vT_IndiceThom name: vT IndiceThom capabilities: value_sensor: attributes: value: expr: xT_IndiceThom_ID primary_attribute: value_sensor.value type: ValueSensor
LUX Funzionale
- id: vB1_LuxFunzionale name: vB1 LuxFunzionale capabilities: value_sensor: attributes: value: expr: xB1_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vB2_LuxFunzionale name: vB2 LuxFunzionale capabilities: value_sensor: attributes: value: expr: xB2_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vCS_LuxFunzionale name: vCS LuxFunzionale capabilities: value_sensor: attributes: value: expr: xCS_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vH_LuxFunzionale name: vH LuxFunzionale capabilities: value_sensor: attributes: value: expr: xH_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vK_LuxFunzionale name: vK LuxFunzionale capabilities: value_sensor: attributes: value: expr: xK_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vL_LuxFunzionale name: vL LuxFunzionale capabilities: value_sensor: attributes: value: expr: xL_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR1_LuxFunzionale name: vR1 LuxFunzionale capabilities: value_sensor: attributes: value: expr: xR1_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR2_LuxFunzionale name: vR2 LuxFunzionale capabilities: value_sensor: attributes: value: expr: xR2_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR3_LuxFunzionale name: vR3 LuxFunzionale capabilities: value_sensor: attributes: value: expr: xR3_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR4_LuxFunzionale name: vR4 LuxFunzionale capabilities: value_sensor: attributes: value: expr: xR4_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vT_LuxFunzionale name: vT LuxFunzionale capabilities: value_sensor: attributes: value: expr: xT_Lux_Funzionale_ID primary_attribute: value_sensor.value type: ValueSensor
LUX AMBIENTALE
- id: vB1_LuxAmbientale name: vB1 LuxAmbientale capabilities: value_sensor: attributes: value: expr: xB1_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vB2_LuxAmbientale name: vB2 LuxAmbientale capabilities: value_sensor: attributes: value: expr: xB2_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vCS_LuxAmbientale name: vCS LuxAmbientale capabilities: value_sensor: attributes: value: expr: xCS_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vH_LuxAmbientale name: vH LuxAmbientale capabilities: value_sensor: attributes: value: expr: xH_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vK_LuxAmbientale name: vK LuxAmbientale capabilities: value_sensor: attributes: value: expr: xK_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vL_LuxAmbientale name: vL LuxAmbientale capabilities: value_sensor: attributes: value: expr: xL_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR1_LuxAmbientale name: vR1 LuxAmbientale capabilities: value_sensor: attributes: value: expr: xR1_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR2_LuxAmbientale name: vR2 LuxAmbientale capabilities: value_sensor: attributes: value: expr: xR2_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR3_LuxAmbientale name: vR3 LuxAmbientale capabilities: value_sensor: attributes: value: expr: xR3_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vR4_LuxAmbientale name: vR4 LuxAmbientale capabilities: value_sensor: attributes: value: expr: xR4_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor - id: vT_LuxAmbientale name: vT LuxAmbientale capabilities: value_sensor: attributes: value: expr: xT_Lux_Ambientale_ID primary_attribute: value_sensor.value type: ValueSensor
State
EnergyMode10
- id: v00_EnergyMode_Biorario template: Binary Sensor name: v00 EnergyMode Biorario - id: v00_EnergyMode_Sovraccarico_Locale template: Binary Sensor name: v00 EnergyMode Sovraccarico Locale - id: v00_EnergyMode_Lavatrice_State name: v00 EnergyMode Lavatrice State template: String Sensor capabilities: string_sensor: attributes: value: expr: xB1_Lavatrice_State - id: v00_EnergyMode_Oven_State name: v00 EnergyMode Oven State template: String Sensor capabilities: string_sensor: attributes: value: expr: xK_Oven_State - id: v00_EnergyMode_Fridge_State name: v00 EnergyMode Fridge State template: String Sensor capabilities: string_sensor: attributes: value: expr: xK_Fridge_State - id: v00_EnergyMode_Dishwasher_State name: v00 EnergyMode Dishwasher State template: String Sensor capabilities: string_sensor: attributes: value: expr: xK_Dishwasher_State
HVACMode20
TEST100
- id: testA template: Binary Switch name: TestA Binary - id: testB template: Binary Switch name: TestB Binary - id: testC template: Binary Switch name: TestC Binary - id: testD template: Binary Switch name: TestD Binary - id: testE template: Binary Switch name: TestE Binary - id: testF template: Value Sensor name: TestF ValueSensor - id: testG template: Value Sensor name: TestG ValueSensor - id: testH template: String Sensor name: TestH StringSensor - id: testI template: String Sensor name: TestI StringSensor - id: testJ template: String Sensor name: TestJ StringSensor - id: virtual_therm name: Test Virtual Heating Thermostat capabilities: - hvac_control - hvac_heating_unit - temperature_sensor primary_attribute: hvac_control.state - id: color_dimmerA template: Dimmer name: Test Virtual Color DimmerA capabilities: - color_temperature - id: color_dimmerB template: Dimmer name: Test Virtual Color DimmerB - id: testK template: Binary Sensor name: TestK Binary - id: mode name: Test ModeHouse StringSensor template: String Sensor capabilities: string_sensor: attributes: value: expr: x00_HouseMode_StartState - id: testA_dynamic_vec name: TestA Dynamic Vec Binary capabilities: value_sensor: attributes: value: expr: x00_HouseMode_StartState primary_attribute: string_sensor.value type: StringSensor - id: testB_dynamic_vec name: TestB Dynamic Vec Binary capabilities: binary_sensor: attributes: state: expr: primary_attribute: state_sensor.value type: BinarySensor - id: testC_dynamic_vec name: TestC Dynamic Vec Binary capabilities: value_sensor: attributes: value: if_expr: test - id: testD_dynamic_vec name: TEST D Dynamic VEC capabilities: binary_sensor: attributes: state: expr:
END VIRTUAL DEVICE
-
-
@noelab said in Need Testers:
- id: testB_dynamic_vec
name: TestB Dynamic Vec Binary
capabilities:
binary_sensor:
attributes:
state:
expr:
primary_attribute: state_sensor.value
type: BinarySensor
In this config, you are declaring
binary_sensor
as a capability, but then trying to set the primary attribute tostate_sensor.value
-- the capabilitystate_sensor
is not a Reactor-defined capability. That should be corrected.I have other changes coming today to help examine the timer complaints. The logging issue seems like it should be more rare -- it's a race condition in the asynchronous writing of streams. I expect to see this messages occasionally, rarely (as I said), but in runs like you are getting is not cool, but I think it's induced by the timer complaint. Stay tuned.
Also, @noelab and @tunnus the log snippets you posted look heavily filtered. That removes important context, so please don't do that. I'd rather see one error message with full context than 20 with none.
- id: testB_dynamic_vec
-
Build 24133 userauth
- Force logging to be synchronous, to work around many long-standing issues with the naturally asynchronous nature of nodejs streams.
- Enhance diagnostic reporting for short timers (temporarily, in this branch only).
- Fix some issues with post-login redirection (if you were pushed to the login page by the dashboard or admin UI).
- VirtualEntityController: preserve attribute values if possible across system updates that affect capability definitions or implementations.
- DynamicGroupController: it is now possible to set the primary attribute of a dynamic group, and determine its value dynamically. Trivial example: add the
binary_sensor.state
attribute to a group to show true when any light in the group is on. Refer to the docs for DynamicGroupController. - The
dist-config
directory is now copied to the user data virtual path target when running in docker containers so that its contents are more easily accessible to docker users. - Fixed an issue in the core related to deleting certain objects (a Global Reaction was reported, but could have been many object types).
- Dashboard: further improvements to widget type guessing if we're not told by the entity configuration.
- Bring
dist-config
version ofusers.yaml
into line with current reality. - Fix the
rpi-status.sh
tool to work better with HTTPS enabled. - HassController: Bless Hass to 2024.5.2
Note that this build incorporates all changes for
userauth
as well aslatest
branch changes not yet deployed for everyone else. -
@toggledbits I think you also forgot to push userauth-amd64. thanks.
-
@tunnus said in Need Testers:
@tunnus said in Need Testers:
Not sure if others are experiencing the same, but with this build (userauth-24120-7745fb8d) if I go to Manual (i.e. http://reactor-ip:8111/docs/), it takes a very long time before search can be used as it stays in "initializing search" state. At least I haven't noticed the same with earlier releases. Browser is Chrome.
Anyone else noticed this? @noelab, @Pabla? Other thing I found out today was that there's some kind of a log problem going on, Docker logs claim that reactor.log file cannot be found:
Just quickly tested this and yes it says initializing search. I am using Safari
-
@toggledbits can confirm that this new build (24133) fixes "delete problem", i.e. global reactions can now be deleted without them reappearing. Not seeing those "Reactor Log Panic" messages, but other errors:
2024/05/13 18:51:08 stdout at async /opt/reactor/server/lib/Rule.js:885:17 2024/05/13 18:51:08 stdout at process.processTicksAndRejections (node:internal/process/task_queues:95:5) 2024/05/13 18:51:08 stdout at Rule._evaluate (/opt/reactor/server/lib/Rule.js:943:69) 2024/05/13 18:51:08 stdout at Timer.delayms (/opt/reactor/server/lib/Timer.js:151:132) 2024/05/13 18:51:08 stdout at Timer.at (/opt/reactor/server/lib/Timer.js:147:381) 2024/05/13 18:51:08 stdout Error: Delay Source Traceback 2024/05/13 18:51:08 stdout [userauth-24133]2024-05-13T15:51:08.837Z <Timer:CRIT> Error: Delay Source Traceback [-] 2024/05/13 18:51:08 stdout [userauth-24133]2024-05-13T15:51:08.837Z <Timer:null> Timer#rule-ktmrcd6d just a note: I'm setting a delay of only 8ms (from 1715615468837<5/13/2024, 6:51:08 PM> to 1715615468845<5/13/2024, 6:51:08 PM>) 2024/05/13 18:51:00 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/13 18:51:00 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/13 18:50:03 stdout [userauth-24133]2024-05-13T15:50:03.915Z <wsapi:CRIT> wsapi: client "172.18.0.1#5" sending authreq 2024/05/13 18:50:02 stdout [userauth-24133]2024-05-13T15:50:02.092Z <wsapi:CRIT> wsapi: client "172.18.0.1#4" sending authreq 2024/05/13 18:49:33 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/13 18:49:33 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/13 18:48:15 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/13 18:48:15 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/13 18:47:07 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/13 18:47:07 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/13 18:46:29 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/13 18:46:29 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/13 18:46:25 stdout at async /opt/reactor/server/lib/Rule.js:885:17 2024/05/13 18:46:25 stdout at process.processTicksAndRejections (node:internal/process/task_queues:95:5) 2024/05/13 18:46:25 stdout at Rule._evaluate (/opt/reactor/server/lib/Rule.js:943:69) 2024/05/13 18:46:25 stdout at Timer.delayms (/opt/reactor/server/lib/Timer.js:151:132) 2024/05/13 18:46:25 stdout at Timer.at (/opt/reactor/server/lib/Timer.js:147:381) 2024/05/13 18:46:25 stdout Error: Delay Source Traceback 2024/05/13 18:46:25 stdout [userauth-24133]2024-05-13T15:46:25.384Z <Timer:CRIT> Error: Delay Source Traceback [-] 2024/05/13 18:46:25 stdout [userauth-24133]2024-05-13T15:46:25.384Z <Timer:null> Timer#rule-ktmrcd6d just a note: I'm setting a delay of only 8ms (from 1715615185384<5/13/2024, 6:46:25 PM> to 1715615185392<5/13/2024, 6:46:25 PM>) 2024/05/13 18:45:10 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/13 18:45:10 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/13 18:44:15 stdout at async /opt/reactor/server/lib/Rule.js:885:17 2024/05/13 18:44:15 stdout at process.processTicksAndRejections (node:internal/process/task_queues:95:5) 2024/05/13 18:44:15 stdout at Rule._evaluate (/opt/reactor/server/lib/Rule.js:943:69) 2024/05/13 18:44:15 stdout at Timer.delayms (/opt/reactor/server/lib/Timer.js:151:132) 2024/05/13 18:44:15 stdout at Timer.at (/opt/reactor/server/lib/Timer.js:147:381) 2024/05/13 18:44:15 stdout Error: Delay Source Traceback 2024/05/13 18:44:15 stdout [userauth-24133]2024-05-13T15:44:15.282Z <Timer:CRIT> Error: Delay Source Traceback [-] 2024/05/13 18:44:15 stdout [userauth-24133]2024-05-13T15:44:15.281Z <Timer:null> Timer#rule-grp11myuubg just a note: I'm setting a delay of only 5ms (from 1715615055281<5/13/2024, 6:44:15 PM> to 1715615055286<5/13/2024, 6:44:15 PM>) 2024/05/13 18:44:00 stdout stream /var/reactor/logs/reactor.log not open boolean [33mfalse[39m 2024/05/13 18:44:00 stdout /var/reactor/logs/reactor.log size hit rotation limit, rotating 2024/05/13 18:43:44 stdout [userauth-24133]2024-05-13T15:43:44.182Z <wsapi:CRIT> wsapi: client "172.18.0.1#3" sending authreq 2024/05/13 18:43:36 stdout [userauth-24133]2024-05-13T15:43:36.850Z <wsapi:CRIT> wsapi: client "172.18.0.1#2" sending authreq 2024/05/13 18:43:27 stdout [userauth-24133]2024-05-13T15:43:27.574Z <default:null> Module NotifyTelegram v21221 2024/05/13 18:43:27 stdout [userauth-24133]2024-05-13T15:43:27.564Z <wsapi:CRIT> wsapi: client "172.18.0.1#1" sending authreq 2024/05/13 18:43:27 stdout [userauth-24133]2024-05-13T15:43:27.293Z <Notifier:null> Module Notifier v22283 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.989Z <SystemController:null> Module SystemController v24076 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.980Z <OWMWeatherController:null> Module OWMWeatherController v22294 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.893Z <HubitatController:null> Module HubitatController v24076 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.830Z <HassController:null> Module HassController v24128 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.526Z <MQTTController:null> Module MQTTController v24120 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.488Z <VirtualEntityController:null> Module VirtualEntityController v24117 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.431Z <VeraController:null> Module VeraController v24050 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.431Z <TaskQueue:null> Module TaskQueue 24113 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.416Z <InfluxFeed:null> Module InfluxFeed v23341 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.312Z <wsapi:null> Module wsapi v24115 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.284Z <httpapi:null> Module httpapi v24121 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.275Z <Engine:null> Module Engine v24113 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.273Z <GlobalReaction:null> Module GlobalReaction v24099 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.268Z <Rule:null> Module Rule v24115 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.265Z <AlertManager:null> Module AlertManager v24099 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.259Z <Predicate:null> Module Predicate v23093 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.243Z <GlobalExpression:null> Module GlobalExpression v24099 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.232Z <default:null> Module Rulesets v24099 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.231Z <default:null> Module Ruleset v24099 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.219Z <Controller:null> Module Controller v24099 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.215Z <Entity:null> Module Entity v24108 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.208Z <TimerBroker:null> Module TimerBroker v22283 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.180Z <Plugin:null> Module Plugin v22300 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.115Z <Capabilities:null> Module Capabilities v23331 2024/05/13 18:43:21 stdout [userauth-24133]2024-05-13T15:43:21.112Z <Structure:null> Module Structure v24110 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.999Z <app:null> Local date/time using configured timezone and locale formatting is "5/13/2024, 6:43:20 PM" 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.998Z <app:null> Loaded locale en-US for en-US 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.969Z <app:null> Configured locale (undefined); selected locale(s) en-US.UTF-8 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.966Z <app:null> Resolved timezone=Europe/Helsinki, environment TZ=Europe/Helsinki; offset minutes from UTC=180 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.814Z <app:null> NODE_PATH=/opt/reactor:/opt/reactor/node_modules 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.814Z <app:null> Basedir /opt/reactor; data in /var/reactor/storage 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.813Z <app:null> Process ID 1 user/group 0/0; docker; platform linux/x64 #69057 SMP Fri Jan 12 17:02:28 CST 2024; locale (undefined) 2024/05/13 18:43:20 stdout [userauth-24133]2024-05-13T15:43:20.770Z <app:null> Reactor build userauth-24133-a2bf8846 starting on v20.10.0 /usr/local/bin/node 2024/05/13 18:43:20 stdout NODE_PATH /opt/reactor:/opt/reactor/node_modules 2024/05/13 18:43:20 stdout Reactor userauth-24133-a2bf8846 app [33m24127[39m configuration from /var/reactor/config
Also that manual/docs related problem ("initializing search") is still there, but that was expected as there were no mention of it being fixed.
-
@tunnus The log snippets you are posting appear to be from something other than the log files produced by Reactor. This looks like console log for a container, and I'm not interested in looking at that, mostly because they are a limited subset of message severity and, as I said earlier, drop a lot of other context is valuable to me in evaluating what I'm seeing. Reactor produces its own log files for a reason. If you want to post snippets from that, with context in accordance with the posting guidelines, I'm ready to look at those any time.
-
@toggledbits ok, sorry about that snippet, it was indeed a console log for a container.
I looked through reactor logs, but couldn't see any errors there. A bit challenging as there's a lot going on (a couple of variables update very often, every 1-3 seconds and those cause a lot of log events as rules referencing them get evaluated). Btw, there's a related request/question in this thread.
-
@tunnus If your specific concern is the Timer messages, don't be concerned. Long story shortened: these Timer objects were originally used for long intervals and the warning was put in place to help identify bugs in computing the length of those intervals, but they ended up being used for short intervals, too, and there are some conditions where those intervals can randomly be so short that they trigger the warning. I think it's time for this warning to be relegated to debug level or removed.
You clearly have something that's logging a lot, and that's causing frequent log rotations. This not only wastes time/cycles and bytes on the disk, but it also makes debugging anything on your system harder for the reason you stated: you can't find the errors, likely because (a) it's logging too much (you're sipping from a firehose), and (b) it's rotating so frequently that you're losing useful data quickly.
What I'd suggest is that you identify the affected rules (i.e. those that are responding to the frequent entity attribute changes), and increase their specific log levels to reduce the log output. You can do that for an individual rule. You need to get the rule ID, which you can derive from the UI by opening the rule's state in the list. You can also see it logged. Let's pick on
rule-ktmrcd6d
... we modifyconfig/logging.yaml
by adding the following:# The line below is indented two spaces; the line after is indented four spaces. "Rule#rule-ktmrcd6d": level: 3
This puts the logging minimum level at "NOTICE", so all of the INFO messages associated with rule evaluation will be suppressed for that rule only (other rules will not be affected). That should considerably reduce your log traffic.
I saw your question about the frequent updates as it relates to InfluxDB, and that's a sticky business. I'll respond to that over there.