It’s got a nice component to go with it, so setting up is easier. I particularly use it for scheduling thermostats, and find it much more user friendly. Sure I could do it with automations, but I’d either have one, massively unwieldy one with lots of states and triggers, or lots of individual ones.
I don’t know how tech savvy you are, but I’m assuming since your on lemmy it’s pretty good :)
The way we’ve solved this sort of problem in the office is by using the LLM’s JSON response, and a prompt that essentially keeps a set of JSON objects alongside the actual chat response.
In the DND example, this would be a set character sheets that get returned every response but only changed when the narrative changes them. More expensive, and needing a larger context window, but reasonably effective.