Im using Ollama on my server with the WebUI. It has no GPU so its not quick to reply but not too slow either.
Im thinking about removing the VM as i just dont use it, are there any good uses or integrations into other apps that might convince me to keep it?
Yeah even gpt4o couldn’t keep track of encounters, run battles etc. in my case…
I think if you wanted to do it mechanically consistently you’d probably need to integrate it into a vtt where you give it context and potentially fine-tune it to give quest related summaries & gming rather than just “stuff”
I don’t know how tech savvy you are, but I’m assuming since your on lemmy it’s pretty good :)
The way we’ve solved this sort of problem in the office is by using the LLM’s JSON response, and a prompt that essentially keeps a set of JSON objects alongside the actual chat response.
In the DND example, this would be a set character sheets that get returned every response but only changed when the narrative changes them. More expensive, and needing a larger context window, but reasonably effective.
VTT integration would be one hell of a job to do.