Papers
arxiv:2605.12178

Do Enterprise Systems Need Learned World Models? The Importance of Context to Infer Dynamics

Published on May 12
· Submitted by
Patrice Bechard
on May 13
Authors:
,
,
,
,
,
,
,
,
,
,
,
,
,

Abstract

Enterprise discovery agents that read system configuration at runtime outperform traditional world models in configurable environments where dynamics change over time.

AI-generated summary

World models enable agents to anticipate the effects of their actions by internalizing environment dynamics. In enterprise systems, however, these dynamics are often defined by tenant-specific business logic that varies across deployments and evolves over time, making models trained on historical transitions brittle under deployment shift. We ask a question the world-models literature has not addressed: when the rules can be read at inference time, does an agent still need to learn them? We argue, and demonstrate empirically, that in settings where transition dynamics are configurable and readable, runtime discovery complements offline training by grounding predictions in the active system instance. We propose enterprise discovery agents, which recover relevant transition dynamics at runtime by reading the system's configuration rather than relying solely on internalized representations. We introduce CascadeBench, a reasoning-focused benchmark for enterprise cascade prediction that adopts the evaluation methodology of World of Workflows on diverse synthetic environments, and use it together with deployment-shift evaluation to show that offline-trained world models can perform well in-distribution but degrade as dynamics change, whereas discovery-based agents are more robust under shift by grounding their predictions in the current instance. Our findings suggest that, in configurable enterprise environments, agents should not rely solely on fixed internalized dynamics, but should incorporate mechanisms for discovering relevant transition logic at runtime.

Community

Paper author Paper submitter

We study the question of enterprise models in enterprise software, where many transition rules are not hidden in the environment but stored explicitly as workflows, business rules, SLAs, schemas, and configuration records. We compare trained world models against discovery agents that inspect the live system at inference time to predict state changes.

The main finding: learned world models can look strong in-distribution, but they are brittle under enterprise shift. Discovery agents are more robust because they recover the relevant dynamics from the active system itself.

Would love to engage in researchers and people interested in world modeling and long-horizon planning under distribution shift.

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2605.12178
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2605.12178 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2605.12178 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2605.12178 in a Space README.md to link it from this page.

Collections including this paper 2