Architecture discussion: The missing infrastructure for continuously running AI Agents
From an engineering perspective, the current AI agent stack feels incomplete. We have frameworks (LangChain), execution runtimes (sandboxes/Browserbase), and harnesses (DeepAgents/Claude Code). But they all share a fundamental flaw for long-running systems: they are trigger-based.
If you are tasked with building an agent that operates continuously and sustainably on its own, an Agent Harness isn't enough. What we actually need is a dedicated Agent Runtime Environment.
To clarify, I'm not talking about an Agent Execution Runtime Env (where the agent safely executes Python). I'm talking about the persistent daemon/supervisor layer—the environment that gives the agent a continuous lifecycle, manages its state, handles self-healing when the LLM inevitably hallucinates a crash, and provides a heartbeat for proactive background work.
How are you all architecting this? Are you just wrapping your agents in Kubernetes cronjobs and temporal workflows, or is there a better pattern emerging for true persistent agent environments?