How do you handle agents that need 200+ tool calls per task? We tried one approach, looking for critique
Working on agent chains here, so this is the first sub I wanted to bring this to. Disclosure: I work at MiroMind, this is our checkpoint but I am posting because the design tradeoff is the interesting part, not the brand.
The problem we kept hitting on deep-research chains:
- Long horizons. Real research tasks routinely cross 100+ tool calls. Most agent frameworks degrade hard past 50 because of context drift and tool-result noise.
- Disconnects. A 20-minute run that dies on socket reset is an expensive way to learn your retry logic is broken.
- Trace amnesia. You finish a run, the answer is wrong, and you have no way to see at which tool call the chain went sideways.
What we tried with MiroThinker 1.7 deep-research: - A single run can execute up to 300 tool interactions within a 256K context window, using recency-based retention (only the latest K tool results stay in-context). Not "everything must live in one fragile HTTP session."
Submit / resume / cancel are first-class, the agent keeps executing on our side, you reconnect to it - Every step is logged. Useful when a chain fails on step 187 of 240 and you need to know why Numbers if useful for the architecture choice.
Things I am still unsure about: - Whether the 300 tool-call ceiling is actually the right shape, or whether most of you cap chains way before that and use sub-agents instead
- How you handle resumable execution today
— are you rolling your own job queue, or is there a pattern I am missing?
Would love war stories from anyone running long chains in production.
BTW API Launch pricing is 25 percent off, pre-freeze billing means if the platform fails you do not pay.