I often need to deal with tasks requires high context usage. Having GPT plus + opencode go sub, the only 1M context model I have is deepseek v4, which I find not very reliable.
And the context compression is really a disaster. The critical instructions and info are often lost. I would rather ask the agent the summerize this session before run out of context space and start a new one by reading the summarized txt, which works but is still annoying.
I've tried the hermes-lcm. It didn't improve much. Only the first 3 message of the session is not compressed, meaning if the agent want to find any following instructions after the first 3 message, it needs to go a long way searching through the DAG.
I also see the context-mode which seems to be an amazing project but hermes is not compatible to that :(
Any suggestion & recommendation?