u/DaMoot

Memory activation layer/framework?

I'm really quite frustrated with the current state of memory retrieval affairs for pretty much everything AI-related. Not memory storage, but memory activation / retrieval.

The disconnect between 'I know this thing' and actively recognizing that it is in memory, no matter if it is a .md or .jsonl file, or in a PostgreSQL vector database. The “I recognize this situation because something already in memory is lighting up.”

All it takes is a session reset between interactions for the agent to go 'huh?' when you respond to something that happened before the reset as there is no per-orientation context. Several times I've come back between minutes and hours later to respond to a question from the agent and it all but goes 'Whatchu talkin' bout, Willis?' until I prompt it to look in memory, where it usually finds the previous session and we can continue.

Before I go about reinventing the wheel to try and create this memory activation layer naggin' in my noggin, I wanted to see who might be using these or other existing projects with Hermes Agent that do similar.

  • Meta-memory
  • Mem0
  • LangMem
  • Letta
  • Signet

Anything similar like BMAM, or projects that use Associative Spreading Activation theory.

reddit.com
u/DaMoot — 2 days ago

Disclaimer; I'm not a super smart AI researcher or expert computer guru. I prompted the framework to fix itself because I complained it was slow and was curious why my agent was spending so much time prompting me that it was just waiting. Not doing anything I could see, just waiting for... my electric meter to make another turn, the meaning of life, who knows.

First, here's the PR so you can check it out directly; https://github.com/NousResearch/hermes-agent/pull/18547 One fix is a duplicate of a 3 week old PR and the other appears unique.

Short-short, a bad system prompt and a bad timestamp application causes large amounts of wasted re-compute. For my V100 32GB LS w/ ~820GB/s memory bandwidth it adds up to several minutes in some tasks.

This is a compound compute + memory bandwidth issue that makes stuff slower than it should be. My understanding of the issue is that this will affect even things like the DGX Spark or Mac Mini/Studio.

I expressed frustration to Claude that this was all so much of a black box and it said 'Oh, well turn on verbose logging for llama.cpp and now the black box becomes largely transparent. Upload the log to me and consume your entire week's worth of Pro usage in under an hour.'

u/DaMoot — 13 days ago