u/Careful-Crow9831

long-term memory and personality for any model
▲ 1 r/OpenAssistant+1 crossposts

long-term memory and personality for any model

Does anyone have experience with the team behind Anda DB? They claim to provide long-term memory and personality for any model. In my specific use case, I am looking for a tool or database that allows my local model to query the changelog history of my Laravel applications and other related data. here is the link https://anda.ai/#project-andadb

u/Careful-Crow9831 — 1 day ago

Zed + llamacpp + context overflow

Hello community, I have a problem with Qwen3.6 iq4 xs 35ba3b + Zed. My hardware: RTX 3090 24GB and 64GB DDR4 RAM. My model fits entirely in VRAM and works perfectly, but Zed occasionally throws an error saying the context limit has been exceeded. I don't understand why, because in Zed's config file the limit is set to 121k, and in llama.cpp the limit is 128k. The problem is that sometimes Zed sends 190k tokens or 140k tokens to llama.cpp in a single request. Why does this happen and how can I fix it?"

reddit.com
u/Careful-Crow9831 — 5 days ago
▲ 1 r/ollama

Qwen3.6 + Zed context problem

Hello community, I have a problem with Qwen3.6 iq4 xs 35ba3b + Zed. My hardware: RTX 3090 24GB and 64GB DDR4 RAM. My model fits entirely in VRAM and works perfectly, but Zed occasionally throws an error saying the context limit has been exceeded. I don't understand why, because in Zed's config file the limit is set to 121k, and in llama.cpp the limit is 128k. The problem is that sometimes Zed sends 190k tokens or 140k tokens to llama.cpp in a single request. Why does this happen and how can I fix it?

reddit.com
u/Careful-Crow9831 — 5 days ago