Local LLM for coding
Hey everyone, I just got started yesterday trying to set up a local LLM for coding tasks. I'm used to Claude Code since I use it for work, so I'm trying to get it set up for local use. I have docker model runner pulling the LLM's, but I've come across a few issues getting started:
First I tried Gemma4, but I got a ton of tool errors in Claude Code.
Next I tried Qwen3-coder-next, but docker refused to offload processing to my GPU. Overall it was unusable because it took too much time to process anything (I don't think I had enough memory).
After that I tried deepcoder, but for some reason it refused to write anything to my filesystem.
Querying the models directly through dockers chat agent (at least for Gemma4 and deepcoder) was a decent experience though.
I have a 5090 and 9800x3d with 32gb of ram. Which model should I be running in docker for claude code? Or am I going about this all wrong and should be using a different software stack altogether?
Appreciate any advice!