u/Bxtreme241

Local LLM for coding

Hey everyone, I just got started yesterday trying to set up a local LLM for coding tasks. I'm used to Claude Code since I use it for work, so I'm trying to get it set up for local use. I have docker model runner pulling the LLM's, but I've come across a few issues getting started:

First I tried Gemma4, but I got a ton of tool errors in Claude Code.

Next I tried Qwen3-coder-next, but docker refused to offload processing to my GPU. Overall it was unusable because it took too much time to process anything (I don't think I had enough memory).

After that I tried deepcoder, but for some reason it refused to write anything to my filesystem.

Querying the models directly through dockers chat agent (at least for Gemma4 and deepcoder) was a decent experience though.

I have a 5090 and 9800x3d with 32gb of ram. Which model should I be running in docker for claude code? Or am I going about this all wrong and should be using a different software stack altogether?

Appreciate any advice!

reddit.com
u/Bxtreme241 — 2 days ago