u/nunodonato

▲ 13 r/Vllm+1 crossposts

Getting a lot of garbage results with Qwen3.6-27B :(

I'm running Qwen3.6-27B with vLLM at FP16. There are a few known issues with the chat template (I think), and I do get occasional stop in OpenCode or other harnesses.

But in OpenWebUI is 100x worse. The model stops, sometimes gives me garbage words in a loop and other times fails tool calls due to bad json. It's a 50% chance to actually manage to use it or not.

I don't get it, I'm using the default values and yes, Native tool calls. In vLLM I'm using the recommended params.

What else can I try?

reddit.com
u/nunodonato — 23 hours ago