Getting a lot of garbage results with Qwen3.6-27B :(
I'm running Qwen3.6-27B with vLLM at FP16. There are a few known issues with the chat template (I think), and I do get occasional stop in OpenCode or other harnesses.
But in OpenWebUI is 100x worse. The model stops, sometimes gives me garbage words in a loop and other times fails tool calls due to bad json. It's a 50% chance to actually manage to use it or not.
I don't get it, I'm using the default values and yes, Native tool calls. In vLLM I'm using the recommended params.
What else can I try?
u/nunodonato — 23 hours ago