r/CLine

▲ 5 r/CLine

Why does Cline recommend Claude when using GLM 5.1

Guys, I’m trying to understand something about Cline.

When I use GLM 5.1 in Cline, I sometimes receive an error or a recommendation suggesting that I should use Claude instead. It gives me the option to ignore the warning, but I want to understand what it actually means.

Does this mean the model is not strong enough for the task, or could there be an issue with Cline itself?

I’m asking because I never received messages like this in Roo Code, even on heavy tasks. I’m new to Cline and recently switched from Roo Code, so I’m trying to understand how it works.

Thanks for your help.

reddit.com
u/Comfortable-Mix-7805 — 3 days ago
▲ 6 r/CLine

Running Cline/OpenHands on-prem with 4×RTX 3090: 30B vs 70–80B, vLLM/SGLang, SaaS cost reduction?

I’m evaluating an on-prem coding-agent setup to optimize Claude/GPT SaaS/API spend.

Hardware target: 4×RTX 3090, 96GB total VRAM, Linux, likely vLLM or SGLang exposing an OpenAI-compatible endpoint.

Tools: Cline and OpenHands.

Questions:

  1. Is Qwen3-Coder-30B-A3B enough as a daily driver, or is it underusing the hardware?
  2. Has anyone run Qwen3-Coder-Next 80B-A3B, Llama/Qwen 70B-class, or similar models on 4×3090 for coding agents?
  3. What quantization actually works well for tool-use and long-horizon coding tasks: FP8, AWQ, GPTQ, Q8, Q4?
  4. What context length is realistic before throughput collapses?
  5. Has this meaningfully reduced your Claude/OpenAI spend, or do you still need cloud fallback for hard tasks?

I’m especially interested in real-world results: tokens/s, accepted PRs/tasks, failure modes, model configs, and whether OpenHands/Cline behave reliably with local endpoints.

reddit.com
u/EmbarrassedBeach1069 — 2 days ago
▲ 76 r/CLine

Hello from the Cline team 👋

We've been heads-down on a ground-up rewrite of our extension and CLI for the past couple months, and we need help beta testing to bring it over the finish line.

Why we did this

The original Cline shipped right after Claude 3.5 Sonnet dropped. It was the first coding agent that actually worked, and we built it extension-first for the IDE. Great in its time, but the architecture got tightly coupled to IDE semantics, which made it painful to evolve the harness for the CLI and to extend into things like flexible agent profiles and agent teams.

So we took every hard lesson from the past year and started fresh. We built an SDK with better performance and token efficiency, then rebuilt the CLI and extension on top of it. The SDK has a plugin architecture for providers, models, LSP servers, code search, themes, all of it. The goal is for Cline to be custom-fitted to how you work.

What we need from you

We're at the last mile and want folks in the beta with us. There will be breaking changes and missing features, and we'd like your help to close the gaps.

To help kick things off:

  • $20 in credits to get started
  • A bounty program rewarding contributors who help us fix bugs and ship plugins

Drop into the cline discord's #beta channel to get set up, share feedback, and build alongside us.

u/saoudriz — 13 days ago
▲ 22 r/CLine+6 crossposts

Everyone's facing insane costs and rate limits from Claude Code it's gotten ridiculous these last few months. I needed a better alternative to save my money, so I found Cline, started clining , and it was amazing. But I kept thinking: imagine bringing Cline as a provider into Claude Code mature environment to test... and it rocked I combined Cline cost/performance models with Claude Code ecosystem into one product, handling cache control, API calls, ToS schemas, and building req/response to fit perfectly. Now I've got 13 providers and Cline one of my faves you guys gotta try it: https://github.com/AbdoKnbGit/tau

u/JhonDoe191ee — 6 days ago
▲ 14 r/CLine

They say on their page it is 1.6 billion credit and mimo v2.5 pro takes 2 credit per token, mimo v2.5 takes 1 credit per token but here is how they get you, cached token is still billed the same credit per round trip, absolutely not suitable for coding cli then, because every single one of them by design would keep going back and forth with toolcalls, that's how they work, normally inference providers charge 10% for the pre existing cached context, but Xiaomi takes the full amount, I did 10 small tasks like not even that deep, small tasks and it is already at 12 or so million credit used, it used probably under a million context tasks were that mini, like saying hello, and mv this folder around, write some sql etc, like 10 total prompts same session, credit cost keeps snow balling, they don't mention nothing of this sort in the token plan docs or anything anywhere, for a big task it would be what 200 million token uncached, so 400million credit if you used mimo v2.5 pro, so with max 100$ plan you can use it for 4 tasks PER MONTH, honestly get anything over mimo token/coding plan, 40m token task(input+output) would be like 400million, cache hit rate is avg 90%

reddit.com
u/FearlessGround3155 — 10 days ago