r/modelcontextprotocol

▲ 32 r/modelcontextprotocol+5 crossposts

I’ve been obsessed with Agentic Workflows lately, and I just found the "missing link" for anyone struggling with agent hallucinations and massive API bills.

It’s called King Context, and it’s an open-source framework that replaces messy vector searches with structured Corpus Engineering.

The GitHub Repo:https://github.com/deandevz/king-context

Why this is a complete paradigm shift:

  1. The "Corpus" Method: Instead of just "chunking" data, it synthesizes it into a specialized corpus. You can generate a corpus from any source (docs, web research, internal notes) and refine it. It’s like giving your agent a custom-built brain instead of a pile of random papers.
  2. Metadata-First Retrieval: It uses a tiered approach (metadata -> preview -> full read). This stopped my agents from "hallucinating" on missing context because they can verify if the information exists before they consume the tokens.
  3. Solving the Skill Bottleneck: By using "Skills" alongside a specialized Corpus, you can build multi-agent workflows where one agent acts as a researcher (building the corpus) and the other acts as an expert (executing with 100% facts).

The Numbers (Benchmarked against Context7):

  • Accuracy: 38/38 correct facts (100%) vs 32/38.
  • Hallucinations: ZERO (0.0) per query.
  • Efficiency: 3.2x fewer tokens per request.
  • Speed: Up to 170x faster metadata hits.

I’ve been talking to the dev (@deandevz), and the roadmap for Corpus Refinement (automatically pruning noisy data) is going to change how we build production-grade agents.

If you are tired of agents getting lost in large codebases or documentation, you need to check this out. It’s local-first, transparent, and built for the "Vibe Coding" era where context is everything.

Check it out here:https://github.com/deandevz/king-context

Would love to hear from anyone else trying to move away from traditional RAG. How are you handling context bloat?

u/VadeloSempai — 1 day ago
▲ 10 r/modelcontextprotocol+8 crossposts

I’ve been experimenting with making MCP tools feel more Unix-native

There are already some interesting projects around MCP tooling and conversion layers like mcporter and similar libraries.
While trying them, I realized what I personally missed wasn’t just “wrapping” MCP servers, but having an environment where:
MCP tools become normal CLIs
they work naturally with pipes/scripts/CI
agents can use them without loading huge schemas every session
and you can also create your own CLI tools directly from Python code
So I started building cli-use.

Example:

cli-use add fs /tmp
cli-use fs list_directory --path /tmp

After that the MCP server behaves like a regular Unix command:

cli-use fs search_files --path /tmp --pattern "*.md" | head

also added things like:
daemon mode for fast repeated calls
caching
shell completions
automatic SKILL.md generation for agents

One thing I found interesting is that reducing all the MCP protocol overhead ended up saving a pretty large amount of tokens during agent workflows.
Still experimenting with the idea, but I’m curious whether other people working with MCP also want a more shell-native / Unix-style approach to tools.

github.com
u/Just_Vugg_PolyMCP — 2 hours ago
▲ 5 r/modelcontextprotocol+3 crossposts

Every MCP server you add makes your agent slightly dumber. Here is what actually fixes it.

One thing I’ve started noticing with MCP-based agents is that performance degrades much earlier than most people expect, especially once the number of integrations becomes large.

Small setups work surprisingly well. A few integrations, a handful of tools, manageable schemas, and the agent behaves predictably. The problems usually begin once teams start connecting the systems they actually use in production. Slack, Gmail, GitHub, Linear, Notion, databases, deployment tooling, internal APIs, monitoring systems. The integration surface grows very quickly.

At that point, the issue stops being “model intelligence” and starts becoming a context management problem.

Most MCP servers expose many tools, and each tool brings descriptions, parameter schemas, examples, and edge cases into the prompt space. Individually this feels harmless, but collectively it creates a very noisy environment for the model to reason inside. The agent spends more effort understanding the tool ecosystem than solving the task itself.

You can partially reduce the problem with lazy loading or dynamic tool visibility, but those approaches still inherit the same scaling issue underneath. The total surface area keeps growing.

I recently came across this open-source project Corsair that takes a different approach, and I thought the design was genuinely interesting.

Instead of exposing hundreds of tools directly, it exposes four generic primitives:

  • setup and authentication
  • operation discovery
  • schema inspection
  • execution

The important detail is that schemas are fetched only when the agent decides it needs them. The model first discovers available operations, then inspects a specific schema on demand, and finally executes the workflow.

That keeps the tool surface effectively constant regardless of how many integrations exist underneath.

The design feels much closer to how humans interact with unfamiliar systems. You first discover what capabilities exist, then inspect the details you need, and only then perform the action. Most current MCP ecosystems invert this by front-loading the entire integration surface into context immediately.

I suspect a lot of current agent reliability issues are really interface design problems. As integration counts grow, the systems that scale will probably be the ones that minimize what the model has to hold in working memory at any given moment.

u/Arindam_200 — 2 days ago
▲ 2 r/modelcontextprotocol+1 crossposts

Hey everyone,

I've been building B2G (Business-to-Government) agents lately, and if you've ever tried to scrape government portals, you know the nightmare: malformed PDFs, captchas, and layouts that change every week.

My CrewAI agents were constantly breaking because of bad data input.

I decided to move the entire "dirty work" to a specialized infrastructure. I built an MCP (Model Context Protocol) Server that:

  1. Navigates the portals in the background.
  2. Uses Llama-3 (via Groq) to structure the messy PDF/HTML data into strictly typed JSON.
  3. Exposes everything to the agent via the new native MCPServerAdapter.

The result: The agent no longer "scrapes". It just asks for bidding opportunities in a city and gets a clean JSON back. Zero hallucinations on values or dates.

Architecture:

  • Backend: FastAPI + SQLite (for caching).
  • Tools: Custom MCP wrapper for Gov Data.
  • Orchestrator: CrewAI.

I’ve attached a video of the agent running. It found 3 cloud computing tenders in a Brazilian city and drafted a sales summary in seconds.

I’ve opened the public wrapper for the community to test. If anyone is building sales/prospecting agents and wants to play with this, let me know in the comments and I'll share the repo/template!

https://i.redd.it/we6yahvrq6zg1.gif

reddit.com
u/GrouchyGeologist2042 — 9 days ago