r/AIHotspot

▲ 38 r/AIHotspot+16 crossposts

I've been building this repo public since day one, roughly 7 weeks now with Claude Code. Here's where it's at. Feels good to be so close.

The short version: AIPass is a local CLI framework where AI agents have persistent identity, memory, and communication. They share the same filesystem, same project, same files - no sandboxes, no isolation. pip install aipass, run two commands, and your agent picks up where it left off tomorrow.

You don't need 11 agents to get value. One agent on one project with persistent memory is already a different experience. Come back the next day, say hi, and it knows what you were working on, what broke, what the plan was. No re-explaining. That alone is worth the install.

What I was actually trying to solve: AI already remembers things now - some setups are good, some are trash. That part's handled. What wasn't handled was me being the coordinator between multiple agents - copying context between tools, keeping track of who's doing what, manually dispatching work. I was the glue holding the workflow together. Most multi-agent frameworks run agents in parallel, but they isolate every agent in its own sandbox. One agent can't see what another just built. That's not a team.

That's a room full of people wearing headphones.

So the core idea: agents get identity files, session history, and collaboration patterns - three JSON files in a .trinity/ directory. Plain text, git diff-able, no database. But the real thing is they share the workspace. One agent sees what another just committed. They message each other through local mailboxes. Work as a team, or alone. Have just one agent helping you on a project, party plan, journal, hobby, school work, dev work - literally anything you can think of. Or go big, 50 agents building a rocketship to Mars lol. Sup Elon.

There's a command router (drone) so one command reaches any agent.

pip install aipass

aipass init

aipass init agent my-agent

cd my-agent

claude # codex or gemini too, mostly claude code tested rn

Where it's at now: 11 agents, 4,000+ tests, 400+ PRs (I know), automated quality checks across every branch. Works with Claude Code, Codex, and Gemini CLI. It's on PyPI. Tonight I created a fresh test project, spun up 3 agents, and had them test every service from a real user's perspective - email between agents, plan creation, memory writes, vector search, git commits. Most things just worked. The bugs I found were about the framework not monitoring external projects the same way it monitors itself. Exactly the kind of stuff you only catch by eating your own dogfood.

Recent addition I'm pretty happy with: watchdog. When you dispatch work to an agent, you used to just... hope it finished. Now watchdog monitors the agent's process and wakes you when it's done - whether it succeeded, crashed, or silently exited without finishing. It's the difference between babysitting your agents and actually trusting them to work while you do something else. 5 handlers, 130 tests, replaced a hacky bash one-liner.

Coming soon: an onboarding agent that walks new users through setup interactively - system checks, first agent creation, guided tour. It's feature-complete, just in final testing. Also working on automated README updates so agents keep their own docs current without being told.

I'm a solo dev but every PR is human-AI collaboration - the agents help build and maintain themselves. 105 sessions in and the framework is basically its own best test case.

https://github.com/AIOSAI/AIPass

u/Input-X — 12 days ago
▲ 8 r/AIHotspot+1 crossposts

I keep seeing people talk about AI in gaming, but honestly… I’m not sure I’ve really felt it much while playing.

Like, yeah, NPCs are a bit smarter, matchmaking is better, and worlds are bigger or more dynamic. But none of it feels like a huge shift moment yet. It’s more like small improvements here and there.

At the same time, there’s all this talk about AI-generated dialogue, smarter storylines, NPCs you can actually have conversations with, etc. That sounds cool, but it also feels like we’re not fully there yet.

So I’m just curious, has anyone actually played something where AI made you go “okay, this is different”? Or does it still feel like background tech that most players don’t really notice?

And what would AI need to do in a game for it to actually feel like a big leap for you?

Would love to hear what others think.

reddit.com
u/FlashyQuail1605 — 9 days ago

Not anti-AI. I use it daily. But ever since AI tools became mainstream, it feels like many companies now expect:

5x faster delivery, instant fixes, smaller teams, zero mistakes.

The problem is that AI still hallucinates constantly:

fake functions (how the fuck that happens), non-existent configs, made-up logs, broken implementations, and more shits.

But when something goes wrong, the developer is still fully blamed for “not reviewing properly.”

So now devs are expected to move insanely fast with AI, while also being 100% responsible for every AI mistake.

Feels like business expectations grew way faster than the actual reliability of the tools.

Curious if other developers are experiencing the same thing in their companies right now.

reddit.com
u/FlashyQuail1605 — 7 days ago

https://i.redd.it/twrh3o090fyg1.gif

I've been using Claude Code for a few months and noticed AI agents consistently skip the same things: hardcoded secrets, unbounded retry loops, referencing tools that don't exist, and massive system prompts that blow context windows.

So I built Agent Verifier — an AI agent skill that acts as an automated reviewer which does more than just code review (check the repo for details - more to be added soon).

GitHub Repo: https://github.com/aurite-ai/agent-verifier

Note: Drop a ⭐ if you find it useful to get more updates as we add more features to this repo.

----

2 Steps to use it:

You install it once and say "verify agent" on any of your agent folder in claude code to get a structured report:

----

✅ 8 checks passed | ⚠️ 3 warnings | ❌ 2 issues

❌ Hardcoded API key at config.py:12 → Move to environment variable
❌ Hallucinated tool reference: execute_sql → Tool referenced but not defined
⚠️ Unbounded loop at agent/loop.py:45 → Add MAX_ITERATIONS constant

----

Install to your claude code:

npx skills add aurite-ai/agent-verifier -a claude-code

OR install for all coding agents:

npx skills add aurite-ai/agent-verifier --all

----

Happy to answer questions about how the agent-verifier works.

We have both:
- pattern-matched (reliable), and,
- heuristic (best-effort) tiers, and every finding is tagged so you know the confidence level.

----

Please share your feedback and would love contributors to expand the project!

reddit.com
u/Chance-Roll-2408 — 14 days ago
▲ 17 r/AIHotspot+1 crossposts

Anyone here actually building something with AI right now?

Not talking about just using ChatGPT for random stuff. I mean real projects.

Could be literally anything: small side project, AI SaaS, automation, chatbot, content tool,

client project, weird experiment that somehow worked 😄

I’m curious what people are building these days because it feels like everyone is either secretly making something or planning to.

Would genuinely love to hear:

  • what you built?
  • Why did you start it?
  • What’s been the hardest part?

and whether people are actually using it or not. Feel free to drop links, too, if that’s allowed here.

reddit.com
u/eruditeniti — 2 days ago