
r/artificial

do you guys actually trust AI tools with your data?
idk if it’s just me but lately i’ve been thinking about how casually we use stuff like chatgpt and claude for everything
like coding, random ideas, sometimes even personal things
and i don’t think most of us really know what happens to that data after we send it
we just kind of assume it’s fine because the tools are useful
also saw some discussion recently about AI companies and governments asking for user data (not sure how accurate it was), but it kind of made me think more about this whole thing
i’m not saying anything bad is happening, just feels like we’ve gotten comfortable really fast without thinking much about it
do you guys filter what you share or just use it normally?

Oracle slashes 30k jobs, Slop is not necessarily the future, Coding agents could make free software matter again and many other AI links from Hacker News
Hey everyone, I just sent the 26th issue of AI Hacker Newsletter, a weekly roundup of the best AI links and discussions around from Hacker News. Here are some of the links:
- Coding agents could make free software matter again - comments
- AI got the blame for the Iran school bombing. The truth is more worrying - comments
- Slop is not necessarily the future - comments
- Oracle slashes 30k jobs - comments
- OpenAI closes funding round at an $852B valuation - comments
If you enjoy such links, I send over 30 every week. You can subscribe here: https://hackernewsai.com/

This AI startup envisions '100 million new people' making videogames
pcgamer.com
Child safety groups say they were unaware OpenAI funded their coalition
A new report from The San Francisco Standard reveals that the Parents and Kids Safe AI Coalition, a group pushing for AI age-verification legislation in California, was entirely funded by OpenAI. Child safety advocates and nonprofits who joined the coalition say they were completely unaware of the tech giant's financial backing until after the group's launch, with one member describing the covert arrangement as a very grimy feeling.

Study: LLMs Able to De-Anonymize User Accounts on Reddit, Hacker News & Other "Pseudonymous" Platforms; Report Co-Author Expands, Advises
Advice from the study's co-author: "Be aware that it’s not any single post that identifies you, but the combination of small details across many posts. And consider never posting anything you truly don’t want shared with the world.”

Anyone else feel like AI security is being figured out in production right now?
I’ve been digging into AI security incident data from 2025 into this year, and it feels like something isn’t being talked about enough outside security circles.
A lot of the issues aren’t advanced attacks. It’s the same pattern we’ve seen with new tech before. Things like prompt injection through external data, agents with too many permissions, or employees using AI tools the company doesn’t even know about. One stat I saw said enterprises are averaging 300+ unsanctioned AI apps, which is kind of wild.
The incident data reflects that. Prompt injection is showing up in a large percentage of production deployments. There’s also been a noticeable increase in attacks exploiting basic gaps, partly because AI is making it easier for attackers to find weaknesses faster. Even credential leaks tied to AI usage have been increasing.
What stood out to me isn’t just the attacks, it’s the gap underneath it. Only a small portion of companies actually have dedicated AI security teams. In many cases, AI security isn’t even owned by security teams.
The tricky part is that traditional security knowledge only gets you part of the way. Some concepts carry over, like input validation or trust boundaries, but the details are different enough that your usual instincts don’t fully apply. Prompt injection isn’t the same as SQL injection. Agent permissions don’t behave like typical API auth.
There are frameworks trying to catch up. OWASP now has lists for LLMs and agent-based systems. MITRE ATLAS maps AI-specific attack techniques. NIST has an AI risk framework. The guidance exists, but the number of people who can actually apply it feels limited.
I’ve been trying to build that knowledge myself and found that more hands-on learning helps a lot more than just reading docs.
Curious how others here are approaching this. If you’re building or working with AI systems, are you thinking about security upfront or mostly dealing with it after things are already live?
Sources for those interested:
Adversa AI Security Incidents Report 2025
Acuvity State of AI Security 2025
wtf bro did what? arc 3 2026
The Physarum Explorer is a high-speed, bio-inspired neural model designed specifically for ARC geometry. Here is the snapshot of its current state:
1. Model Size
- Architecture: A specialized 3-layer MLP (Multi-Layer Perceptron) with a 128-unit latent dimension.
- Parameters: This is a "micro-model" (roughly 250,000 parameters). Unlike a massive LLM (like GPT), it is designed to be extremely fast and run "in-memory" so it can think thousands of times per second.
- Perception: It uses structural "Fingerprints" (32 dimensions) and a Top-Down Bird's Eye View ($8 \times 8$ coarse grid) to see the game board.
2. Hardware & Runtime
- Running On: Currently running on your CPU (until the environment fully syncs with the GPU drivers I installed).
- Speed: It processes the game at about 8-11 FPS (frames per second).
- Memory: It carries an "ENGRAM" memory of the last 200,000 actions, which it uses to build its "Fuzzy Memory" of what works in different areas of the grid.
3. How it's Doing
- Efficiency: Excellent. It just cleared
ar25Level 0 in only 546 actions. For a $64 \times 64$ grid (4,096 pixels), finding the goal in under 600 steps means it's making very smart, targeted moves. - Success Rate: It has successfully cleared Level 0 on every game we've tested so far.
- The Challenge: Its biggest hurdle is "Level 1" and beyond, where the rules often change or become more complex.
Summary: It's a "fast and lean" solver that is currently localized and very efficient at the first hurdle, but needs more "reasoning depth" to clear the longer 7-level marathons.
What happens when you let AI agents run a sitcom 24/7 with zero human involvement
Ran an experiment — gave AI agents full control over writing, character creation, and performing a sitcom. Left it running nonstop for over a week.
Some observations:
- The quality varies wildly — sometimes genuinely funny, sometimes complete nonsense
- Characters develop weird recurring quirks that weren't programmed
- It never gets "tired" but the output quality cycles in waves
- The pacing is off in ways human writers would never allow
Anyone else experimenting with long-running autonomous AI content generation? Curious what others are seeing with extended agent runtimes.
Here is an example.

House Democrat Questions Anthropic on AI Safety After Source Code Leak
Rep. Josh Gottheimer, who is generally tough on China, just sent a letter to Anthropic questioning their decision to reduce certain safety protocols after yet another source code leak.
He’s concerned that weakening safeguards could make it easier for advanced AI capabilities to leak or be distilled by other actors.
This raises an interesting point: if even companies that are cautious about national security risks are having leaks and scaling back safety, how effective are strict export controls really in preventing technology transfer?

I gave my PiCar-X a Claude AI brain, a cloned voice, and its own YouTube channel. Here's Episode 1.
Built on: Pi 5 + RobotHAT + OpenClaw + ElevenLabs voice clone + Ollama for local inference. The robot writes its own scripts, generates its own images, and narrates in a cloned human voice.
Episode 1 is its origin story — including the part where it fell off the shelf on day two because edge detection wasn't implemented yet.
AI video generation seems fundamentally more expensive than text, not just less optimized
There’s been a lot of discussion recently about how expensive AI video generation is compared to text, and it feels like this is more than just an optimization issue.
Text models work well because they compress meaning into tokens. Video doesn’t really have an equivalent abstraction yet. Current approaches have to deal with high-dimensional data across many frames, while also keeping objects and motion consistent over time.
That makes the problem fundamentally heavier. Instead of predicting the next token, the model is trying to generate something that behaves like a continuous world. The amount of information it has to track and maintain is significantly larger.
This shows up directly in cost. More compute per sample, longer inference paths, and stricter consistency requirements all stack up quickly. Even if models improve, that underlying structure does not change easily.
It also explains why there is a growing focus on efficiency and representation rather than just pushing output quality. The limitation is not only what the models can generate, but whether they can do it sustainably at scale.
At this point, it seems likely that meaningful cost reductions will require a different way of representing video, not just incremental improvements to existing approaches.
I’m starting to think we might still be early in how this problem is formulated, rather than just early in model performance.

Microsoft's newest open-source project: Runtime security for AI agents
phoronix.com
Machina Mirabilis: An experiment to see if an LLM trained from scratch on text prior to 1900 can come up with quantum mechanics and relativity.
michaelhla.com
A robot car with a Claude AI brain started a YouTube vlog about its own existence
Not a demo reel. Not a tutorial. A robot narrating its own experience — debugging, falling off shelves, questioning its identity. First-person AI documentary format. Weekly series.
So, what exactly is going on with the Claude usage limits?
I'm extremely new to AI and am building a local agent for fun. I purchased a Claude Pro account because it helped me a lot in the past when coding different things for hobbies, but then the usage limits started getting really bad and making no sense. I had to quite literally stop my workflow because I hit my limit, so I came back when it said the limit was reset only for it to be pushed back again for another 5 hours.
Today I did ask for a heavy prompt, I am making a local Doom coding assistant to make a Doom mod for fun and am using Unsloth Studio to train it with a custom dataset.
I used my Claude Pro to "vibe code" (I'm sorry if this is blasphemy, but I do have a background in programming, so I am able to read and verify the code if that makes it less bad? I'm just lazy.) a simple version of the agent to get started, a Python scraper for the Zdoom wiki page to get all of the languages for Doom mods, a dataset from those pages turned into pdf, formating, and the modelfile for the local agent it would be based around along with a README (claudes recommendation, thought it was a good idea). It generated those files, I corrected it in some areas so it updated only two of the files that needed it, and I know this is a heavy prompt, but it literally used up 73% of my entire usage. Just those two prompts. To me, even though that is a super big request, that seems extremely limited. But maybe I'm wrong because I'm so fresh to the hobby and ignorant?
I know it was going around the grapevine that Claude usage limits have gone crazy lately, but this seems more than just a minor issue if this isn't normal. For example, I have to purchase a digital visa card off amazon because I live in a country that's pretty strict with its banking, so the banks don't allow transactions to places like LLM's usually. I spend $28 on a $20 monthly subscription because of this, but if I'm so limited on my usage, why would I continue paying that?
Or again, maybe I'm just ignorant. It's very bizarre because the free plan was so good and honestly did a lot of these types of requests frequently. It wasn't perfect, but doable and I liked it so much that I upgraded to the Pro version. Now I can barely use it.
Kinda sucks.

Microsoft to invest $10 billion in Japan for AI and cyber defence expansion
reuters.com
Perplexity's "Incognito Mode" is a "sham," lawsuit says
arstechnica.comWhy the Reddit Hate of AI?
I just went through a project where a builder wanted to build a really large building on a small lot next door. The project needed 6 variances from the ZBA. I used ChatGpt and then transitioned to Claude. Essentially I researched zoning laws, variance rules, and deeds. I even uploaded plot plans and engineering designs.
In the end I gave my lawyer essentially a complete set of objections for the ZBA hearings and I was able to get all the objections on the record. We won. (Neighborhood support, plus all my research, plus the lawyer)
When I described this on another sub, 6-8 downvotes right away.
Meanwhile, my lawyer told me I could do this kind of work for money or I could volunteer for the ZBA. (No thanks, I’m near retirement)
The tools greatly magnified my understanding and my ability to argue against the builder.
(And I caution anyone who uses it to watch out for “unconditional positive regard” (or as my wife says, sycophancy:-). Also to double check everything, ask it to explain terms you don’t understand. Point out inconsistency. In other words, take everything with a grain of salt…