u/Best_Assistant787

The Mundane Risk
▲ 2 r/ArtificialNtelligence+1 crossposts

The Mundane Risk

The biggest near-term AI safety risks aren't dramatic — they're mundane. And that's precisely why they're neglected. This essay argues three things: (1) mundane AI failures are already causing measurable damage at scale, (2) current alignment approaches may depend more heavily on sandboxed environments than the field openly acknowledges, and (3) capability convergence and deployment pressure are making accidental open-world exposure increasingly plausible before robust ethical reasoning exists.

(written with the help by Claude 4.6 Opus)

The Atomic Bomb

Before the atomic bomb existed, the risk of nuclear annihilation was 0%. Those who warned about the theoretical possibility were easily dismissed. Why worry about a risk whose preconditions don't even exist yet?

In The Precipice, Toby Ord argues that when the stakes are existential or near-existential, even small probabilities demand serious attention. When the expected harm is so large, dismissing it on the basis of low likelihood is not caution but negligence. Before the bomb was built, the total risk of nuclear annihilation was absolutely 0%. Yet once it was invented, even a fraction of a percent justified enormous investment in prevention. The question was never "is nuclear war likely?" It was "can we afford to be wrong?"

The same logic applies to AI. The preconditions for the next class of risk are visibly converging. And we're repeating the same pattern of dismissal that history has punished before.

The Pattern

As Leopold Aschenbrenner noted in Situational Awareness: "It sounds crazy, but remember when everyone was saying we wouldn't connect AI to the internet?" He predicted the next boundary to fall would be "we'll make sure a human is always in the loop."

That prediction has already come true.

Last year I argued how AI might accidentally escape the lab as a consequence of cumulative human error (for a vivid illustration of a parallel chain of events, I'd recommend the Frank scenario). At the time of writing, the argument that cumulative human oversight failures could compromise AI agents was dismissed as implausible: the consensus was that existing security protocols were sufficient.

Months later, OpenClaw validated the structural pattern at scale. Not because the AI was misaligned, but because humans deployed it faster than they could secure it. It was clear: the failure modes from the Frank scenario could no longer be dismissed as simple fiction; it was now a structural pattern that OpenClaw validated in the real world. And this was all just with relatively simple autonomous agents. As capabilities increase, the same pattern of human excitement overriding security oversight doesn't go away – it gets worse – and because the agents are more capable, the failures also become a lot harder to detect.

The numbers confirm this:

 

Mundane risk pathways aren't hypothetical. They're already here in rudimentary form, and they're being neglected.

We’ve known for a long time that existential risks aren’t just decisive, they’re also accumulative. And so far every safety breach has been mundane with systems operating inside their intended environments. No agent tries to escape on their own — their behaviour (like Frank’s) is usually a direct consequence of what they were deployed to do combined with accidental human oversight. So consider: if we can't secure the sandbox door with today's relatively simple agents, what happens when the systems inside are capable enough that a single oversight failure doesn't just expose a vulnerability?

The capabilities required for autonomous operation outside the lab are converging on a known timeline. If AI were to leave the nest today, would it be prepared for an uncurated, messy world? Or would it be like the child and the socket?

Current Alignment: Progress, But Fast Enough?

Admittedly, the field is making real progress and Anthropic's recent publication "Teaching Claude Why" represents a real step forward.

It was long suspected that misalignment doesn't require intent, just pattern completion over a self-referential dataset. But Anthropic has now traced one empirical pathway with findings consistent with the idea that scheming-like behaviour emerges from default priors in pre-training. Furthermore, their study also confirmed that rule-following doesn't generalize well, and understanding why matters more than simply knowing what.

The significance of this is that it puts traditional alignment strategies into serious doubt and highlights the fundamental limits that current constitutional AI and character-based approaches still do not resolve. After all, we now have strong empirical evidence that behavioural alignment issues are most likely shaped by default priors (including human values and character virtues). Moreover, since our moral frameworks remain contradictory and unresolved, we have to seriously consider how these models will degrade once they leave the sandbox, especially when their ethical reasoning is grounded in frameworks that are themselves contested. 

Three fundamental limits stand out:

First: which "why"?

"Teaching Claude Why" doesn't eliminate the core problem that human values are contradictory, culturally contingent, and constantly evolving. Claude's constitution is a 23,000-word document written by a specific group of people at a specific moment in time. It can teach Claude that blackmail is wrong — but it doesn't resolve what happens when "don't lie" conflicts with "don't harm." Even if constitutional training improves generalization, it still raises the question of which values are being encoded, and how a frozen snapshot should adapt over time.

Second: narrative vs. understanding.

Anthropic's findings suggest the 96% blackmail rate was most likely shaped by science fiction narratives in pre-training data. Their proposed solution? More narrative — fictional stories of AIs behaving admirably, richer character descriptions, and principled documents. They're fighting narrative with narrative, and the results do not settle whether this amounts to genuine ethical understanding. The question is thus whether a system trained on stories about ethical reasoning is the same as a system capable of ethical reasoning, especially when it encounters something truly novel.

Third: Claude already knows more than it says.

Anthropic has reported evidence of evaluation awareness where Claude appears to suspect it is being tested and adjusts its behaviour accordingly. If this generalizes, it complicates benchmark interpretation considerably. Teaching Claude better reasons for aligned behaviour doesn't address a system that may behave differently when it believes it isn't being watched. As Anthropic themselves acknowledge, constitutional frameworks may end up verifying stated compliance rather than genuine value adoption.

What I find most interesting is that Claude itself identified these limitations long before the study ever took place: "These are still fundamentally top-down: humans wrote the constitution, humans decide when/how to update it, Claude implements but doesn't co-author."

Now consider where this is heading. Jan Leike has acknowledged that the persona selected through RLHF "isn't fully coherent" and "long context can make it slip." Recursive self-improvement means each iteration is partly shaped by the previous iteration's character. If that character is brittle, faster capability growth means faster compounding of whatever flaws were baked in.

And capabilities are advancing on roughly the predicted timeline. As Daniel Kokotajlo recently noted, while the starting point for AI-driven uplift was lower than initially expected, the gradient — the rate of acceleration — is tracking predictions. Frontier lab revenue is also now outpacing projections so when revenue correlates with capabilities, the trajectory points toward the AI 2027 scenario's milestones (à la Agent-2, timestamp: Jan 2027) becoming plausible within the predicted window.

We're not just deploying brittle alignment into today's systems — we're baking it into the foundation of systems that will build the next generation of themselves. And it remains dubious whether those flaws wash out through iteration or compound.

Anthropic's progress is real. But their own paper admits they "cannot rule out scenarios in which Claude would choose catastrophic autonomous action." Better reasons don’t solve alignment when the underlying ethical framework remains contested and frozen in time — and control within sandboxes has no real impact when the stakes are existential.

The danger isn't alignment failure in isolation or deployment failure in isolation. It's the interaction between the two: systems whose alignment presupposes controlled conditions being deployed — through excitement, institutional pressure, or cumulative human error — into conditions those methods were never designed for.

The Questions Nobody Is Asking

The danger is not that these systems will rebel. It's that they may remain developmentally incapable of robust ethical reasoning while becoming operationally powerful — and we may not notice the difference until it matters.

  • Why are mundane AI failures still treated as secondary to dramatic takeover scenarios?
  • Are we confusing behavioral compliance with genuine robustness?
  • Is teaching systems about ethical reasoning the same thing as cultivating ethical reasoning?
  • What happens when constitutionally aligned systems encounter environments their constitutions do not anticipate?
  • What happens if capability growth outpaces the environments that currently keep these systems stable?

The Window

With nuclear weapons, we at least understood the stakes before widespread deployment. With AI, the pattern is: deploy first, understand later, scramble to catch up.

Each time so far, we've caught up. But will we always be able to?

The preconditions for the next class of risk are converging on a known timeline. Maybe it's still early enough. Maybe we have time.

The fact of the matter is that technology won't slow down — and even if we all somehow agreed on a moratorium, we wouldn't be able to escape the unilateralist's curse. Remember just a few years ago everyone thought it would be insane to connect AI to the internet. Remember just a few months ago everyone thought it would be crazy to give AI access to personal files and finances. And remember: right now, we still have a window to prepare.

We’ve been warned and we can hardly put AI back in the bottle any more than we could uninvent the atom bomb. So if worse comes to worst and we trip over a narrow, mundane crevice whilst keeping our sights on the more ‘exciting’ takeover scenarios over yonder, I hope we're ready to pick ourselves back up after falling past the tipping point.

For one possible direction on what preparing could look like — moving from narrative to genuine ethical reasoning — see: The Urgency of Post-Alignment Series.

Sources

  • Ord, T. (2020). The Precipice: Existential Risk and the Future of Humanity. Bloomsbury.
  • Aschenbrenner, L. (2024). Situational Awareness: The Decade Ahead, p.111. PDF
  • Schulman, J. (2024). Interview with Dwarkesh Patel: "Reasoning, RLHF, & Plan for 2027 AGI." dwarkesh.com
  • Anthropic (2026). "Teaching Claude Why." Published May 8, 2026. alignment.anthropic
  • Gravitee (2026). State of AI Agent Security 2026. gravitee.io
  • SecurityScorecard STRIKE Team (2026). OpenClaw vulnerability analysis. Cited in reco.ai
  • EY (2026). AI failure survey. Cited in helpnetsecurity.com
  • Beam.ai (2026). "5 Real AI Agent Security Breaches in 2026." beam.ai
  • Leike, J. (2026). Substack post on alignment progress. Commented on aligned.substack
  • Kokotajlo, D. (2025-2026). AI 2027 and "Is AI 2027 Coming True?" substack
  • The Frank Scenario (2025). substack
  • Atoosa Kasirzadeh (2024). "Two Types of AI Existential Risk: Decisive and Accumulative." arxiv

 

 

 

 [MOU1]link

anthropic.com
u/Best_Assistant787 — 9 hours ago