u/Large-Reporter-1746

🔥 Hot ▲ 598 r/Futurology

AI targeting systems have made war crimes structurally unaccountable

Israel's Lavender system assigned assassination scores to 37,000 people using mass surveillance data, communication patterns, social graphs, phone contacts. Human review per target: 20 seconds, solely to confirm the person's biological sex. Known error rate: 10%, meaning ~3,700 people with zero militant connection were marked for killing by design, not accident.

The US's Project Maven (now run by Palantir) compressed targeting timelines from 743 minutes to under 1 minute. In the Iran campaign launched February 2026, Maven's pipeline identified 15,000 targets in 10 days across 177 cities. 900 strikes in the first 12 hours. $5.6 billion in munitions in 48 hours. Impossible without AI.

Under the Rome Statute, individual criminal responsibility requires proving a specific person ordered a specific unlawful act. When an algorithm recommends, a commander batch-approves a queue, and an operator rubber-stamps in 20 seconds, that chain of individual intent collapses. No single human "decided" to kill those 3,700 civilians, the system did. Officers themselves described it: "Everything was automatic. I had zero added value as a human, apart from being a stamp of approval."

The ICRC has stated that lawfulness under IHL "cannot be assessed by a machine." The UN Special Rapporteur called for an immediate moratorium on autonomous targeting. Nothing happened. Instead, after the Iran campaign, Palantir stock surged 12.4% in a single week.

We are watching the field test of a new doctrine: that AI-assisted mass targeting is both militarily optimal and legally unprosecutable. If that conclusion holds, every future conflict will look like this.

reddit.com
u/Large-Reporter-1746 — 5 hours ago