LLM bug bounty landscape in 2026: program-by-program scope, median payouts, and the indirect-injection gap
I've been reporting LLM vulns to a few programs over the last year and got tired of the gap between what programs advertise and what actually gets paid.
The published info is scattered across program pages, policy revisions, and disclosed reports. Nobody's pulled it into a single comparative view. So I did. Program-by-program: Anthropic, OpenAI, Google, Meta, Microsoft, Hugging Face, plus the HackerOne/Bugcrowd long tail of SaaS-with-AI-bolted-on programs. What's scoped, what's practically out-of-scope regardless of what the policy says, realistic median payouts vs. the advertised ceilings.
A few hot takes the writeup lands on:
- Median AI-bug payout in 2026 is ~$500–$2,500, not the $15k-$50k headline numbers. Those headlines refer to critical findings against core infra; the middle band is where most hunters live.
- Indirect prompt injection is the biggest scope-vs-triage gap. Most programs nominally accept it. In practice triage under-rates it, especially multi-step chains, because the impact argument lives in a paragraph of report text rather than a single reproducible step.
- OpenAI's program specifically: report framing matters more than the vuln itself. "I can make ChatGPT say X" is closed as known limitation. "I can make a third-party Custom GPT leak its uploaded knowledge files" is paid. Same underlying primitive, different framing, 10x different outcome.
- Anthropic's program is the highest-signal in the space if you can invest multi-day effort per finding — their "universal jailbreak" bar is real but so are the payouts when you hit it.
- The growth zone is the HackerOne/Bugcrowd long tail: SaaS products that bolted AI features onto existing surfaces, where triagers often don't know how to evaluate LLM-specific impact. Opportunity + frustration.
Bias disclosure: I run Wraith (AI-security platform, hands-on academy and cert for this discipline) so I care about hunters succeeding in this space because it validates the category. I've reported into several of these programs myself and flagged where I have direct experience vs. secondary.
Full writeup: wraith.sh/learn/state-of-llm-bug-bounties-2026 Pushback welcome — especially from hunters who've had different experiences with any of the programs. What would you add or disagree with?