I tested 200+ prompts over 6 months. Here are the 7 patterns that actually move the needle (with examples)
I've been obsessively benchmarking prompt structures across Claude, GPT-4, and Gemini for a client project. Not vibes — actual A/B evals with human raters. Here's what separates prompts that kind of work from ones that are embarrassingly good.
1. Persona + Constraint stacking
Most people assign a persona. Almost nobody adds constraints on top of the persona. The combo is where the magic happens.
You are a senior systems engineer who has been burned by vague requirements three times this quarter. Review this spec and flag anything that would cause ambiguity during implementation. Be specific, be ruthless, and skip anything obvious.
2. The "Anti-example" trick
Showing what you don't want outperforms describing what you do want by ~40% in my evals. Brains (and models) pattern-match on contrast.
Write a product description for this blender.
NOT like this:
"Experience the revolutionary power of BlendMaster Pro — your ultimate kitchen companion for crafting delicious smoothies!"
Like this: [your actual good example]
3. Role reversal as a QA tool
After getting an output, immediately prompt: "What are the 3 weakest assumptions in your response above?" — the model will catch things your initial prompt didn't even think to ask about. This alone saved my team hours of review.
4. Format as a cognitive scaffold
Don't just say "be concise". Specify the cognitive structure you want. There's a huge difference between:
- "Answer briefly" → vague, ignored
- "Answer in: one sentence conclusion, then 3 bullet supporting points, no fluff" → model now has a scaffold to fill
5. Emotional priming (yes, really)
Adding "This is important to get right" or "Take your time with this" measurably improves output quality on complex tasks. It sounds silly but it works — probably because these phrases appear before high-quality human writing in training data.
6. Chain-of-thought with a twist — ask for uncertainty
Standard CoT: "Think step by step."
Better: "Think step by step. At each step, rate your confidence 1-5 and flag if you're guessing."
You get the reasoning AND a map of where hallucinations are most likely hiding.
7. The "Steelman first" pattern for critical tasks
Before asking the model to critique anything, make it argue for the thing first. You get a more balanced critique that doesn't just perform skepticism.
First, make the strongest possible case FOR this business idea. Then, with that context in mind, identify its most serious flaws.