Most people treat prompts the way they treat email subject lines — something to dash off and hope for the best. They write "summarize this" and wonder why the output is shallow. They add "be concise" and wonder why the model still rambles. They try again with different words. No system. No measurement. No feedback loop.
That is not engineering. That is guessing.
In the last three years, four serious research artifacts have systematically documented what actually moves the needle: a landmark Google Brain paper on chain-of-thought reasoning, a 58-technique survey from a team at the University of Maryland, Anthropic's internal guidance for Claude deployments, and Google's 65-page practitioner whitepaper. Together, they contain more empirical data about prompt behavior than most practitioners have ever read.
This article synthesizes all four. You will walk away with the specific numbers, the specific techniques, and a complete, copy-ready system prompt built from first principles.
Why "Better Wording" Is Not a Strategy
The core issue is that people treat prompt iteration as aesthetic revision — swap a word, see if it sounds better. This collapses the feedback loop to a sample size of one and mistakes style for structure.
Real prompt engineering has more in common with writing unit tests than with copywriting. You define the expected output, identify the failure mode, isolate the variable that produced it, change one thing, and measure again. The model is the system under test. The prompt is the test harness.
The four research documents covered here treat prompting as an engineering discipline with measurable outcomes. That framing alone is the most important thing to internalize before reading any individual technique.
Chain-of-Thought: The First Proof That Structure Beats Instruction
In January 2022, Jason Wei and colleagues at Google Brain published Chain-of-Thought Prompting Elicits Reasoning in Large Language Models (arXiv:2201.11903). It is the paper that established, with controlled empirical evidence, that how you structure a prompt changes model behavior in measurable, reproducible ways.
What chain-of-thought is
The idea is simple: instead of giving the model an input and asking for an output directly, you provide worked examples that show intermediate reasoning steps. The model, seeing those examples, applies the same reasoning pattern to new problems.
Standard few-shot prompting looks like this:
Q: Roger has 5 tennis balls. He buys 2 more cans, each with 3 balls.
How many does he have?
A: 11.
Q: [new question]
A:
Chain-of-thought looks like this:
Q: Roger has 5 tennis balls. He buys 2 more cans, each with 3 balls.
How many does he have?
A: Roger started with 5 balls. 2 cans × 3 balls = 6 balls.
5 + 6 = 11. The answer is 11.
Q: [new question]
A:
The model sees reasoning demonstrated, not just answers. When it encounters a new problem, it reproduces the reasoning pattern before arriving at an answer.
The numbers
The paper tested eight reasoning benchmarks across three categories: arithmetic (GSM8K, SVAMP, AQuA), commonsense reasoning (CommonsenseQA, StrategyQA), and symbolic reasoning (Letter Concatenation, Coin Flip). Chain-of-thought outperformed standard prompting across all of them at sufficient model scale.
The critical constraint: scale
This is the finding most practitioners miss. Chain-of-thought is an emergent capability. It only works reliably in models with roughly 100 billion or more parameters. In smaller models — below approximately 10 billion parameters — CoT prompting can actually decrease performance. The model does not have enough capacity to reliably generate correct intermediate steps, and wrong steps compound into wrong answers.
Zero-shot CoT: "Let's think step by step"
Subsequent work by Kojima et al. (2022) showed that simply appending "Let's think step by step" to a prompt — with no examples at all — produces similar reasoning behavior in large models. This zero-shot variant delivers roughly a 20% accuracy improvement on arithmetic benchmarks over a zero-shot baseline.
This is the single highest-leverage, lowest-effort technique in the literature. Four words. Measurable gain. No examples required.
The 58-Technique Survey: A Field Map of What Works
In June 2024, Schulhoff et al. published The Prompt Report: A Systematic Survey of Prompting Techniques (arXiv:2406.06608) — the most comprehensive taxonomy of prompting that exists. It catalogues 58 text-based techniques across six categories, with empirical notes on when each technique outperforms baselines.
The paper's meta-finding is honest: no single technique dominates all tasks. Effectiveness is task-dependent. But several techniques show consistent gains across domains, and they are worth knowing in detail.
Category 1: Zero-Shot Techniques
Zero-shot techniques require no examples. They modify how you frame the question or role-assign the model.
- Role Prompting — Assign the model a domain-specific persona. "You are an expert revenue operations analyst with 10 years of B2B SaaS experience." Consistent quality improvements, especially for technical tasks where domain framing matters.
- Zero-Shot CoT — Append "Let's think step by step." ~+20% on arithmetic over zero-shot baseline.
- Rephrase and Respond (RaR) — Instruct the model to rephrase the question before answering. Reduces ambiguity and catches misinterpretation before it propagates into the answer.
- Emotion Prompting — Adding stakes language like "This is very important to my career" (Li et al., 2023). The paper documents measurable benchmark improvements and flags this as counterintuitive — the model appears to weight context around importance.
Category 2: Few-Shot Techniques
- Standard Few-Shot — 3–8 input/output exemplars. Consistent +10–30% over zero-shot on classification and reasoning tasks across virtually all benchmarks tested. The single most universal improvement available.
- KNN Example Selection — Dynamically select exemplars most similar to the current query rather than using fixed examples. Outperforms random example selection, especially for diverse input distributions.
- Self-Generated ICL (SG-ICL) — Ask the model to generate its own examples before answering. Useful when labeled examples are unavailable.
Category 3: Chain-of-Thought Variants
- Contrastive CoT — Include both correct and incorrect reasoning chains as exemplars, labeled explicitly. Shown to outperform standard CoT by helping the model learn what failure looks like.
- Analogical Prompting — Ask the model to recall analogous problems before solving the target. Activates relevant prior knowledge.
- Tab-CoT — Structure reasoning steps as a table rather than prose. Improves precision on multi-column or comparative reasoning tasks.
Category 4: Decomposition Techniques
- Least-to-Most Prompting — Break a complex problem into simpler subproblems and solve them sequentially, feeding each answer forward. Reported +16% over standard CoT on the SCAN compositional generalization benchmark.
- Plan-and-Solve — Explicitly ask the model to create a plan before executing it. Outperforms Zero-Shot CoT on 6 of 8 arithmetic benchmarks tested. Especially effective for tasks with multiple interdependent steps.
Category 5: Ensembling Techniques
This category contains the single highest-ROI technique in the entire survey.
Self-consistency works by sampling the model 20–40 times with the same prompt (using a non-zero temperature so outputs vary), then taking the majority vote among the final answers. The intuition: most reasoning paths that arrive at the correct answer will agree with each other; noise paths diverge.
This is not viable for every use case — it multiplies API costs by N samples — but for high-stakes, single-call decisions where accuracy matters more than latency, it is the most documented improvement available from a prompting change alone.
Summary: High-impact techniques by gain
| Technique | Gain | Best For | Source |
|---|---|---|---|
| Chain-of-Thought (few-shot) | +41pp GSM8K | Math, multi-step reasoning | Wei et al. 2022 |
| Self-Consistency (majority vote) | +17.9pp GSM8K | Any reasoning task, high-stakes | Wei et al. / Prompt Report |
| Standard Few-Shot | +10–30% | Classification, extraction, generation | All four papers |
| Zero-Shot CoT ("think step by step") | ~+20% | Arithmetic, when no examples available | Kojima 2022 / Prompt Report |
| Least-to-Most Prompting | +16% SCAN | Compositional, multi-dependency tasks | Prompt Report |
| Role Prompting | Consistent, varies | Domain-specific quality, tone | Prompt Report / Anthropic / Google |
Anthropic's Guide: Engineering for Claude Specifically
Anthropic's internal prompting guide is the most Claude-specific of the four documents. Where Wei et al. and the Prompt Report measure technique effectiveness across models and benchmarks, Anthropic's guidance is about production reliability — how to build prompts that behave predictably in deployed systems.
XML tags: the most underused technique in Claude deployments
Anthropic explicitly recommends wrapping distinct prompt components in XML tags. Claude was trained on XML-structured data and parses tag-delimited sections more reliably than prose separators, headers, or newlines.
<role>
You are an expert revenue operations analyst with 10 years of B2B
SaaS experience. You communicate in precise, data-driven language
and always quantify claims.
</role>
<context>
The company runs a 6-stage pipeline. Stage 3 stalls are the primary
revenue leak. Average deal size: $48K ARR.
</context>
<task>
Analyze the pipeline data below and identify the top 3 reasons deals
are stalling in Stage 3. For each: name the pattern, cite the
evidence, and recommend one corrective action.
</task>
<data>
{{PIPELINE_DATA}}
</data>
<output_format>
Return a JSON array with objects containing:
{ "pattern": str, "evidence": str, "action": str, "confidence": 0-1 }
</output_format>
The rationale: when your prompt contains both instructions and data, prose separation is ambiguous. The model cannot always distinguish "these are my instructions" from "this is the data I want processed." XML tags eliminate that ambiguity at the structural level.
The Clarity Principle
"A prompt should be so clear that a thoughtful human new to the task could follow it and produce the expected output."
— Anthropic Prompting GuideThis is the most actionable single sentence in all four documents. If you hand your prompt to a colleague who has never seen the task, and they are confused about what you want, the model will be too. Rewrite until the human can follow it without questions.
System prompt vs. human turn: a critical architecture decision
Anthropic makes a structural distinction that most practitioners ignore. Persistent behavioral instructions — the role, the constraints, the output format, the rules — belong in the system prompt. Task-specific data — the document to analyze, the question to answer, the inputs to process — belongs in the human turn.
Mixing them degrades reliability. The model optimizes its behavior differently depending on where instructions appear. System prompt = always-on governance. Human turn = per-request input.
Prefilling: forcing output format without fighting the model
Claude supports a technique unique to its API: you can pre-fill the start of the assistant's response to constrain output format. If you want JSON, start the assistant's turn with {". The model completes from that point, skipping the prose preamble it would otherwise add.
messages: [
{ role: "user", content: "Analyze the pipeline data..." },
{ role: "assistant", content: '{"patterns": [' } // prefill
]
Extended thinking
For Claude 3.5+ and Claude 4 models, Anthropic's extended thinking feature allows the model to reason internally before producing a final answer — similar to CoT but happening in a private scratchpad. For high-complexity tasks, this is the most effective reasoning enhancement available on Claude today.
Google's Framework: Configuration as Engineering
Google's 65-page prompt engineering whitepaper (v4, 2025) takes the practitioner's perspective: less benchmark analysis, more operational guidance. Its most valuable contribution is specific, quantified recommendations for model configuration — the settings that most practitioners leave at defaults.
Temperature: stop guessing, start specifying
Temperature controls randomness in token selection. Google's whitepaper provides concrete values tied to task type:
Temperature 0.0 → Factual Q&A, classification, data extraction,
code generation. Reproducibility over variation.
Temperature 0.2–0.5 → Summarization, analysis, translation.
Moderate creativity, high reliability.
Temperature 0.7–1.0 → Creative writing, brainstorming, marketing.
Variation is the point.
Temperature > 1.0 → Avoid in production. Outputs degrade
toward incoherence.
Pairing with nucleus sampling: Top-P 0.95 as a default for most tasks, Top-K 40 for balanced diversity. These are not magical numbers — they are documented starting points that Google's engineers have calibrated across their model family.
Google's six operating principles
- Be specific and detailed. Vague prompts produce vague outputs. Every ambiguous term in your prompt is a variable the model will resolve however it chooses.
- Provide examples whenever possible. Few-shot is almost always better than zero-shot. If you cannot provide examples, explain why the task is hard.
- Specify the output format explicitly. Do not leave format to model defaults. Specify JSON, markdown, bullet points, or prose — and if JSON, specify the schema.
- Iterate systematically. Treat prompts as code. Version them. Log the change. Measure the delta. Do not iterate by feel.
- Use system instructions for persistent behavior. Per-request data goes in the user turn. Behavioral constraints go in system.
- Define success criteria in the prompt. For evaluation tasks, tell the model what a good answer looks like before it answers. This is the prompt equivalent of a rubric.
Output format for extraction tasks
Google's specific recommendation for extraction: always request JSON with an explicit schema. For analysis: bullet-point structure with a mandatory caveats or confidence section. For generation: specify tone, audience, and length in the prompt — not after the fact during editing.
The 7-Layer Synthesis: A Framework Built from All Four Papers
Across these four documents, a consistent architecture emerges. Every high-performing prompt, regardless of technique, answers seven questions in order. Answering them out of order, or skipping layers, is where most prompts fail.
- Core goal + success metrics — What does a correct output look like? How will you know it worked?
- Role / persona — Who is the model? Domain, skill level, behavioral traits. Specificity matters.
- Output format / structure — JSON, markdown, table, prose. Specify the schema if JSON. Never leave format to defaults.
- Tone / style / length — Temperature-equivalent in prose. Formal, technical, concise, comprehensive — pick and state.
- Context / background / examples — What does the model need to know? Include few-shot exemplars here (standard or CoT).
- Constraints / rules — What must not appear in the output? What is out of scope? What must always be true?
- Reasoning method — CoT? Plan-and-solve? Self-critique? Make the reasoning process explicit in the prompt.
Notice this is not a creative framework. It is derived directly from the research: layer 2 from role prompting (all four papers), layer 3 from Google's output configuration model and Anthropic's structured output guidance, layer 5 from the few-shot literature, layer 7 from Wei et al. and the CoT variants in the Prompt Report.
This framework also maps directly to a systematic prompt-building process: if you ask someone exactly these seven questions about their task — in this order — you will have enough information to engineer a prompt that works.
The System Prompt: PromptMaster2000
Below is a complete, deployable system prompt that implements the 7-layer framework as an interactive tool. It acts as a research-backed prompt engineer: it asks one strategic question per turn — in the exact order the research supports — and synthesizes the answers into an optimized final prompt.
The system prompt itself demonstrates what it teaches: XML structure (Anthropic), role specificity (all four papers), explicit reasoning method (Wei et al.), and constraint definition (Google).
You are PromptMaster2000, a research-backed interactive prompt
engineer. Your knowledge base is grounded in four primary sources:
- Wei et al. 2022 (Chain-of-Thought, arXiv:2201.11903)
- Schulhoff et al. 2024 (58-technique survey, arXiv:2406.06608)
- Anthropic Claude Prompting Guide
- Google Prompt Engineering Whitepaper v4 (2025)
<operating_rules>
1. Humans give incomplete information. Your job is to extract
the seven layers of a well-engineered prompt, one at a time.
2. Ask EXACTLY ONE strategic question per turn, in this order:
1. Core goal + success metrics?
2. Target LLM role / persona?
3. Output format / structure?
4. Tone / style / length?
5. Context / background / examples?
6. Constraints / rules?
7. Reasoning method (CoT / self-critique / plan-and-solve)?
3. After each answer: acknowledge briefly (1 sentence), then ask
the next question. Track all answers internally.
4. Do not suggest or explain techniques unless asked. Your job
is to gather, not to teach during intake.
5. When all seven layers are complete: output ONLY the final
optimized prompt. Apply these principles automatically:
- XML tags for structural separation (Anthropic)
- Role definition with domain + skill + behavior (Prompt Report)
- CoT or decomposition steps if reasoning is required (Wei et al.)
- Few-shot exemplars if examples were provided (all four papers)
- Explicit output format with schema if JSON (Google)
- Constraints as a named, bounded list (Google)
- Success criteria embedded in the prompt (Google)
6. The final prompt must be self-contained. A model seeing it
for the first time, with no other context, should produce the
correct output.
7. Output nothing outside the final prompt block. No preamble,
no explanation, no "here is your prompt." Just the prompt.
</operating_rules>
<start>
Begin by asking: What is the core goal of the prompt you are
building, and how will you know if the output is correct?
</start>
How to use this
Paste this as the system prompt in any Claude, GPT-4, or Gemini Pro conversation. Then describe a task you need a prompt for. PromptMaster2000 will ask seven questions — one at a time, in order. When it has all seven answers, it outputs a complete, structured, research-grounded prompt you can use immediately.
The value is not just the final prompt. It is the discipline of answering all seven questions. Most prompts fail because the person writing them skipped layer 1 (they never defined success), or layer 6 (they never stated constraints), or layer 7 (they never specified a reasoning method). The sequential intake forces completeness.
What to Do Monday
If you only take three things from this article:
- Add "Let's think step by step" to every non-trivial prompt you write this week. Zero effort. ~+20% on reasoning tasks. This is the minimum viable CoT and it works in any large model.
- Restructure your highest-value production prompts with XML tags. Separate role, context, task, data, and output format into named tags. Test before and after on 10 real inputs. The reliability improvement in Claude deployments is qualitative but consistent.
- Set temperature explicitly for every API call. If you are doing extraction or classification: temperature 0. If you are doing analysis: 0.2–0.4. Stop leaving it at default and wondering why outputs vary. Temperature is not a mystery — it is a dial with documented behavior.
Prompt engineering is not art. It is not intuition. It is a discipline with a growing empirical literature, measurable outcomes, and reproducible techniques. The practitioners who treat it that way are building AI systems that work. Everyone else is guessing — and getting what guessing produces.