Skip to main content
Back to Blog
Claude 4.6GPT-5Gemini 2.5prompt engineeringreasoning modelsextended thinking

Advanced Prompt Engineering in 2026: Claude 4.6, GPT-5.4, and Gemini 2.5 Deep Think

The 2026 playbook for prompting reasoning models. Learn how to use Claude's adaptive thinking, GPT-5.4's reasoning effort levels, and Gemini Deep Think — plus the old techniques that stopped working.

SurePrompts Team
April 8, 2026
15 min read

The prompt engineering advice you learned in 2023 is wrong for 2026's frontier models. Here's what actually works on Claude 4.6, GPT-5.4, and Gemini 2.5 Deep Think.

The Rules Changed in 2026

Three years ago, the best prompt engineering trick was "let's think step by step." You added it to a prompt, the model wrote out its reasoning, and accuracy went up. It was magic.

In 2026, that same phrase is either useless or actively counterproductive. The frontier reasoning models — Claude 4.6, GPT-5.4, and Gemini 2.5 with Deep Think — already think before they answer. Telling them to think "step by step" doesn't unlock hidden reasoning. At best it wastes thinking budget the model already allocated. At worst it trains the model to narrate its reasoning in the final answer instead of letting the private reasoning tokens do the work.

The new game is not about coaxing reasoning out of a model. It's about telling the model how much reasoning to do, when to do it, and where the structure of your request should guide its thinking. Each of the three major labs exposes these controls differently, and each one rewards a different prompting style.

This is the 2026 playbook. It assumes you already know the basics. If you don't, start with the complete guide to AI prompt engineering and come back.

ControlClaude 4.6GPT-5.4Gemini 2.5 Pro
Reasoning mechanismAdaptive thinkingReasoning effort + verbosityDeep Think (parallel hypotheses)
Effort levelslow / medium / high / maxnone / low / medium / high / xhighToggle on/off (Gemini app)
Thinking between tool callsInterleaved (built-in)Supported via Thinking modeN/A for Deep Think
Context window1M tokens standard1M tokens (5.4 tier)1M tokens
Native structureXML tagsMarkdown + structured outputsData tables + multimodal
Mid-response steeringNoYes (add instructions while thinking)No

Claude 4.6: Adaptive Thinking and Interleaved Reasoning

Claude Opus 4.6 and Sonnet 4.6 both ship with adaptive thinking as the recommended mode. You set an effort budget — low, medium, high (default), or max — and Claude decides how much of that budget to spend on each request. At low effort Claude may skip thinking entirely on a simple question; at max it runs deep reasoning chains on gnarly problems.

The important thing to understand: adaptive thinking is a dial, not a prompt trick. You control it through the API parameter, not through natural language. Writing "please think hard" in the user message does nothing that setting effort to high doesn't already do — and it consumes tokens that could be carrying your actual task.

When to use which effort level

  • Low — classification, simple rewrites, lookups against provided context, short translations. Any task where the answer is close to the surface.
  • Medium — analysis of short documents, most business writing, straightforward code changes, single-file refactors.
  • High (default) — multi-document analysis, architectural reviews, code generation for non-trivial features, nuanced writing where tone matters.
  • Max — mathematical proofs, research synthesis, debugging across many files, anything where a wrong answer is expensive to catch.

Start at high. Drop to medium if latency matters and quality holds. Only climb to max when you've seen high miss.

Interleaved thinking changes how you prompt tool-using agents

On Opus 4.6, interleaved thinking is built in — no beta header, no feature flag. When Claude calls a tool, it can think between the tool result and the next action. For prompting, this means you should stop writing giant multi-step instructions into the system prompt. Claude can plan, act, observe, replan. Your job is to give it clean tool definitions and a clear objective, not a pre-baked algorithm.

Tip

If you're porting a Claude 3.5 agent to 4.6, delete the "plan before acting" scaffolding from your system prompt. It's doing nothing, and in some cases it crowds out the model's natural interleaved reasoning.

XML tags still matter — but use them for input, not meta-instructions

Claude is still trained on XML-structured inputs, and structured output reliability is meaningfully higher when you wrap your context in tags. The shift in 2026 is what you put inside them. In 2023 people wrote things like <thinking_instructions>first consider X, then Y</thinking_instructions>. Don't do this anymore. The model's thinking is already allocated. Reserve tags for content — the document you want analyzed, the code you want reviewed, the constraints you want enforced.

Here's the pattern that works in 2026:

code
You are reviewing a database migration for safety.

<migration>
ALTER TABLE orders ADD COLUMN customer_tier VARCHAR(20) NOT NULL DEFAULT 'standard';
CREATE INDEX idx_orders_customer_tier ON orders(customer_tier);
</migration>

<constraints>
- Table has ~50M rows
- PostgreSQL 16
- Zero-downtime deployment required
- Concurrent writes are expected during migration
</constraints>

<review_criteria>
1. Lock escalation risks
2. Backfill safety under concurrent writes
3. Rollback path
</review_criteria>

Identify risks and propose the safest sequence of steps.

No "think step by step." No persona stacking ("you are a senior DBA with 20 years of experience"). Just clean context and a crisp ask. On high effort, Claude will reason through each <review_criteria> item in its thinking tokens and return a focused answer.

Cache the static parts of your prompt

If you're hitting the API repeatedly with the same system prompt and tool definitions, prompt caching is free money. Cache writes cost 25% more than a normal input token, and cache reads cost 90% less. On any workflow longer than two turns, the math tips in your favor fast. The ephemeral cache has a 5-minute TTL, so keep requests flowing or reset the cache explicitly.

The rule: put everything stable at the top of the prompt (system instructions, tool schemas, large reference context), mark the last stable block with cache_control: { type: "ephemeral" }, and put the changing user message after it. Don't try to cache the user turn — it defeats the point.

GPT-5.4: Reasoning Effort and Verbosity Are Separate Dials

GPT-5.4 exposes two independent controls that most prompters still conflate: reasoning effort and verbosity. Reasoning effort — none, low, medium, high, or xhigh — governs how much the model thinks before responding. Verbosity governs how long the answer is. They are not the same knob.

You can have a model that thinks hard (xhigh) and answers tersely (verbosity: low). That combination is gold for things like math proofs, legal analysis, or complex diagnostic questions where you want the model's full reasoning power but don't want a five-paragraph explanation wrapped around a one-sentence answer.

When to use each reasoning effort

  • none — pure classification, simple extraction, format conversion. GPT-5.4 responds almost like a traditional instruction-tuned model at this level.
  • low — standard Q&A, summarization, short code snippets.
  • medium — analysis tasks, moderate refactoring, writing with constraints.
  • high — multi-step problems, research synthesis, complex code generation.
  • xhigh — competition math, formal reasoning, high-stakes debugging.

Every step up costs latency and tokens. xhigh is slow. Use it when you'd otherwise have to double-check the answer yourself.

The biggest prompting mistake on GPT-5.4

Do not prepend "think step by step" or "let's reason carefully" to an xhigh prompt. It's an anti-pattern. The reasoning is already happening in the private thinking tokens. Adding that phrase to the visible prompt trains the model to also narrate reasoning in the answer, which bloats output and — in testing — sometimes causes the answer to drift from the private thinking chain. If you need the reasoning shown, ask for a "brief justification" after the answer. Otherwise, trust the hidden work.

Warning

Mixing "think step by step" with high reasoning effort is a common hand-me-down from 2023 guides. On GPT-5.4 it often lowers answer quality because it forces the model to duplicate work: once in hidden reasoning tokens and again in the visible answer.

Mid-thinking steering is the sleeper feature

GPT-5.4 Thinking lets you inject additional instructions while the model is mid-reasoning. You don't have to wait for a wrong answer and retry. If you're running an interactive workflow and notice the model heading down a bad path, you can nudge it mid-stream. This is unique to GPT-5.4 in 2026 — neither Claude nor Gemini exposes the equivalent.

A clean GPT-5.4 API call

code
response = client.responses.create(
    model="gpt-5.4",
    input=[
        {
            "role": "developer",
            "content": "You are auditing a TypeScript codebase for security bugs. Report findings as a numbered list with severity (Critical/High/Medium/Low) and file:line references."
        },
        {
            "role": "user",
            "content": "<attached: auth_middleware.ts, session_store.ts, jwt_utils.ts>\n\nAudit these three files. Focus on authentication bypass and session fixation."
        }
    ],
    reasoning={"effort": "high"},
    text={"verbosity": "low"}
)

No "think carefully." No chain-of-thought primer. The reasoning.effort parameter does that work. The verbosity: low setting keeps the answer tight even though reasoning was high. This separation is what advanced prompting looks like in 2026.

Gemini 2.5 Pro Deep Think: Parallel Hypothesis Prompting

Gemini 2.5 Pro's Deep Think mode is the most architecturally different of the three. Instead of a single reasoning chain with adjustable length, Deep Think runs parallel thinking — it generates multiple hypotheses at once, considers them simultaneously, and can revise or combine them before committing to an answer.

This changes what a "good" prompt looks like for Deep Think. You're not asking a single reasoner to think harder. You're asking a committee of parallel reasoners to explore a space. Prompts that narrow the space too early waste that capability. Prompts that open the space with multiple framings get dramatically better results.

Deep Think is available as a toggle in the Gemini app for Google AI Ultra subscribers — you turn it on in the prompt bar before sending.

Best-fit tasks for Deep Think

Google's own positioning is clear on where Deep Think wins: hard math, competition-level coding, strategic planning, iterative design, and research tasks that benefit from comparing multiple approaches. It scores well on benchmarks like USAMO (the USA Mathematical Olympiad) and LiveCodeBench. For day-to-day writing and summarization, regular Gemini 2.5 Pro is faster and good enough.

The technique: ask for parallel exploration explicitly

Deep Think does parallel reasoning natively, but your prompt can amplify it. Instead of asking "what's the best approach to X?", ask "explore at least three distinct approaches to X, compare their tradeoffs, then recommend one." You're matching the prompt's surface structure to the model's internal parallelism. In practice the outputs get more honest — Deep Think is more likely to surface the runner-up approach and explain why it lost.

The technique: stack constraints

Single-pass models get overwhelmed when you pile on constraints. Deep Think tends to handle layered constraints well because it can weigh them in parallel. Don't be shy. "Design a caching layer that supports sub-10ms p99 reads, graceful degradation during Redis outages, stampede protection, and per-tenant isolation, for a system with 500k concurrent users and a 2GB working set" is the kind of prompt Deep Think actually wants.

Pair Deep Think with long context

Gemini's 1M-token window and Deep Think's parallel reasoning complement each other. If you're analyzing a large document — a contract, a research paper stack, a codebase dump — upload the full thing and let Deep Think work on it. This is also where multimodal prompting pays off: mix text, images, and tables in the same request and Deep Think will reason across modalities.

code
Upload: quarterly_results.pdf (87 pages), q3_earnings_call.mp3, product_roadmap.png

Analyze this quarter's performance. Explore at least three narratives that
explain the Q3 revenue dip, using evidence from the PDF, the earnings call
transcript, and the roadmap timeline. For each narrative, list the strongest
supporting evidence and the strongest counter-evidence. Then recommend which
narrative leadership should adopt in the public messaging.

Notice what's not in this prompt: no "think step by step," no "you are a financial analyst," no "let's think carefully." Deep Think is already going to think carefully. Your job is to frame the exploration, not coach the reasoning.

The Techniques That Stopped Working in 2026

Five hand-me-downs from the 2023 prompting era that are now dead weight on frontier models:

Before

You are a senior software engineer with 15 years of experience. Let's think step by step before answering. Take a deep breath. If you're unsure about anything, say so. Now, review this code for bugs.

After

<code>[code here]</code>

Identify bugs. For each bug, state the file:line, the failure mode, and the minimal fix. Prioritize correctness bugs over style issues.

  • Verbose "step by step" prefixes. Reasoning models already allocate thinking budget. The phrase doesn't unlock anything on Claude 4.6, GPT-5.4, or Gemini 2.5 Deep Think — it just adds noise and sometimes causes the model to duplicate its reasoning in the visible answer.
  • Elaborate persona stacking. "You are a senior X with 15 years of experience at top-tier firms" is a 2023 trick. On reasoning models, detailed personas are less effective than direct task framing. State the task, provide the context, set the evaluation criteria — skip the costume.
  • "Take a deep breath" emotional primers. These showed small effects in GPT-3.5-era evals. On models with native reasoning, they do nothing measurable. Drop them.
  • Few-shot examples for pure reasoning tasks. Few-shot prompting still wins for pattern-matching and format-following. But for tasks where the model should reason from scratch, examples can actually anchor the model to your specific solution path and reduce solution diversity. If you want a creative solution, don't show it a template answer.
  • Confidence-eliciting phrases. "If you're unsure, say so" was good advice for non-reasoning models that would confidently hallucinate. Reasoning models self-check during their thinking tokens and already flag uncertainty when it matters. The phrase is now a tax on token budget with diminishing returns. (It's still useful on non-reasoning models — don't remove it from your GPT-4o-mini or Haiku workflows.)

The Wharton School's 2025 Prompting Science Report found that chain of thought prompting adds negligible benefit on reasoning models that already think step-by-step. That's the same principle operating across all five of these techniques: the model is already doing the work you're trying to prompt into existence.

A Universal 2026 Prompt Framework

Across all three models, a well-structured 2026 prompt follows the same five steps:

1

State the task cleanly. No filler, no personas, no emotional primers. One sentence on what you want.

2

Provide structured context. XML tags for Claude, markdown headings for GPT-5.4, data tables or uploaded documents for Gemini. Match the model's native affordances.

3

Set reasoning effort via API, not via language. Use reasoning.effort on GPT-5.4, thinking.effort (or equivalent) on Claude 4.6, the Deep Think toggle on Gemini. Don't try to coerce effort through the user message.

4

Define success criteria. What does a correct answer look like? What should the model prioritize? What should it explicitly ignore?

5

Request the specific output format. Ask for the shape you need: a numbered list, a JSON object, a code block, a markdown table. Don't leave this to chance.

Step 2 is where most people still slip. Claude wants XML. GPT-5.4 wants clean markdown and will often do its best work with structured outputs / JSON schemas. Gemini wants tables, data, and multimodal input. You'll write the same task three different ways — not because the models are fundamentally different, but because each is easiest to steer through its native input vocabulary.

Key Takeaways

  • The old advice is outdated. "Let's think step by step," elaborate personas, and emotional primers were tricks for 2023's non-reasoning models. On Claude 4.6, GPT-5.4, and Gemini 2.5 Deep Think, they're noise.
  • Control reasoning via the API, not the prompt text. Set effort levels through parameters. The visible prompt is for your task and context.
  • Separate reasoning depth from answer length. GPT-5.4's split between reasoning.effort and verbosity is the cleanest example, but the same principle applies everywhere.
  • Match input structure to each model. XML for Claude, clean markdown / structured outputs for GPT-5.4, tables and multimodal uploads for Gemini.
  • Cache what's stable and exploit interleaved thinking. Claude 4.6's prompt caching and interleaved tool use are how you build production-grade agents without re-paying for your system prompt on every turn.

The frontier has moved. Prompts you wrote in 2023 will still run — but they're paying a tax, and a cleaner 2026-native prompt will usually beat them on both latency and quality. Go audit one workflow this week and see.


Ready to stop reinventing prompts for each model? SurePrompts generates model-optimized variations automatically — structured for Claude, cleaned for GPT-5.4, and tuned for Gemini. Try the Claude Prompt Generator, ChatGPT Prompt Generator, or Gemini Prompt Generator — or browse every prompt engineering technique in our research-backed pillar guide.

Ready to Level Up Your Prompts?

Stop struggling with AI outputs. Use SurePrompts to create professional, optimized prompts in under 60 seconds.

Try AI Prompt Generator