PromptingBasics

ReAct Prompting

By
Dan Lee
Dan Lee
Dec 20, 2025

If you’ve ever watched an AI give a confident answer that’s almost right (and occasionally totally wrong), you’ve felt the need for a better approach. Enter ReAct prompting—short for Reason + Act. It’s a practical technique that helps models think through a problem and take actions (like using tools, checking facts, or asking clarifying questions) before finalizing an answer.

For AI engineers, ReAct is a blueprint for tool-using agents. For non-technical professionals, it’s a simple way to get AI to show its work, validate assumptions, and avoid “fabricated confidence.”

What ReAct Means

ReAct = the model alternates between reasoning (“what should I do next?”) and actions (“use a tool / query data / check constraints”) to reach a better answer.

Why It Helps

ReAct is valuable because it reduces two common failure modes:

  • Jumping to conclusions: The model answers too early without checking.
  • Guessing missing information: The model fills in gaps instead of asking.

Instead, ReAct encourages a loop: think → do → observe → refine → answer.

In practice, “Act” can mean different things depending on your workflow:

  • Engineers: call tools (search, database, calculator), run code, inspect logs
  • Business users: request missing details, extract key facts, create structured drafts, double-check logic

How to Prompt It

You don’t need magical keywords. You need two things:

  1. Permission to take actions (ask questions, verify, use provided data)
  2. A required output shape so you can trust what you get

Example 1: Tool-Ready Engineering Task

You’re debugging a production issue and want a disciplined workflow, not a one-shot guess.

Text
Context: You are a senior backend engineer. We’re seeing intermittent 502s from a FastAPI service behind Nginx.
Instruction: Use a ReAct-style approach: (1) list what you need to check, (2) propose the next action, (3) explain what you expect to learn, and only then (4) give a likely root cause and fix.
Input Data:
- Nginx error: "upstream timed out (110: Connection timed out) while reading response header from upstream"
- App logs: periodic "DB pool exhausted" warnings
Output Indicator: Return a Markdown table with columns: Step, Reasoning, Action, Expected Signal, Decision. End with a short mitigation plan (max 6 bullets).

This prompt forces sequence and verification. Instead of “it’s probably X,” you get a guided investigation plan that’s easy to execute and audit.

Example 2: ReAct for a Non-Technical Workflow

Let’s make it practical for a founder or exec assistant dealing with a sensitive message.

Text
Context: I’m a CEO replying to a partner who is unhappy about a delayed deliverable. The relationship matters and I need to keep trust.
Instruction: Use a ReAct-style approach: first ask 3 clarifying questions you need answered, then draft two email options (direct + diplomatic).
Input Data: Partner message: "We’re disappointed. This delay impacts our launch timeline. What happened?"
Output Indicator: Provide:
1) 3 questions (bulleted)
2) Email Option A (direct, <=140 words)
3) Email Option B (more diplomatic, <=170 words)
4) A final 3-bullet checklist to verify tone and commitments.

ReAct keeps the model from inventing details (“the delay was due to X”) and instead guides it to ask what it needs, then produce drafts with guardrails.

Don’t Leak Hidden Reasoning

In production systems, you often shouldn’t ask models to reveal detailed internal reasoning. Instead, have them output structured steps, checks, assumptions, and tool results you can audit.

Takeaway

ReAct prompting is a simple mental model: don’t just answer—think, act, observe, then answer. It helps engineers build safer tool-using agents and helps non-technical teams get more accurate, less “hand-wavy” outputs. When the stakes are high or the task is multi-step, ReAct turns AI from a guesser into a deliberate problem-solver.

Dan Lee

Dan Lee

DataInterview Founder (Ex-Google)

Dan Lee is an AI tech lead with 10+ years of industry experience across data engineering, machine learning, and applied AI. He founded DataInterview and previously worked as an engineer at Google.