Human-in-the-Loop

A checkpoint in an AI workflow where the system pauses and waits for a human to review, approve, or redirect its output before continuing — ensuring that human judgement governs high-stakes decisions.


What is it?

AI agents can research, draft, calculate, and even take actions in the real world. But there are moments in any workflow where the cost of an error is too high, the decision is too nuanced, or the stakes are too significant for a machine to act alone. Human-in-the-loop (HITL) is the design pattern that addresses this: it inserts deliberate pause points where a human reviews the system’s work before the next step proceeds.1

The core insight is that autonomy is not binary. Every AI system sits on a spectrum between fully manual (the AI suggests, the human does everything) and fully autonomous (the AI acts without any human involvement). Both extremes fail in practice. Full manual supervision eliminates the speed advantage of AI. Full autonomy creates unacceptable risk when things go wrong — and things always go wrong eventually.2

The parent concept, orchestration, introduces human-in-the-loop as one of its core coordination patterns. Where orchestration concerns the overall management of agents, tools, and decision flows, HITL concerns a specific type of orchestration decision: when and where to pause the machine and involve a person. The orchestrator is the system that decides to pause; the HITL checkpoint is the pause itself.

Human-in-the-loop is not a failure of automation. It is a deliberate architectural choice that places human judgement where it matters most while letting agents handle routine work autonomously. The best HITL systems route only the genuinely uncertain or high-risk actions to humans — typically 5-20% of total volume — while handling the rest without interruption.3

In plain terms

Human-in-the-loop is like a safety valve on a pressure system. The machine runs continuously under normal conditions, but when pressure exceeds a threshold, the valve opens and a human intervenes. The valve does not mean the machine is broken — it means the system is designed to handle exceptional situations safely.


At a glance


How does it work?

Human-in-the-loop is built from three core patterns, a decision framework for when to apply them, and a feedback mechanism that improves the system over time.

1. Pre-action approval — the gatekeeper pattern

The most common HITL pattern pauses execution before a high-stakes action and asks a human to approve it. The agent plans the action, presents it with supporting context and reasoning, and waits. The human reviews the proposal and either approves, modifies, or rejects it.1

This pattern is appropriate when:

  • The action is irreversible (sending an email, publishing content, executing a financial transaction)
  • The action has a large blast radius (affecting many users, records, or systems)
  • Regulatory compliance requires human sign-off (medical, legal, financial decisions)

For example: a medical AI that analyses patient symptoms and recommends a treatment plan. The AI does the analysis; a doctor reviews and approves the recommendation before it reaches the patient. The AI handles the computation; the human owns the decision.4

Think of it like...

A bank requiring two signatures on cheques above a certain amount. The first person prepares and signs the cheque (the agent proposes). The second person reviews and co-signs (the human approves). Neither signature alone is sufficient — the process requires both to proceed.


2. Post-action review — the audit pattern

Post-action review lets the agent act first, then surfaces the action for human inspection within a defined window. If the reviewer finds a problem, the system triggers a correction — a rollback, a follow-up message, or an alert.2

This pattern is appropriate when:

  • Actions are reversible within a reasonable time window
  • The volume of actions is too high for pre-action review, but spot-checking adds value
  • The cost of latency from pre-action approval exceeds the cost of occasional corrections

For example: a customer service agent that drafts and sends responses automatically, but flags every response for a four-hour post-send review. Reviewers scan a sample. Flagged responses trigger follow-up corrections. At 95% quality, this catches edge cases without blocking the agent.3

Key distinction

Pre-action approval prevents errors before they happen but adds latency. Post-action review maintains speed but accepts that some errors will reach the user and must be corrected. The choice depends on the cost of latency versus the cost of the error.


3. Confidence-based routing — the threshold pattern

The most scalable HITL pattern uses a calibrated confidence score to route actions automatically. Actions above the threshold proceed autonomously. Actions below the threshold enter an approval queue. No human decides which actions need review — the routing is automated based on the agent’s own uncertainty.3

This pattern concentrates human capacity on exactly the cases where it adds the most value: the uncertain ones. A well-calibrated system might autonomously handle 80% of actions (high confidence, low risk) while routing 20% to humans (low confidence, high risk, or edge cases).1

The critical engineering challenge is calibration. An overconfident model routes too many wrong actions autonomously. An underconfident model routes too many correct actions to the queue, overwhelming reviewers and eliminating the automation benefit. Production systems tune thresholds per action type based on error cost analysis: how expensive is a mistake for this specific action?3


4. Where to place checkpoints — the risk decision framework

Not every action needs a human checkpoint. The goal is to identify the subset of actions where human judgement adds value that exceeds the latency cost of the review. Four dimensions determine whether an action is a HITL candidate:3

DimensionQuestionExample
IrreversibilityCan the action be undone?Sending an email vs saving a draft
Blast radiusHow many people or records does it affect?Updating one profile vs bulk notification to 50,000 users
Compliance exposureDoes it create legal or regulatory obligations?Medical recommendation vs internal summary
ConfidenceHow certain is the agent about correctness?0.95 confidence vs 0.45 confidence

Actions that score high on multiple dimensions are strong HITL candidates. Actions that score low on all dimensions can proceed autonomously. The framework provides a systematic alternative to the gut-feel approach of “this seems risky, let’s add a human.”2

Think of it like...

A hospital triage system applied to AI actions. Not every patient needs the emergency room — some need a quick consultation, some need monitoring, and some are fine to go home. The triage framework prevents two failure modes: sending everyone to the ER (overwhelming the doctors) and sending everyone home (missing the emergencies).


5. Feedback loops — learning from human decisions

The most sophisticated HITL systems treat human review decisions as training data that improves the system over time. When a human modifies or rejects an agent-proposed action, that event is a labelled example: the agent was wrong, and here is the correct answer.3

Over time, aggregating these signals enables:

  • Threshold recalibration: If reviewers approve 99% of actions at a given confidence level, the threshold can safely be lowered to reduce unnecessary reviews
  • Error pattern detection: Systematic failures (the agent consistently overestimates confidence on a specific action type) become visible and fixable
  • Model improvement: Human corrections serve as fine-tuning data for the underlying model

This creates a virtuous cycle: human oversight improves the agent, which reduces the need for human oversight on routine cases, which frees human capacity for genuinely difficult decisions.5

Concept to explore

See orchestration for how feedback loops fit into the broader pattern of iterative orchestration and the evaluator-optimiser pattern.


Why do we use it?

Key reasons

1. Safety for irreversible actions. Some mistakes cannot be undone. Publishing incorrect medical advice, executing a flawed financial transaction, or sending an embarrassing message to 50,000 customers requires human review before execution — not after. HITL provides the structural guarantee that high-stakes actions pass through human judgement.4

2. Calibrated trust. Full autonomy requires complete trust in the system. HITL allows organisations to deploy AI agents with partial trust — autonomous for well-understood tasks, supervised for novel or risky ones — and increase autonomy as the system proves itself over time.2

3. Regulatory compliance. Many industries require human oversight for certain categories of decisions. Healthcare, finance, legal, and government applications often mandate that a licensed professional reviews and approves AI recommendations, regardless of the AI’s accuracy.4

4. System improvement. Every human review is a data point. Approvals confirm the agent is working correctly. Modifications and rejections identify where it is not. This feedback loop is the mechanism through which HITL systems improve over time, gradually reducing the proportion of actions that need human review.5


When do we use it?

  • When the AI takes irreversible actions in the real world (sending messages, spending money, publishing content)
  • When errors carry significant financial, legal, or reputational risk
  • When regulatory frameworks require human sign-off on certain decisions
  • When the agent operates in a new domain where its accuracy has not been established
  • When the agent’s confidence is low on a particular action
  • When actions have a large blast radius — affecting many users or records simultaneously

Rule of thumb

If the cost of a wrong action exceeds the cost of a human reviewing it, add a checkpoint. If the cost of delay exceeds the cost of an occasional error, use post-action review instead of pre-action approval.


How can I think about it?

The surgical team

Human-in-the-loop works like a surgical team in an operating theatre.

  • The AI agent is the surgical team — monitors, instruments, imaging systems, robotic arms — handling data collection, routine procedures, and precision tasks
  • The surgeon (human) makes the critical decisions: where to cut, when to change approach, when to stop
  • Pre-action approval is the surgical timeout: the team pauses before the first incision to confirm the patient, the procedure, and the site
  • Post-action review is the post-operative check: reviewing what happened to catch complications early
  • Confidence-based routing is the escalation protocol: a routine procedure runs with minimal oversight, but an unexpected finding triggers a pause for the surgeon’s judgement

The surgeon does not do everything. The team handles the routine work autonomously. But the surgeon intervenes at the moments that matter — and the system is designed so those moments are clearly identified in advance.

The publishing editor

Human-in-the-loop works like the editorial process at a newspaper.

  • Reporters (AI agents) research, gather facts, and draft articles
  • The editor (human) reviews each article before publication, checking for accuracy, bias, and legal risk
  • Pre-action approval is editorial review: no article goes to print without the editor’s sign-off
  • Post-action review is the corrections column: errors that slip through are identified and publicly corrected
  • Confidence-based routing is editorial triage: a routine weather report might go straight to print, but an investigative piece about a public figure always gets full editorial review
  • Feedback loops are editorial standards evolving: patterns of errors lead to updated style guides and training for reporters

The editor does not write every article. They focus on the decisions that require judgement: is this story accurate, fair, and safe to publish? The reporter handles the production; the editor handles the accountability.


Concepts to explore next

ConceptWhat it coversStatus
orchestrationThe broader coordination layer that manages when and how HITL checkpoints are triggeredcomplete
playbooks-as-programsStructured instructions that define where checkpoints belong in a workflowcomplete
llm-pipelinesThe multi-stage workflows into which HITL checkpoints are insertedcomplete
ragRetrieval-augmented generation, where human review can validate retrieved contextstub
dynamic-load-shiftingContinuously redistributing work between AI and humans based on confidence, risk, and capacitycomplete

Some cards don't exist yet

A broken link is a placeholder for future learning, not an error.


Check your understanding


Where this concept fits

Position in the knowledge graph

graph TD
    ORCH[Orchestration] --> HITL[Human-in-the-Loop]
    ORCH --> PP[Playbooks as Programs]
    LP[LLM Pipelines] -.->|prerequisite| HITL
    HITL --> DLS[Dynamic Load Shifting]
    style HITL fill:#4a9ede,color:#fff

Related concepts:

  • orchestration — HITL is one of the core orchestration patterns; the orchestrator decides when to invoke a human checkpoint
  • playbooks-as-programs — playbooks define where in a workflow human checkpoints should be placed, embedding HITL into the procedure
  • rag — retrieval-augmented generation can benefit from human review of retrieved context to catch hallucination or irrelevant retrieval

Sources


Further reading

Resources

Footnotes

  1. Renner, K. (2026). Human-in-the-Loop Agent Patterns: When Agents Should Ask for Help. Harness Engineering Academy. 2 3

  2. Thallus. (2026). Human-in-the-Loop AI: Why the Best AI Agents Know When to Ask Permission. Thallus. 2 3 4

  3. MyEngineeringPath. (2026). Human-in-the-Loop Patterns for AI Agents (2026). MyEngineeringPath. 2 3 4 5 6

  4. Schluntz, E. and Zhang, B. (2024). Building Effective Agents. Anthropic. 2 3

  5. Cordum. (2026). Human-in-the-Loop AI Patterns: 5 Production Designs. Cordum. 2