Agentic AI in H1 2026 stopped being a frontier bet and started behaving like infrastructure. We compiled one hundred reported agentic deployments across the half, normalised them onto a shared pattern-and-outcome grid, and read out the four trend lines that define the period: the orchestrator pattern moved from interesting to dominant, eval-first rollout went from talking point to default, governance shifted from policy decks to enforcement, and observability became table stakes rather than differentiator.
The sample is not a random survey. It is a curated mix of client engagements we ran, public case studies from major vendors and user companies, and documented production deployments referenced in conference talks, podcasts, and engineering blogs between January 1 and April 30, 2026. Skew is toward customer-facing agents and engineering-productivity workflows because that is where the public reporting concentrates. Treat the numbers as directional rather than census-grade.
What this retrospective covers: why one hundred deployments is the right unit of analysis, the four pattern classes the deployments collapse into, the outcome metrics they report, the failure modes that dominate post-mortems, an industry-level breakdown, the four trends shaping H2 2026, and a forward projection of where the production state of the art is heading next.
- 01Orchestrator pattern is dominant.Roughly half of H1 2026 production deployments now run an explicit orchestrator coordinating sub-agents or tools, up sharply from H2 2025. Single-agent and RAG-grounded patterns persist but cede ground at the higher-blast-radius workloads.
- 02Eval-first deployments are rising.Teams shipping agents in H1 increasingly built the evaluation harness before the agent itself — golden-set replay, behaviour assertions, regression suites tied to deploy gates. Eval-first is now the dividing line between resilient deployments and brittle ones.
- 03Governance enforcement is maturing.H1 2026 was the half policy documents turned into enforced controls — approval gates wired into orchestrator state, signed prompts, data-handling allowlists, audit trails queryable for compliance. Governance theatre lost ground to governance plumbing.
- 04Observability is table stakes, not differentiator.Trace-per-run, captured tool I/O, latency and cost per stage, replay harnesses for incident response — these moved from optional to assumed in H1. Deployments without them are now considered prototypes, not products.
- 05Industry adoption patterns are diverging.Engineering-productivity workloads cluster around orchestrators and code agents; customer-service deployments stay closer to RAG-grounded with supervised review; regulated industries lean heavily on supervised patterns with explicit human gates. One pattern does not fit all sectors.
01 — Why 100 DeploymentsOne hundred is the smallest sample where patterns beat anecdotes.
Quarterly AI retrospectives have a well-known failure mode: the sample size is one or two flagship case studies and the conclusions are extrapolated from there. That works for marketing posts; it does not work for engineering decisions. One hundred deployments is the practical floor where pattern signal starts to survive the noise of vendor reporting bias, survivorship bias, and cherry-picked outcome metrics.
The methodology is straightforward. Each deployment in the sample had to meet three criteria: shipped to real production traffic in H1 2026 (not just announced or piloted), enough public or shared documentation to classify pattern and at least one outcome metric, and either a named owner team or a documented vendor reference. That filter drops the long tail of unverifiable claims and keeps the data set anchored in workloads we can reason about.
Within those constraints we normalised across pattern class, workload type, industry, outcome metric, failure mode, and maturity stage. The normalisation is the slow part — public reporting uses inconsistent vocabulary for the same thing, and different vendor teams will describe an identical orchestrator pattern as either an "agent network" or a "multi-agent system" or simply "our AI platform". Cleaning the labels takes the lion's share of the analysis time.
The retrospective is also unapologetically point-in-time. The field is moving fast enough that an H2 2026 read-out four months from now will look meaningfully different — pattern share will shift, new failure modes will surface, and the eval and governance bars will rise. The point of compiling H1 now is to give teams a benchmark for where their own deployment sits relative to the production state of the art today, not to forecast it forever.
02 — PatternsFour pattern classes, unevenly adopted.
The hundred deployments collapse into four pattern classes. They are not mutually exclusive — many production systems combine two or three — but each deployment has a dominant pattern that drives its architecture. The pattern share has shifted sharply over H1 2026, with orchestrator-led systems pulling ahead of the single-agent and pure-RAG approaches that dominated H2 2025.
Orchestrator + sub-agents
An orchestrator owns workflow state and routes work to specialist sub-agents or tool calls. Dominant pattern in H1 2026 — roughly half the sample. Strong fit for multi-step workflows with clear hand-offs (research, code review, customer triage with escalation).
≈50% of H1 sampleSingle-agent
One agent loop with a fixed toolset and direct user interaction. Simpler to build and reason about; dominant in chat-style interfaces, single-purpose code agents, and structured-task automation. Roughly a quarter of the sample — losing ground at higher-blast-radius workloads.
≈25% of H1 sampleRAG-grounded answerer
Retrieval over a curated corpus feeds a generation step; agency is minimal — the agent answers from grounded context rather than taking multi-step actions. Dominant in customer-service knowledge bases, internal documentation Q&A, and regulated-industry research support.
≈15% of H1 sampleSupervised / human-in-loop
Every material action passes a human checkpoint before execution. Dominant in regulated workflows (legal, healthcare, financial), high-blast-radius internal ops, and trust-rebuild deployments after a prior incident. Slower throughput, dramatically lower incident rate.
≈10% of H1 sampleThe interesting story in the pattern data is not which class wins outright — it is the increasing willingness of teams to combine patterns inside a single product. An orchestrator with a RAG-grounded research sub-agent and a supervised checkpoint before any external mutation is now a recognisable shape rather than an experimental composition. The H2 2025 single-agent-only deployments that survived into H1 either added an orchestrator layer or shrank their scope to genuinely single-purpose tasks.
The orchestrator pattern's rise tracks the maturation of the tooling around it. Workflow platforms with durable execution (Temporal, Inngest, Restate, the Vercel Workflow DevKit and similar) made the orchestrator layer cheap to build correctly, and the new generation of agent SDKs encode the orchestrator- plus-sub-agent shape as the default rather than an advanced pattern. Teams who would have rolled a single-agent loop in mid-2025 now reach for the orchestrator-shaped scaffolding by default.
"The single-agent loop is now a niche tool, not the default starting point. By mid-2026 the orchestrator-plus-sub-agents shape will be assumed in the same way that REST APIs are assumed for synchronous backends."— Production audit, March 2026
03 — OutcomesWhat the deployments actually shipped.
Outcome reporting is where vendor marketing and engineering reality diverge most sharply. We normalised every deployment against the same four metric classes — productivity lift, cost-per-task reduction, customer-experience scores, and deflection / containment — and dropped any claim without a stated baseline. The chart below shows the median lift reported within each metric class, with the spread captured in the sub-labels. Bars are not absolute scores; they are relative magnitudes of typical reported lift.
Reported outcome lift by metric class · H1 2026 sample
Source: Digital Applied H1 2026 deployment retrospective, n=100The honest version of the outcome story is that the median lift is real but the spread is wide. A median +28% productivity lift means half the deployments delivered more and half delivered less; the 8% lift cases were typically deployments where the human-in-loop overhead consumed most of the agent's throughput gain, and the 52% cases were almost always engineering-productivity workflows (code review, test generation, doc generation) where the baseline was a high-cost human task.
The 20% failure / restart rate at the bottom of the chart is the metric that is least often reported and most worth paying attention to. It captures the percentage of agent runs that require operator intervention — a forced rollback, a compensation step, a manual override of a stuck checkpoint, or an outright restart. That number is the single best predictor of whether a deployment is in the prototype phase or the production phase, and most teams under-report it because it complicates the marketing story.
Productivity workloads
Engineering productivity, content generation, and operational task automation cluster at the high end. Customer-service workloads cluster lower — the human time saved is real but bounded by the supervision overhead.
n=100Cost per task
The most consistent metric across the sample. Customer-service and internal-ops deployments hit this band reliably; the upper-bound −61% cases were narrow workloads with a high pre-agent labour cost.
Most consistentRuns requiring operator
The under-reported metric. The teams that publish this number are also the teams running the most disciplined resilience and eval programs. Treat it as a proxy for production maturity.
Maturity proxyOne pattern worth calling out: the supervised / human-in-loop class delivered the smallest productivity lift but the largest customer-experience and CSAT gains. That is not a contradiction; it is the trade-off made explicit. Supervised deployments slow the throughput in exchange for predictability — fewer surprises, fewer escalations, fewer apology emails. For workloads where the cost of an agent error exceeds the cost of the slower path, supervised is the right answer even though it shows up as a modest line on the productivity chart.
04 — Failure ModesThe failure catalogue that dominated H1 post-mortems.
Failures cluster. The same handful of root causes recurs across otherwise unrelated deployments, and the same handful of preventive controls appears in the deployments that avoided them. Three classes dominated H1 2026 post-mortems by a wide margin: eval gaps, tool-call chaos, and governance theatre. None of these are exotic; all of them are addressable with discipline rather than new technology.
Eval gaps
The most common failure was a deployment that passed human-curated evaluation in development and degraded sharply in production once the input distribution shifted. The pattern was almost always the same — a small golden set used to gate the initial release, no behavioural assertions to catch silent regressions, no production-trace replay against the golden set as part of CI. The fix is the eval-first discipline covered in the trends section: build the evaluation harness before the agent, treat the golden set as a living artefact, and gate deployments on it. That single change separated the H1 deployments that scaled cleanly from the ones that quietly regressed.
Tool-call chaos
The second cluster was failures in the tool-call layer — agents calling tools with malformed arguments, retrying mutating tools without idempotency keys, hanging on tool calls without timeouts, and looping endlessly on tools that returned ambiguous error states. The post-mortems read like a checklist of the resilience audit we published in May. The fix is mechanical — per-tool timeouts, idempotency keys on every mutating call, bounded retries, structured error responses from tools — but it requires engineering investment that many H1 teams deferred until the first incident.
Governance theatre
The third cluster was governance that existed on paper but not in the code path. Policy documents declared that certain data categories required explicit handling and certain actions required approval, while the actual agent implementation routed around those requirements either accidentally or deliberately. The H1 deployments that avoided this failure class had wired the governance controls directly into the orchestrator state — an approval gate that the orchestrator literally could not bypass, a data-handling allowlist enforced at the tool boundary, a signed-prompt requirement validated before execution. Paper governance lost; plumbed governance won.
"Most H1 production incidents were not capability failures — they were resilience failures, evaluation failures, or governance failures. The models were strong enough; the scaffolding was thin."— Digital Applied H1 2026 retrospective
The companion checklist covers the resilience side of this in depth — our agentic workflow resilience audit grades production workflows across timeouts, retries, rollback, human-in-the-loop, observability, and replay. For the failure-mode catalogue applied to a more pointed list of common deployment mistakes, the agentic AI anti-patterns post lays out the ten failure shapes we keep finding inside the sample.
05 — Industry BreakdownFour industries, four different production shapes.
Industry slicing matters more in H1 2026 than it did a year ago, because the pattern share inside each industry is now visibly different. Engineering-productivity workloads have converged on orchestrator-plus-sub-agents; customer-service deployments cluster around RAG-grounded answerers with optional supervised review; regulated industries lean hard on the supervised pattern; and operations / internal-tooling workloads sit closest to single-agent loops with carefully scoped toolsets.
Engineering productivity
Orchestrator dominant · 70% of segmentCode review, test generation, doc generation, repository search and refactor. Highest productivity lifts in the sample and the fastest pattern convergence — orchestrator-plus-sub-agents is now the default shape, and single-agent deployments here look increasingly dated.
Highest productivity liftCustomer service
RAG-grounded dominant · 55% of segmentKnowledge-base Q&A, ticket triage, deflection bots, agent-assist suggestions. Deflection rates of 40-60% are achievable on narrow scopes but degrade sharply on open-domain. Supervised review on edge cases is the difference between deflection and customer-experience damage.
Highest CSAT varianceRegulated industries
Supervised dominant · 60% of segmentLegal review, healthcare documentation, financial reconciliation, compliance support. Throughput trades against predictability — the segment's lower productivity lift is the price of the highest CSAT and lowest intervention rates in the sample.
Lowest failure rateOperations & internal tooling
Single-agent dominant · 45% of segmentInternal automations, ops runbooks, alert triage, lightweight data pulls. Single-agent loops persist here because the scope is genuinely narrow and the orchestrator overhead is hard to justify. The 20% intervention rate sits roughly at the sample median.
Single-agent strongholdThe industry breakdown also exposes where the field has the most ground left to cover. Regulated-industry deployments are disproportionately supervised because the cost of an agent error is high — not because supervision is the optimal long-term answer. As governance plumbing matures and evaluation harnesses tighten, expect a meaningful migration in those segments from heavily supervised patterns toward orchestrator-with-explicit- gates patterns over the next two halves. Customer-service deployments will polarise: narrow-scope deflection bots stay RAG-grounded, while general-purpose support agents migrate toward orchestrators with supervised review on uncertain cases.
06 — Four TrendsThe four lines that shaped H1 2026.
Four trend lines run through the hundred deployments and define the half. None of them is a single technology; each is a shift in how teams build, evaluate, govern, and operate agents in production. Together they describe the maturation of the field from frontier-of-the-month to industrial-discipline-in-progress.
Trend 01 · Orchestrator pattern dominance
The single-agent loop ceded the high-blast-radius workloads to orchestrator-plus-sub-agents during H1. The shift was enabled by cheap durable-execution platforms, by agent SDKs that encode the orchestrator shape as default, and by an industry-wide reading of mid-2025 incidents that pinned single-agent fragility as a recurring root cause. The single-agent pattern is not dying — it keeps the genuinely-narrow workloads — but the centre of gravity has moved.
Trend 02 · Eval-first deployments rising
The teams shipping in H1 increasingly built the evaluation harness before they built the agent. Golden sets curated up front, behavioural assertions written alongside the prompts, regression suites tied to deploy gates, production traces looping back into the golden set on every meaningful incident. Eval-first is the H1 hallmark of a deployment likely to survive the year; the eval-after deployments populated most of the failure-mode catalogue.
Trend 03 · Governance enforcement maturing
Policy documents lost ground to enforced controls in H1. Approval gates wired into the orchestrator state, data-handling allowlists enforced at the tool boundary, signed prompts validated before execution, audit trails queryable for compliance review. The governance-theatre deployments showed up in post-mortems as written policy that the running code cheerfully ignored; the plumbed-governance deployments showed up as audits that closed without findings.
Trend 04 · Observability becoming table stakes
Trace-per-run with captured tool I/O, latency and cost per stage, structured error categorisation, and at least an early replay harness for incident response — these moved from optional to assumed in H1 2026. Deployments without them are now considered prototypes regardless of how clean their happy path looks. The observability bar is rising fast enough that the deployments shipping in H2 will look back on H1 instrumentation as thin.
07 — H2 ProjectionWhat H2 2026 is likely to look like.
Forecasts in this field age badly, so the projections below are framed as directional rather than point. Three shifts are likely enough that teams should plan as though they will happen, and two more are worth tracking even if they only land in 2027.
Highly likely
Orchestrator share will keep climbing as the tooling matures. Expect the orchestrator pattern to account for 60-65% of H2 production deployments — drawing share from single-agent loops first and from pure-RAG patterns at the edges. The orchestrator-shaped scaffolding is becoming default in the agent SDKs and the workflow platforms, which means new teams will land there without consciously choosing it.
Eval-first will move from rising to assumed. The spread between eval-first and eval-after deployments in H1 is wide enough that publishing the absence of an evaluation harness will become embarrassing rather than tolerated. Expect the H2 failure-mode catalogue to feature far fewer eval-gap incidents and far more incidents tied to the gaps that come after evaluation matures — distribution shift detection, eval-set staleness, behavioural drift inside graded categories.
Governance enforcement will harden in regulated sectors first — legal, healthcare, financial — and from there will migrate into adjacent industries as the patterns generalise. Expect explicit governance plumbing (approval gates, allowlists, signed prompts, audit trails) to be a checklist item in H2 RFPs rather than a differentiator.
Worth tracking
Inference cost will keep falling fast enough to change the architecture math. The H1 efficiency story on open-weight models — DeepSeek V4 most visibly — already shifts on-prem long-context economics. If that trend continues, expect H2 deployments to make different routing decisions between closed-frontier and open-weight models on a per-workload basis, and expect the orchestrator pattern to gain another tailwind as multi-model routing becomes worth the orchestration overhead.
Multi-tenant agentic deployments will surface their first major incidents — cross-tenant data leakage, shared-prompt poisoning, tool-allowlist confusion across tenants. These are not new vulnerability classes but they have agent-specific shapes that will take a few public incidents to normalise into the standard hardening checklist. Plan for tenant-scoped tracing and tenant-scoped governance enforcement now rather than after the first headline.
For teams planning H2 deployments now, the directional read is simple. Default to the orchestrator pattern unless the workload is genuinely narrow. Build the eval harness before the agent and keep it close to production traces. Plumb governance into the code path; do not rely on policy documents. Instrument observability as if you will need to replay a production incident next week, because you will. If you want help applying this retrospective to a specific roadmap, our AI transformation engagements start exactly here — pattern selection per workload, eval-first rollout playbook, governance enforcement implementation, and observability architecture sized to the deployment.
H1 2026 was the year agentic AI production patterns crystallised.
One hundred deployments analysed, four pattern classes normalised, four trend lines identified — the headline of the half is that agentic AI stopped behaving like a frontier and started behaving like infrastructure. The orchestrator pattern is dominant, eval-first is rising fast enough to become the default, governance is migrating from paper to plumbing, and observability has shifted from optional to assumed. None of those four is a vendor pitch; all four are visible in the sample.
The honest framing is that the median deployment in the sample is still maturing. Median productivity lift was +28%; median cost reduction was −34%; median intervention rate was 20% of runs. Those are real numbers and they are also evidence that the field is closer to early-production than to mature production. The H2 work is not new capabilities — it is the same four trends, deeper. More orchestrators, more eval discipline, more plumbed governance, more observability.
The practical next step for any team with an agent in production is to score its own deployment against this retrospective. Which pattern class is it in, and is that the right one for the workload? Where does it sit on the evaluation, governance, and observability axes relative to the sample medians? And which of the four H1 trends has not yet shown up in its own roadmap? The questions are uncomfortable; the answers are how the H2 2026 sample improves on the H1 one.