Migrating from Gemini 3 Pro to Gemini 3.1 is not a drop-in upgrade. The release moves on four axes at once — a new Deep Think reasoning mode that is not available on every tier, a tier-pricing shift that changes the cost-quality Pareto for several common workloads, a safety-filter overhaul with new categories and altered default thresholds, and a tool-use schema diff that quietly breaks function calling for any agent built against the previous shape.
None of these axes individually justifies an emergency cut-over. Together, they make a hands-off "flip the model string and ship" approach the most expensive mistake teams make. The wrong-workload regression typically shows up two to three days after the cut-over, when the safety-filter change starts refusing the previously-passing tail of customer queries, or the schema diff silently strips an argument from a tool call your agent uses once per thousand invocations.
This playbook is the per-workload migration sequence we run for client engagements — assess, shadow-test, cut over, retire — with the specific gotchas to watch for on each of the four breaking axes. The end state is the same workloads running on 3.1, the cost-curve modelled, and a clean retirement window for the 3 Pro endpoints.
- 01Deep Think availability is per-tier — confirm before designing around it.Deep Think is not a default capability across every Gemini 3.1 surface. Confirm the tier you actually have provisions Deep Think before building workloads that depend on it, and budget for the latency and cost step-up it introduces.
- 02Safety-filter changes can silently regress some workloads.New filter categories and altered default thresholds mean the same prompt that passed on 3 Pro may refuse on 3.1. Audit refusal rates on a representative sample before cut-over — the regression rarely shows up in synthetic evals.
- 03Tier pricing shifts the cost-quality Pareto.3.1's tier structure changes the per-token economics for several common workloads. Re-evaluate model choice per workload — the best answer for bulk classification may now differ from the best answer for long-context retrieval, even within the same provider.
- 04Tool-use schema diff requires per-tool migration.Function-calling and tool-use schemas are not byte-compatible with 3 Pro. Argument typing, response shape, and a handful of metadata fields shifted. Treat each tool definition as its own migration unit — do not assume drop-in compatibility.
- 05Per-workload shadow-testing surfaces the most regressions.Build the eval harness before you flip routing. Shadow-test by mirroring a slice of production traffic to 3.1 and diffing outputs — that's where the long-tail refusal rate, the schema-mismatch failures, and the cost-curve surprises actually become visible.
01 — What's New3.1 ships in four axes — Deep Think, tiers, safety, tools.
Most provider releases are dominated by one breaking change and a few quality-of-life additions. Gemini 3.1 is the opposite — four roughly equal-weight changes shipped in the same release window, and every one of them is consequential for at least one common workload class. Treating the upgrade as a single migration is the top-of-list mistake; treating it as four migrations on one upgrade window is the right mental model.
The four axes are independent but interact. The Deep Think mode is gated by tier, which couples the availability question to the pricing question. The safety-filter overhaul interacts with tool-use because some tool-call arguments now trip filter categories they did not on 3 Pro. Sequencing matters — assess all four before designing the rollout, not one at a time.
Deep Think mode
tier-gated · extended reasoningNew reasoning mode with extended deliberation budgets and stronger performance on multi-step problems. Tier-gated — confirm provisioning before designing around it. Latency and cost step-up are both meaningful.
ReasoningTier pricing shifts
per-token rebalance · per-workload ParetoPer-tier input and output token rates re-balanced. The cheapest answer for any given workload may have moved within the tier ladder — re-evaluate per workload, not org-wide.
CostSafety filters
new categories · new defaultsFilter category set expanded, default thresholds altered for several existing categories. Same prompt may pass on 3 Pro and refuse on 3.1. Audit refusal rates on a representative sample.
BehaviorTool-use schema
argument typing · response shapeFunction-calling schemas shifted in argument typing, response shape, and a handful of metadata fields. Per-tool migration required — no drop-in compatibility guarantee for non-trivial schemas.
Agents02 — Deep ThinkAvailability and access tiers.
Deep Think is Gemini 3.1's headline capability — an extended reasoning mode that holds more state across deliberation steps and delivers materially stronger performance on multi-step, structured problems. It is also the axis most teams plan around first and most often plan around wrong, because Deep Think is not a default capability for every tier or every API path.
The first decision in any 3.1 migration is whether the workloads you want to upgrade actually need Deep Think, and if so, whether your provisioning gives you reliable access to it. Confirm the tier-level availability before designing dependent workloads. The cost and latency of Deep Think are both noticeably above the standard 3.1 reasoning path, so even when it is available, the economic question is whether each workload earns the step-up.
Standard reasoning only
Deep Think not available at this tier. Workloads requiring extended deliberation should either be upgraded in tier or routed to a different provider for the reasoning-heavy slice while keeping the rest on Gemini 3.1 standard.
Route around Deep ThinkDeep Think metered
Deep Think available with metered access — usage caps, per-minute throughput limits, or both. Suitable for moderate-volume workloads where the reasoning lift earns the cost. Plan for fallback to standard mode under burst load.
Use with fallbackDeep Think first-class
Deep Think available as a first-class mode. Use for the reasoning-heavy core of the workload, route lower-complexity sub-tasks to the standard mode in the same call sequence to keep the cost envelope predictable.
Premium tierMulti-vendor routing
If Deep Think feasibility is a hard requirement and tier upgrade is not viable, route the reasoning-heavy slice to a different frontier model and keep general Gemini workloads on 3.1. Document the routing decision and the eval that backs it.
Mixed routingThe most common pattern we see in client repos: a workload that currently runs against Gemini 3 Pro's standard reasoning is assumed to be a Deep Think candidate, the migration is planned around Deep Think, and then provisioning turns out to be metered or entry-tier-only. Budget the access verification at the very start of the migration — it changes the rollout shape for the other three axes.
"The first question on any 3.1 migration is not what Deep Think does for the workload — it is whether your tier actually gives you Deep Think reliably."— Production lesson · Digital Applied AI engagements
03 — PricingTier shifts and the cost-curve impact.
3.1's tier rebalance is the axis most teams underestimate. Per-token rates moved, the spread between tiers widened for some workloads and narrowed for others, and Deep Think's premium reshapes the math when it enters the picture. The org-wide cost curve depends on the workload mix; the per-workload curves can move in opposite directions on the same release.
The horizontal bars below are illustrative of the directional shifts our migrations have hit in client repos. The actual numbers depend on the specific tier, region, and workload — re-evaluate them against current published Google AI pricing before committing to a routing decision.
Per-workload cost ratio · 3.1 modes vs 3 Pro baseline
Illustrative directional shifts — verify against current Google AI pricing before routing decisions.Two patterns emerge in practice. For bulk classification and short-context generation, 3.1 standard at the entry tier is materially cheaper than 3 Pro for comparable quality — the cost-per-correct-answer drops even without Deep Think. For long-context retrieval and structured reasoning, Deep Think on a mid or premium tier earns its premium when the workload genuinely needs the extra deliberation; for shallow reasoning, paying the Deep Think premium is the most common pricing pathology we see.
The right move is to model the cost-per-correct-answer for each workload class separately. The 3.1 release is one of the few where a single org-wide swap can be net-negative on cost even though the sticker price moved in your favor — the Deep Think premium compounds quickly for workloads that did not need it. If you need help structuring the per-workload cost-curve audit, our AI transformation engagements run this as a standalone deliverable before the migration window.
04 — Safety FiltersNew categories, thresholds, override patterns.
The safety-filter overhaul is the quietest of the four axes and the source of the most embarrassing post-cut-over regressions. The change is two-fold: new filter categories were added to the default policy set, and default thresholds for several existing categories were moved. The net effect for many workloads is a higher refusal rate on the tail of production traffic — the prompts that passed on 3 Pro and refuse on 3.1 are almost never the prompts that show up in synthetic evals.
The categories that bite most often in production:
Existing categories
Default thresholds for harassment, hate speech, dangerous content, and sexually explicit content were re-balanced on 3.1. Same prompt can pass on 3 Pro and refuse on 3.1. Override patterns from your previous safety settings may not map cleanly — audit the threshold mapping explicitly.
Threshold driftNew filter categories
3.1 introduces additional filter categories with their own defaults. Workloads handling civic, political, or sensitive editorial content may hit refusals on subject matter that was unmoderated on 3 Pro. Confirm the active category list before cut-over.
New surfaceRefusal-rate auditing
Sample at least 500 representative production prompts and replay them against 3.1 with your intended override settings. Diff the refusal verdicts against 3 Pro. The right number to track is delta-refusal-rate per workload — not absolute refusal rate.
Pre cut-over gateThe override patterns also need re-thinking. On 3 Pro, many teams relied on a single low-threshold override across all categories. On 3.1, that pattern is brittle — the new categories may require explicit settings rather than inheriting a global default, and the re-balanced thresholds mean the previous "safe" level for a category is now the "medium" level for equivalent behavior. Re-derive the override settings from policy, not from the previous config file.
The practical sequence: audit refusal rates first, derive the new override settings from the audit findings second, regression-test the safety behavior third. Skipping the audit and hand-porting the old override settings is the most reliable way to ship a quiet production regression.
05 — Tool UseSchema diff and the function-calling surface.
The tool-use schema diff is the axis that breaks agents quietly. The headline shape — tool definitions in JSON Schema-like form, invocation via function-call response, structured argument passing — is preserved. The details are not. Argument typing has tightened in several places, the response shape has new metadata fields, and a handful of legacy fields that 3 Pro accepted permissively are now rejected with a validation error rather than silently coerced.
For simple tools (one or two arguments, primitive types, no nested objects), the migration is almost a no-op. For non-trivial tools — anything with nested objects, optional fields with default values, or enum-typed arguments — expect to migrate each tool definition explicitly. The drop-in compatibility assumption is the most common source of silent agent breakage in the first week after cut-over.
Treat each tool as a migration unit
Walk the tool definitions one at a time. Re-validate against the 3.1 schema spec. Flag any nested objects, optional fields, or enums for explicit re-derivation. Do not bulk-port tool definitions.
Audit per-toolRecorded-call replay
Replay a sample of recorded tool calls from production 3 Pro traffic against 3.1 in a sandboxed test fixture. Diff the argument shapes and response payloads. Surfaces the silent coercion failures that schema validation alone misses.
Replay before flipSide-by-side tool versions
If the agent runs against multiple model versions during the migration window, version the tool definitions explicitly. tool_v1 for 3 Pro, tool_v2 for 3.1. Avoids the worst case — a single tool definition that drifts under both models.
Versioned toolsPre-call validation
Add a strict client-side schema validation step before any tool call is sent to 3.1. Catches argument-typing mismatches at the call site rather than as opaque validation errors from the model. The week-one debugging cost reduction is significant.
Validate at call siteThe agent-side change is also worth flagging: 3.1's function-call response shape includes new metadata fields that agent scaffolding built tightly around the 3 Pro shape will silently drop. Most scaffolding handles this gracefully because the fields are additive, but any custom parser that asserts on the exact field set will need to be updated. Run the scaffolding through the same recorded-call replay as the tool definitions themselves.
06 — Phased RolloutAssess → shadow-test → cut over → retire.
The migration sequence we run for clients has four phases. Each phase has an explicit exit gate; you do not advance to the next phase until the gate is met. The total elapsed time is typically one to two sprints per workload, depending on the workload's risk profile and traffic volume.
Assess
tier, pricing, schema, filter inventoryConfirm Deep Think availability at your tier. Model the cost-per-correct-answer for each workload class. Inventory every tool definition. Capture a representative sample of production prompts for refusal-rate baseline.
Exit: capability + cost + scope confirmedShadow- test
mirror traffic · diff outputsMirror a slice of production traffic to 3.1 in a shadow path. Diff outputs against 3 Pro for quality, refusal rate, tool-call success, latency. Run for at least one full traffic cycle. Surfaces the long-tail regressions synthetic evals miss.
Exit: regression gates greenCut over
per-workload routing flipFlip the routing for one workload at a time. Monitor refusal rate, tool-call success rate, cost-per-call, and latency in real time for the first 24 to 48 hours. Keep the 3 Pro endpoint warm for rollback.
Exit: one full traffic cycle cleanRetire
decommission 3 Pro endpointsOnce all migrated workloads have run cleanly on 3.1 for at least one billing cycle, decommission the 3 Pro endpoints, archive the tool-version-1 definitions, and update internal docs. Keep the migration audit trail for at least one quarter.
Exit: zero 3 Pro traffic for a weekTwo anti-patterns worth naming explicitly. The first is the single-window cut-over — flipping all workloads in one window because the migration feels coordinated. It is not coordinated; it is one large blast radius. Per-workload flips with at least 24 hours between them is the right pattern. The second is the premature retirement of the 3 Pro endpoints. Keep them warm for at least one full billing cycle after the last workload migrates — the cost of a few days of unused endpoints is trivial against the cost of a rollback you cannot execute because the endpoint is already gone.
For teams running production agents on Gemini today, the migration pattern here is the same shape we apply to any frontier-model upgrade window. The provider-specific gotchas change; the phased sequence does not. Our Claude Opus 4.6 to 4.7 migration playbook applies the same four-phase structure to a different provider — useful read if your stack spans both vendors.
07 — Common PitfallsFour ways the migration trips.
Across the migrations we have run, four failure modes account for most of the post-cut-over incidents. None of them are exotic; each is the predictable result of skipping one of the gates in the phased rollout above. Name them explicitly in your migration plan and most of them stop happening.
- The Deep Think assumption. A workload is designed around Deep Think before tier availability is confirmed. Mitigation: verify provisioning in phase 1, before any architecture work depends on it.
- The org-wide cost-curve swap. Cost modeling is done at the org level rather than per-workload, the Deep Think premium gets averaged across workloads that did not need it, and the actual production bill comes in higher than the projection. Mitigation: model per workload class.
- The week-two refusal-rate regression. Synthetic evals pass, cut-over happens, the long tail of real customer prompts starts hitting the re-balanced filters two days later, support tickets spike. Mitigation: replay 500+ representative real prompts against 3.1 before flipping production routing.
- The silent tool-call breakage. A tool with a nested-object argument gets ported without re-validation; the new schema rejects it cleanly half the time and silently coerces it the other half, and the agent output drifts in ways that are hard to attribute. Mitigation: per-tool audit, recorded-call replay, strict pre-call validation at the agent layer.
For broader context on how pricing alone reshapes provider choice even before considering capability differences, our LLM API pricing index for Q2 2026 tracks per-token rates across the major providers and the cost-curve dynamics teams actually run into.
Gemini 3.1 is a per-workload decision — and a per-workload migration.
The Gemini 3.1 release is one of those rare upgrades where the sensible default — flip the model string, ship the change — is the most expensive thing a team can do. Four axes moved at once. None of them break the same workload, and that is precisely why the single-window cut-over leaves regressions scattered across the production surface, each one showing up two to three days later when the long tail of real traffic finally hits the new thresholds.
The fix is the discipline of the four-phase rollout. Assess Deep Think availability and per-workload cost-curve before any architecture work. Build the eval harness and shadow-test before any routing flip. Cut over one workload at a time with the 3 Pro endpoints kept warm. Retire only after a full billing cycle of clean production runtime. Done that way, the migration is uneventful — which is exactly the right adjective for a frontier model upgrade.
The broader pattern is worth holding on to. Frontier model releases are increasingly multi-axis — one mode change, one pricing change, one safety change, one schema change, in the same release window — and the migration discipline has to scale with that shape. Per-workload eval harnesses, recorded-call replay, and phased per-workload cut-overs are no longer special; they are the baseline operating practice for any team running production AI.