AI vendor risk assessment is the procurement discipline that distinguishes a defensible signing decision from a brochure read — and in 2026 it is the artefact that auditors, board members, and future engineering leaders will read three years from now to understand why your AI stack looks the way it does. The 50-point template below organises the assessment across five domains: security posture, compliance certifications, data handling, model provenance, and operational resilience. Each point is scored, severity-weighted, and gated.
The structure of this assessment matters more than any individual point on the list. The template is built so the output is a single severity-weighted score per vendor, a documented evidence trail per point, and a small set of hard go/no-go gates that disqualify a vendor regardless of total score. Together those three artefacts convert vendor risk from a relationship judgement into a reviewable procurement decision — which is the actual test the assessment is designed to survive.
This guide walks each of the five domains in turn — ten points per domain — followed by the severity weighting scheme and the scorecard gates. It assumes you have already filtered your shortlist to the top three to five vendors through the upstream scorecard described in our Stage 4 vendor selection playbook; risk assessment is the next gate in the same procurement track.
- 01AI vendors outlive their executive sponsors.Average AI exec tenure runs around eighteen months; vendor contracts run twenty-four or longer. The 50-point assessment is the institutional memory that keeps the procurement decision defensible after the original sponsor moves on.
- 02Compliance certificates are necessary, not sufficient.SOC2 and ISO 27001 are the floor — they tell you the vendor has a security programme, not that the programme covers AI-specific failure modes. Treat certificates as gates, then assess the substance behind them on points outside the cert scope.
- 03Training opt-out is the easy-to-miss control.The single most-skipped clause in 2026 AI contracts. Without an explicit training opt-out, every prompt and document you send the vendor becomes potential training data, with no consent and no compensation. Insist on it; vendors that refuse are filtering themselves out.
- 04Model provenance distinguishes serious vendors.Training data composition, RLHF process, eval coverage, and watermarking are the four signals that separate vendors who treat the model as a serious engineered artefact from vendors who treat it as a black-box capability bolted on for the demo.
- 05Deprecation policy is the under-discussed risk.Capability deprecation is the silent contract failure mode — the vendor stops supporting the feature you signed for, the contract still runs, the cost still bills. A written deprecation policy with notice periods and migration support is the protection your future self will thank you for.
Run the full 50 points against each shortlisted vendor; collect evidence per point in the same document; score severity-weighted; check the gate list; produce a single signed scorecard per vendor. The output goes into the procurement file alongside the scorecard from Stage 4, the RFP responses, the reference call notes, and the contract red-line. That bundle is what survives the executive transition and the renewal audit.
01 — Why Vendor RiskAI vendors outlive most exec sponsors — assess accordingly.
The average tenure of a senior AI executive in 2026 sits around eighteen months — meaningfully shorter than the typical AI vendor contract, materially shorter than the twenty-four-month roadmap window most teams plan against. That asymmetry is the single most under-appreciated feature of vendor risk assessment. The person who signs the contract is rarely the same person who renews it, defends it after a model-update breaks production, or migrates off it when the vendor strategy shifts. The artefact has to survive the person.
That changes how the assessment should be run. The output of a vendor risk assessment is not an opinion — it is a documented severity-weighted scoring artefact with named evaluators per point, an evidence file per point, and a hard go/no-go gate list applied independently of the weighted total. The procurement decision falls out of those artefacts; without them, vendor choice collapses into the relationship in the room on signing day, which is exactly what audit reviewers will flag eighteen months later.
The second reason this matters more than its budget line suggests: vendor lock-in compounds. A choice made casually at signing propagates through implementation, into integration depth, into the data the vendor accumulates about your usage, into the renewal negotiation, and into the cost and complexity of an exit. A deliberate 50-point assessment at signing produces optionality at every subsequent stage; a casual one produces a sequence of expensive renegotiations under duress.
"The vendor decision is downstream of the artefact. Build a severity-weighted scorecard your replacement can read three years from now, and the renewal conversation falls out cleanly."— Digital Applied procurement engagements, on AI vendor assessments
Two failure modes show up reliably when teams skip the artefact work. The first is the relationship-driven approval — a vendor cleared because the AI lead worked with their CTO at a previous company, with the risk write-up produced after the decision to justify it. The relationship may be a fine signal, but a write-up authored after the decision is no longer an audit trail; it is justification. The second is the demo-driven approval — a vendor cleared because the demo was polished, even though demo capability is the easiest part of the offering to replicate and the operational reality is closer to a different vendor's. The 50-point template, the gates, and the severity weighting are designed to neutralise both failure modes.
Treat the 50-point scorecard as a signing-day artefact, not a post-signing audit. Risk discovered at signing is leverage; risk discovered six months in is a problem. Run the assessment before the contract goes to legal, not after.
02 — Security PostureAuth, encryption, secrets, audit logs — ten points.
Security posture is the first of the five domains and the one most likely to be over-relied on as a proxy for the rest. The ten points below cover authentication, encryption in transit and at rest, secrets handling, audit logging, network controls, and the incident response programme. Score each one against the evidence the vendor produces in response to the questionnaire, not against the marketing copy on the security page.
The grid that follows is the operating shape of the ten points. Five control families, each with multiple sub-points, scored from 1 to 5 with severity weights applied as described in Section 07. Evidence per point lives in the procurement file; unverified claims score below the floor.
Authentication & identity
SSO, MFA, SCIM, IdP integrationSAML / OIDC SSO with named IdP support, mandatory MFA on admin accounts, SCIM provisioning for deprovisioning at scale, named-IdP integration tested. Two sub-points cover service-account hygiene and machine-identity controls.
Critical · 2 pointsEncryption everywhere
TLS 1.2+, AES-256, customer keysTLS 1.2 or higher in transit, AES-256 at rest, customer-managed encryption keys available for regulated workloads. Sub-points cover backup encryption, vendor key rotation cadence, and the cryptographic agility story.
Critical · 2 pointsSecrets & credentials
Vault integration, rotation, auditProduction secrets stored in a managed vault (HashiCorp Vault, AWS Secrets Manager, equivalent), rotation cadence documented and enforced, full audit trail per credential. Sub-points cover API key scoping and customer-side credential hygiene.
High · 2 pointsAudit logs & SIEM-ready
Retention 12mo+, SIEM exportFull audit log of authentication, configuration changes, and data access, retained 12 months minimum, exportable to customer SIEM in standard formats (JSON, CEF, syslog). Sub-points cover tamper resistance and log integrity verification.
High · 2 pointsIncident response
Documented runbook, tested annuallyDocumented incident response plan, tested via tabletop annually, customer notification within 72 hours for confirmed data incidents, named security contact. Sub-points cover post-incident review process and credit policy.
High · 2 pointsThe most-skipped sub-points across this domain are typically audit-log retention and SIEM export support. Vendors will often offer audit logs in their own console with no export pathway, which makes the logs useful to the vendor's support team but useless to your own security operations centre. A SIEM-ready export — JSON or CEF, ideally with a published schema — is the difference between a security feature you can build into your own detection and response stack and a security feature you can only inspect when you log into the vendor's UI.
The under-discussed point in the encryption family is cryptographic agility. Most vendors meet the headline AES-256 and TLS 1.2 requirement; few publish what their migration plan looks like when post-quantum cryptography enters the mainstream, or when an algorithm gets deprecated. A vendor that cannot describe their crypto-migration story is a vendor whose security posture is frozen at today's standards, which means it is degrading every year.
A vendor failing any one of the five Control 01 to Control 05 families at the floor (score of 1) is a hard disqualifier regardless of weighted total. Security posture below the floor on any control family is the kind of signing-day finding that no commercial discount makes acceptable. The gate is the artefact you point to when the executive sponsor pushes back on the disqualification.
03 — Compliance CertsSOC2, ISO 27001, HIPAA, regional.
Compliance certifications are the second domain — ten points covering the certification programmes themselves, the sector-specific overlays (healthcare, finance, public sector), and the regional regulatory regimes (EU GDPR, UK, California, Singapore, Australia). The framing matters: certifications are a floor, not a ceiling. SOC2 Type II tells you the vendor has a security programme audited annually against the AICPA framework; it does not tell you the programme covers AI-specific failure modes, training-data provenance, or model-update governance.
Use the choice matrix below to assess each shortlisted vendor against the four certification families that matter most in 2026 AI procurement. Sector-specific certifications (HIPAA, HITRUST, PCI-DSS, FedRAMP, IL5) layer on top depending on the workload class.
SOC2 Type II
5 = current SOC2 Type II report within 12 months, shareable under NDA, all five trust services criteria (Security, Availability, Processing Integrity, Confidentiality, Privacy) in scope. 3 = Type II current but limited TSC scope, or report older than 12 months. 1 = Type I only, or no current report.
Baseline gateISO 27001
5 = current ISO 27001 certification with 27017 (cloud) and 27018 (PII) extensions, ISO 42001 (AI management system) for AI-specific scope. 3 = base 27001 only, no AI-specific extension. 1 = lapsed or pending certification.
International gateSector overlays
5 = sector-appropriate certification current (HIPAA + HITRUST for healthcare, PCI-DSS for payments, FedRAMP for federal, IL5 for defense). 3 = sector certification pending or in audit. 1 = no sector-specific certification despite workload requiring it.
Workload-specificRegional regimes
5 = active GDPR Article 28 DPA, UK transfer mechanism documented, California CCPA agent, Singapore PDPA mechanism, Australia Privacy Act covered. 3 = EU-only coverage or partial regional mapping. 1 = no documented regional regime coverage.
Geography gateAudit cadence
5 = annual external audit on every certification, results available to customers under NDA, gap analyses against new frameworks (NIST AI RMF, EU AI Act) published. 3 = audit cadence documented but reports not shareable. 1 = self-attestation rather than external audit.
Programme maturitySub-processor register
5 = public sub-processor list, change-notification within 30 days, customer right to object documented. 3 = sub-processor list available on request, notification cadence ad hoc. 1 = sub-processors undisclosed or change without notification.
Transparency gatePen-test cadence
5 = annual third-party penetration test, results shareable under NDA, AI-specific red-teaming included. 3 = annual pen-test but generic, no AI-specific testing. 1 = no recent pen-test or test results not shareable.
Adversarial validationAI-specific frameworks
5 = documented NIST AI RMF mapping, EU AI Act risk-tier alignment, ISO 42001 in-scope. 3 = self-mapped to one framework, no third-party verification. 1 = no AI-specific framework mapping despite AI being the core product.
AI maturity gateThe two AI-specific compliance signals worth weighting heavily in 2026 are ISO 42001 scope (the new AI management system standard) and documented EU AI Act risk-tier alignment. ISO 42001 is the first internationally-recognised certification that addresses AI-specific governance — training-data management, model lifecycle, AI incident response — rather than treating AI as a generic IT system. Vendors who hold ISO 42001 within their AI product scope are signalling that they treat AI governance as a distinct discipline from general information security; vendors who do not are not.
The EU AI Act risk-tier alignment matters even for non-European buyers because the regulation establishes the most concrete framework for AI risk classification globally — general-purpose AI obligations, high-risk system requirements, transparency obligations, and prohibited practices. A vendor who can articulate which risk tier their product falls into, and what obligations attach, is a vendor who has done the regulatory homework. A vendor who treats the EU AI Act as a problem for someone else's lawyer is a vendor whose compliance posture is reactive rather than programmatic.
Compliance certificates are necessary, not sufficient. SOC2 and ISO 27001 tell you the vendor has a programme; the other forty-two points in the assessment tell you whether the programme covers the things that matter for AI procurement specifically. Use the certificates as gates; assess the substance on the points behind them.
04 — Data HandlingResidency, retention, training opt-out.
Data handling is the third domain — ten points covering data residency, retention policy, deletion process, training opt-out, output-rights ownership, and the chain of custody from prompt submission to model output. This is the domain where AI vendor risk diverges most sharply from generic SaaS risk: the data submitted to an AI vendor is not just data being processed, it is data that may be used to train future models, may be retained as part of a fine-tuning dataset, and may surface in outputs to other customers.
The training opt-out clause is the single most-skipped control in 2026 AI procurement. Without an explicit opt-out in the contract — not just in the marketing copy — every prompt and document submitted to the vendor is potentially training data for future model versions. The clause has to be enumerated in the contract, named per data category, and verifiable.
Data residency options
Multiple geographic regions for storage and processing — EU, UK, US, APAC at minimum. Customer chooses region at provisioning, vendor commits in contract.
SovereigntyRetention policy named
Per data category (prompts, outputs, logs, fine-tuning data, audit data), retention duration named in days, deletion process documented, deletion verifiable on request.
LifecycleTraining opt-out
Customer data must not be used for model training without explicit opt-in. Enumerated per data category. Verifiable. The single most-skipped clause in 2026.
Non-negotiableOutput rights
Outputs owned by customer; vendor licence to operate the service only. No vendor right to use outputs for marketing, case studies, or training without separate opt-in.
IP gateCross-tenant isolation
Customer prompts and outputs cannot surface to other tenants. Isolation model documented (logical / physical), tested annually, results shareable.
Leakage preventionDeletion verifiable
Customer right to delete on request, deletion completed within named SLA (30 days typical), deletion verifiable via signed attestation. Includes backup expiry.
GDPR Article 17Sub-processor data flow
Where customer data flows to sub-processors (cloud infrastructure, fine-tuning vendors, evaluation services), the path is documented and the same opt-out applies downstream.
Chain of custodyLogging discipline
Vendor's own logs of customer data access are minimised, scoped to support and security functions only, retention bounded, access audit trail available.
Internal accessBackup geography
Backup storage geography matches primary residency commitment, encryption keys customer-managed where required, restoration tested annually.
Disaster recoveryExport tooling
Customer right to export all submitted data in standard formats on demand, no per-export fees, no data-format lock-in. The migration insurance policy.
Exit rampTwo operational notes on this domain. First, the training opt-out clause has to be enumerated rather than blanket. A single line saying "customer data will not be used for model training" is weaker than a clause that enumerates the data categories (prompts, documents, outputs, feedback signals, fine-tuning data, telemetry) and applies the opt-out to each named category. The enumerated version is harder to argue around at renewal and produces clearer audit evidence.
Second, deletion verifiability is more important than the deletion SLA itself. A vendor who deletes within 30 days but cannot produce a signed deletion attestation is functionally indistinguishable from a vendor who deletes within 90 days with attestation — and worse than a vendor who deletes within 45 days with attestation. The attestation is the artefact that satisfies the GDPR Article 17 right-to-erasure obligation; without it, the customer cannot demonstrate compliance to their own regulators.
"Every training-data clause you skip at signing becomes a hostage negotiation at renewal. Enumerate the data categories, name the opt-out, verify the implementation."— Standard advice from our AI procurement engagements
05 — Model ProvenanceTraining data, RLHF, evals, watermarking.
Model provenance is the fourth domain — ten points covering the model's lifecycle from training corpus through fine-tuning, evaluation, deployment, and ongoing monitoring. Provenance assessment is the domain that most distinguishes serious AI vendors from less serious ones; the points below cannot be bluffed in an RFP response, and the vendors who answer them substantively are the vendors whose models you can defend in an audit.
The four signals that matter most in this domain are training data composition disclosure, RLHF process documentation, eval harness coverage, and output watermarking support. Together they tell you whether the vendor treats the model as a serious engineered artefact or as a black-box capability bolted on for the demo.
Training corpus disclosed
Public summary of training-data composition (web crawl scope, licensed corpora, synthetic data proportion, language coverage), with cutoff date.
Foundation transparencyIP-clean training data
Affirmative statement that training corpus excludes copyrighted material licensed expressly for training, with indemnification against third-party IP claims on outputs.
IP exposureRLHF process documented
Reinforcement learning from human feedback process described: rater pool composition, rubric, inter-rater agreement, refusal behaviour calibration.
Alignment maturityEval harness public
Public eval results on standard benchmarks (MMLU, GPQA, HumanEval, MMLU-Pro, MATH, AIME) plus AI-safety evals (HHH, TruthfulQA, BBQ). Methodology documented.
Capability honestyRed-team programme
Internal red-teaming programme with documented test categories (jailbreaks, prompt injection, data extraction), bug bounty pathway, annual external red-team.
Adversarial resistanceOutput watermarking
Cryptographic watermark in generated text or images where applicable, watermark verification tool published, watermark resilience tested against paraphrasing.
AttributionModel card per release
Model card published per model release with intended use, limitations, training-data summary, eval results, and known failure modes. Updated with each version.
Documentation disciplineHallucination calibration
Documented hallucination rate on standard benchmarks, calibration evaluation, confidence-score availability in API responses. The honesty axis.
Output reliabilityBias evaluation
Published bias evaluation across protected categories (gender, ethnicity, age, geography, language), with documented mitigation steps and residual-bias acknowledgement.
Fairness signalCapability scope statement
Explicit statement of capabilities the model is fit for and capabilities it is not — including domain restrictions and high-risk use cases (medical advice, legal, financial).
Scope disciplineThe training-data disclosure question is where vendors most often retreat into vague generalities — "diverse web corpus plus licensed datasets" is the typical answer. Push for the specifics: roughly what proportion is licensed corpora versus web crawl, what is the cutoff date per data source, what is the language distribution, what synthetic-data proportion was used, and what filtering pipeline was applied. Vendors who can answer those questions clearly are vendors with a serious data governance programme; vendors who cannot are vendors whose training pipeline is either undocumented or considered too sensitive to share. Neither of those is acceptable for AI procurement at scale.
On watermarking, the technology is maturing rapidly but adoption is uneven. The 2026 frontier is cryptographic watermarks that survive light paraphrasing on text and standard image transformations on visuals — and that publish a verification tool so the watermark can be checked outside the vendor's infrastructure. Vendors without any watermarking story are signalling that they have not engaged seriously with content-attribution challenges, which is a category of risk that is rising fast as regulators push for AI-generated content disclosure obligations.
Ask the vendor for the model card for the model version you will be deployed against. A current, detailed model card with named limitations and known failure modes is the single best signal of a mature AI vendor. A vendor who refuses, who produces a card with no failure modes listed, or who produces a card more than six months out of date for a currently-shipping model — is a vendor whose model-governance programme is below the standard you should accept.
06 — OperationalSLA, incident response, deprecation policy.
Operational resilience is the fifth and final domain — ten points covering service availability SLAs, incident response, change management, customer support, capability deprecation policy, and the financial-stability signals that predict whether the vendor will be around at the end of your contract term. The under-discussed point in this domain is deprecation policy: the written commitment around what happens when the vendor decides to stop supporting a capability you have built against.
Use the modes grid below to assess each of the five operational sub-domains; sub-points within each cover the more granular questions (named TAM availability, business-review cadence, credit policy details, executive escalation paths).
Service SLA
99.9% / 99.95% uptime, creditsNamed uptime SLA (typically 99.9% for general workloads, 99.95% for production-critical), service-credit formula documented, exclusions enumerated. Per-region SLAs where applicable.
Availability gate · 2 pointsIncident response
P1 ack < 30 min, status page, RCAP1 acknowledgement within 30 minutes, P1 resolution target within 4 hours, public status page, post-incident root-cause analysis within 5 business days. Named escalation contact.
Operational maturity · 2 pointsChange management
Notice 30 days, rollback pathMaterial changes (model updates, API changes, behaviour drift) require 30 days advance notice, rollback path documented, deprecation timelines published. Versioned APIs.
Stability requirement · 2 pointsDeprecation policy
12mo notice, migration supportWritten deprecation policy: minimum 12-month notice for capability deprecation, migration tooling, dedicated migration support, refund-or-credit pathway for contracted capabilities removed before contract end.
Exit insurance · 2 pointsFinancial stability
Funding, customer concentration, audited financialsFunding situation transparent (latest round, runway), customer concentration disclosed (no single customer above 40% of revenue), audited financials available under NDA for enterprise deals.
Survival signal · 2 pointsThe deprecation policy is the under-discussed risk control in most procurement processes. The standard SaaS contract assumes the vendor will continue to support the capabilities you signed for through the contract term; the AI vendor contract has to handle a different reality, where a capability may be deprecated mid-contract because the underlying model has been retired, the vendor has pivoted strategy, or the capability has been folded into a more expensive product tier. A 12-month notice period with named migration support is the baseline; less than that and the contract is exposed to mid-term operational disruption with no recourse.
Financial-stability assessment matters more for AI vendors than for typical SaaS because the AI vendor market is going through a consolidation phase. Vendor acquisitions, strategic pivots, and outright shutdowns are happening more frequently than in mature SaaS categories, and the cost of a vendor going away mid-contract is materially higher because of the integration depth required for production AI workloads. Assess the financial-stability signals seriously; a vendor with eighteen months of runway and a single customer above 40% of revenue is a vendor whose contract you should price down significantly because of the risk premium.
"The deprecation policy is the operational clause your future self will thank you for. Negotiate it when you have signing leverage — at renewal there is none."— Procurement engagements, on operational resilience
Two operational discipline notes. First, the SLA service-credit formula has to produce real consequences. A vendor offering "service credit of 5% of monthly fees" for a multi-day outage is offering a token rather than a remedy; the credit needs to scale with outage duration and severity, with a cap that is high enough to incentivise actual operational investment. A sensible structure caps credits at 30% of monthly fees for severe outages, which is enough to be felt by the vendor's P&L without being existential. Second, the status page has to be public and historically searchable. A status page that only shows the current state and clears history is a status page designed to make the vendor look good rather than to be useful to customers; insist on a 12-month history at minimum.
07 — ScorecardSeverity weighting and go/no-go gates.
The 50 points roll up into a single severity-weighted scorecard per vendor, with explicit go/no-go gates applied independently of the weighted total. The severity weighting scheme below uses four tiers — Critical, High, Medium, Informational — with weights that produce a maximum theoretical score of 100 across the 50 points. Vendors that clear the shortlist threshold (typically 75 of 100) and pass every gate move to contract; vendors that fall below the threshold or fail any gate are disqualified.
Severity weighting matters because not every point on the assessment carries equal procurement consequence. A vendor failing on watermarking support is a different category of risk from a vendor failing on cross-tenant data isolation; the scoring scheme has to reflect that asymmetry.
Severity weighting · 50-point AI vendor assessment
Source: Digital Applied AI vendor risk engagements, 2025-2026The threshold matters as much as the weighting. The default shortlist threshold is 75 of 100 on the weighted total — high enough to filter aggressively, low enough that one or two shortlisted vendors usually clear it. Set the threshold before scoring begins; setting it after produces a moving target that rationalises whatever shortlist the team wanted in the first place. Document the threshold and the weighting scheme in the procurement file as the second artefact after the 50-point scorecard itself.
The gate list applies independently of the score. Any single Critical-tier point scored at the floor (1 out of 5) is a hard disqualifier — auth, encryption, cross-tenant isolation, training opt-out, IP-clean training data. A vendor failing any of those at the floor is disqualified regardless of how well they score on the other 49 points. The gates are non-negotiable because the risks they cover compound across the rest of the portfolio; a single gate failure makes every other strength irrelevant.
The 50-point assessment is annual plus on-event. Run the full assessment at every contract renewal, after any vendor acquisition or strategic pivot, after any material model change (version bump, architecture change, behaviour drift), and after any reported security or AI-safety incident at the vendor. Between assessments, monitor the published trust-page and status-page signals; the assessment is the deliberate gate, the monitoring is the early warning.
The discipline that keeps the assessment honest over time is documenting the baseline and rerunning against it. Store the scored scorecard alongside the contract; on the renewal review, rerun the assessment and diff the new scores against the baseline. Where scores have moved — up or down — note the change, capture the evidence, and update the contract red-line for the renewal cycle. Teams that compound advantage in AI procurement are the teams that rerun the assessment; teams that compound regret are the ones that treat the original signing decision as final.
For teams running their first 50-point assessment, the practical starting point is to score the incumbent vendor first — the one already in production — to calibrate the scoring discipline before the shortlisted alternatives. The incumbent assessment becomes the baseline the alternatives are scored against, and the gaps in the incumbent often reveal which of the 50 points carry the most operational weight in your specific context. From there, the template scales cleanly to the rest of the procurement portfolio. Our AI transformation engagements include the full 50-point assessment as part of the procurement-and-renewal track.
"The signing-day scorecard is the cheapest insurance in the procurement portfolio. Skip it and every clause becomes a hostage negotiation; run it and the renewal conversation falls out cleanly."— Digital Applied procurement engagements, on the 50-point template
Vendor risk is decided at signing — the template prevents regret at renewal.
The 50-point AI vendor risk template is a signing-day artefact first, an audit artefact second, and a renewal artefact third. Five domains — security posture, compliance certifications, data handling, model provenance, operational resilience — ten points per domain, four severity tiers, a clear set of gates, and a single severity-weighted scorecard at the end. The artefact is what survives the executive transition, the renewal cycle, and the inevitable board audit that arrives sometime in between.
The discipline that makes the template work is the willingness to disqualify on a single Critical-tier failure regardless of how strong the rest of the assessment looks. Compliance certs are a floor, not a ceiling. Training opt-out is the easy-to-miss control that becomes a hostage negotiation at renewal if skipped. Model provenance distinguishes serious vendors from less serious ones, and capability deprecation policy is the operational clause your future self will thank you for negotiating at signing. Casual application of this template produces a procurement decision that ages badly; deliberate application produces a scorecard that compounds advantage through the rest of the AI portfolio.
The next concrete step is short. Pick the AI vendor your team is closest to signing — or the AI vendor whose contract is closest to renewal — and run the 50-point assessment against them in the next two days. Score severity-weighted, apply the gates, document the evidence per point, and produce the scorecard. By the time the contract goes to legal, the artefact set is documented, the gates are checked, and the renewal conversation eighteen months from now has a baseline to work against. That is the cadence the template is designed to enable.