SYS/2026.Q1Agentic SEO audits delivered in 72 hoursSee how →
BusinessMethodology17 min readPublished May 15, 2026

Five domains, fifty points — the vendor risk template that distinguishes a defensible signing decision from a brochure read.

AI Vendor Risk Assessment: 50-Point Template 2026

A 50-point AI vendor risk template organised across five domains — security posture, compliance certifications, data handling, model provenance, and operational resilience — with severity weighting and explicit go/no-go gates. The artefact you sign against, defend at audit, and rerun at renewal.

DA
Digital Applied Team
Senior strategists · Published May 15, 2026
PublishedMay 15, 2026
Read time17 min
SourcesField engagements + RFP archives
Total points
50
weighted by severity
Risk domains
5
security to operations
Severity tiers
4
critical to informational
Recommended cadence
Annual+
on event triggers

AI vendor risk assessment is the procurement discipline that distinguishes a defensible signing decision from a brochure read — and in 2026 it is the artefact that auditors, board members, and future engineering leaders will read three years from now to understand why your AI stack looks the way it does. The 50-point template below organises the assessment across five domains: security posture, compliance certifications, data handling, model provenance, and operational resilience. Each point is scored, severity-weighted, and gated.

The structure of this assessment matters more than any individual point on the list. The template is built so the output is a single severity-weighted score per vendor, a documented evidence trail per point, and a small set of hard go/no-go gates that disqualify a vendor regardless of total score. Together those three artefacts convert vendor risk from a relationship judgement into a reviewable procurement decision — which is the actual test the assessment is designed to survive.

This guide walks each of the five domains in turn — ten points per domain — followed by the severity weighting scheme and the scorecard gates. It assumes you have already filtered your shortlist to the top three to five vendors through the upstream scorecard described in our Stage 4 vendor selection playbook; risk assessment is the next gate in the same procurement track.

Key takeaways
  1. 01
    AI vendors outlive their executive sponsors.Average AI exec tenure runs around eighteen months; vendor contracts run twenty-four or longer. The 50-point assessment is the institutional memory that keeps the procurement decision defensible after the original sponsor moves on.
  2. 02
    Compliance certificates are necessary, not sufficient.SOC2 and ISO 27001 are the floor — they tell you the vendor has a security programme, not that the programme covers AI-specific failure modes. Treat certificates as gates, then assess the substance behind them on points outside the cert scope.
  3. 03
    Training opt-out is the easy-to-miss control.The single most-skipped clause in 2026 AI contracts. Without an explicit training opt-out, every prompt and document you send the vendor becomes potential training data, with no consent and no compensation. Insist on it; vendors that refuse are filtering themselves out.
  4. 04
    Model provenance distinguishes serious vendors.Training data composition, RLHF process, eval coverage, and watermarking are the four signals that separate vendors who treat the model as a serious engineered artefact from vendors who treat it as a black-box capability bolted on for the demo.
  5. 05
    Deprecation policy is the under-discussed risk.Capability deprecation is the silent contract failure mode — the vendor stops supporting the feature you signed for, the contract still runs, the cost still bills. A written deprecation policy with notice periods and migration support is the protection your future self will thank you for.
How to use this template

Run the full 50 points against each shortlisted vendor; collect evidence per point in the same document; score severity-weighted; check the gate list; produce a single signed scorecard per vendor. The output goes into the procurement file alongside the scorecard from Stage 4, the RFP responses, the reference call notes, and the contract red-line. That bundle is what survives the executive transition and the renewal audit.

01Why Vendor RiskAI vendors outlive most exec sponsors — assess accordingly.

The average tenure of a senior AI executive in 2026 sits around eighteen months — meaningfully shorter than the typical AI vendor contract, materially shorter than the twenty-four-month roadmap window most teams plan against. That asymmetry is the single most under-appreciated feature of vendor risk assessment. The person who signs the contract is rarely the same person who renews it, defends it after a model-update breaks production, or migrates off it when the vendor strategy shifts. The artefact has to survive the person.

That changes how the assessment should be run. The output of a vendor risk assessment is not an opinion — it is a documented severity-weighted scoring artefact with named evaluators per point, an evidence file per point, and a hard go/no-go gate list applied independently of the weighted total. The procurement decision falls out of those artefacts; without them, vendor choice collapses into the relationship in the room on signing day, which is exactly what audit reviewers will flag eighteen months later.

The second reason this matters more than its budget line suggests: vendor lock-in compounds. A choice made casually at signing propagates through implementation, into integration depth, into the data the vendor accumulates about your usage, into the renewal negotiation, and into the cost and complexity of an exit. A deliberate 50-point assessment at signing produces optionality at every subsequent stage; a casual one produces a sequence of expensive renegotiations under duress.

"The vendor decision is downstream of the artefact. Build a severity-weighted scorecard your replacement can read three years from now, and the renewal conversation falls out cleanly."— Digital Applied procurement engagements, on AI vendor assessments

Two failure modes show up reliably when teams skip the artefact work. The first is the relationship-driven approval — a vendor cleared because the AI lead worked with their CTO at a previous company, with the risk write-up produced after the decision to justify it. The relationship may be a fine signal, but a write-up authored after the decision is no longer an audit trail; it is justification. The second is the demo-driven approval — a vendor cleared because the demo was polished, even though demo capability is the easiest part of the offering to replicate and the operational reality is closer to a different vendor's. The 50-point template, the gates, and the severity weighting are designed to neutralise both failure modes.

The signing-day rule

Treat the 50-point scorecard as a signing-day artefact, not a post-signing audit. Risk discovered at signing is leverage; risk discovered six months in is a problem. Run the assessment before the contract goes to legal, not after.

02Security PostureAuth, encryption, secrets, audit logs — ten points.

Security posture is the first of the five domains and the one most likely to be over-relied on as a proxy for the rest. The ten points below cover authentication, encryption in transit and at rest, secrets handling, audit logging, network controls, and the incident response programme. Score each one against the evidence the vendor produces in response to the questionnaire, not against the marketing copy on the security page.

The grid that follows is the operating shape of the ten points. Five control families, each with multiple sub-points, scored from 1 to 5 with severity weights applied as described in Section 07. Evidence per point lives in the procurement file; unverified claims score below the floor.

Control 01
Authentication & identity
SSO, MFA, SCIM, IdP integration

SAML / OIDC SSO with named IdP support, mandatory MFA on admin accounts, SCIM provisioning for deprovisioning at scale, named-IdP integration tested. Two sub-points cover service-account hygiene and machine-identity controls.

Critical · 2 points
Control 02
Encryption everywhere
TLS 1.2+, AES-256, customer keys

TLS 1.2 or higher in transit, AES-256 at rest, customer-managed encryption keys available for regulated workloads. Sub-points cover backup encryption, vendor key rotation cadence, and the cryptographic agility story.

Critical · 2 points
Control 03
Secrets & credentials
Vault integration, rotation, audit

Production secrets stored in a managed vault (HashiCorp Vault, AWS Secrets Manager, equivalent), rotation cadence documented and enforced, full audit trail per credential. Sub-points cover API key scoping and customer-side credential hygiene.

High · 2 points
Control 04
Audit logs & SIEM-ready
Retention 12mo+, SIEM export

Full audit log of authentication, configuration changes, and data access, retained 12 months minimum, exportable to customer SIEM in standard formats (JSON, CEF, syslog). Sub-points cover tamper resistance and log integrity verification.

High · 2 points
Control 05
Incident response
Documented runbook, tested annually

Documented incident response plan, tested via tabletop annually, customer notification within 72 hours for confirmed data incidents, named security contact. Sub-points cover post-incident review process and credit policy.

High · 2 points

The most-skipped sub-points across this domain are typically audit-log retention and SIEM export support. Vendors will often offer audit logs in their own console with no export pathway, which makes the logs useful to the vendor's support team but useless to your own security operations centre. A SIEM-ready export — JSON or CEF, ideally with a published schema — is the difference between a security feature you can build into your own detection and response stack and a security feature you can only inspect when you log into the vendor's UI.

The under-discussed point in the encryption family is cryptographic agility. Most vendors meet the headline AES-256 and TLS 1.2 requirement; few publish what their migration plan looks like when post-quantum cryptography enters the mainstream, or when an algorithm gets deprecated. A vendor that cannot describe their crypto-migration story is a vendor whose security posture is frozen at today's standards, which means it is degrading every year.

The signing-day security gate

A vendor failing any one of the five Control 01 to Control 05 families at the floor (score of 1) is a hard disqualifier regardless of weighted total. Security posture below the floor on any control family is the kind of signing-day finding that no commercial discount makes acceptable. The gate is the artefact you point to when the executive sponsor pushes back on the disqualification.

03Compliance CertsSOC2, ISO 27001, HIPAA, regional.

Compliance certifications are the second domain — ten points covering the certification programmes themselves, the sector-specific overlays (healthcare, finance, public sector), and the regional regulatory regimes (EU GDPR, UK, California, Singapore, Australia). The framing matters: certifications are a floor, not a ceiling. SOC2 Type II tells you the vendor has a security programme audited annually against the AICPA framework; it does not tell you the programme covers AI-specific failure modes, training-data provenance, or model-update governance.

Use the choice matrix below to assess each shortlisted vendor against the four certification families that matter most in 2026 AI procurement. Sector-specific certifications (HIPAA, HITRUST, PCI-DSS, FedRAMP, IL5) layer on top depending on the workload class.

Cert 01
SOC2 Type II

5 = current SOC2 Type II report within 12 months, shareable under NDA, all five trust services criteria (Security, Availability, Processing Integrity, Confidentiality, Privacy) in scope. 3 = Type II current but limited TSC scope, or report older than 12 months. 1 = Type I only, or no current report.

Baseline gate
Cert 02
ISO 27001

5 = current ISO 27001 certification with 27017 (cloud) and 27018 (PII) extensions, ISO 42001 (AI management system) for AI-specific scope. 3 = base 27001 only, no AI-specific extension. 1 = lapsed or pending certification.

International gate
Cert 03
Sector overlays

5 = sector-appropriate certification current (HIPAA + HITRUST for healthcare, PCI-DSS for payments, FedRAMP for federal, IL5 for defense). 3 = sector certification pending or in audit. 1 = no sector-specific certification despite workload requiring it.

Workload-specific
Cert 04
Regional regimes

5 = active GDPR Article 28 DPA, UK transfer mechanism documented, California CCPA agent, Singapore PDPA mechanism, Australia Privacy Act covered. 3 = EU-only coverage or partial regional mapping. 1 = no documented regional regime coverage.

Geography gate
Cert 05
Audit cadence

5 = annual external audit on every certification, results available to customers under NDA, gap analyses against new frameworks (NIST AI RMF, EU AI Act) published. 3 = audit cadence documented but reports not shareable. 1 = self-attestation rather than external audit.

Programme maturity
Cert 06
Sub-processor register

5 = public sub-processor list, change-notification within 30 days, customer right to object documented. 3 = sub-processor list available on request, notification cadence ad hoc. 1 = sub-processors undisclosed or change without notification.

Transparency gate
Cert 07
Pen-test cadence

5 = annual third-party penetration test, results shareable under NDA, AI-specific red-teaming included. 3 = annual pen-test but generic, no AI-specific testing. 1 = no recent pen-test or test results not shareable.

Adversarial validation
Cert 08
AI-specific frameworks

5 = documented NIST AI RMF mapping, EU AI Act risk-tier alignment, ISO 42001 in-scope. 3 = self-mapped to one framework, no third-party verification. 1 = no AI-specific framework mapping despite AI being the core product.

AI maturity gate

The two AI-specific compliance signals worth weighting heavily in 2026 are ISO 42001 scope (the new AI management system standard) and documented EU AI Act risk-tier alignment. ISO 42001 is the first internationally-recognised certification that addresses AI-specific governance — training-data management, model lifecycle, AI incident response — rather than treating AI as a generic IT system. Vendors who hold ISO 42001 within their AI product scope are signalling that they treat AI governance as a distinct discipline from general information security; vendors who do not are not.

The EU AI Act risk-tier alignment matters even for non-European buyers because the regulation establishes the most concrete framework for AI risk classification globally — general-purpose AI obligations, high-risk system requirements, transparency obligations, and prohibited practices. A vendor who can articulate which risk tier their product falls into, and what obligations attach, is a vendor who has done the regulatory homework. A vendor who treats the EU AI Act as a problem for someone else's lawyer is a vendor whose compliance posture is reactive rather than programmatic.

The certificate trap

Compliance certificates are necessary, not sufficient. SOC2 and ISO 27001 tell you the vendor has a programme; the other forty-two points in the assessment tell you whether the programme covers the things that matter for AI procurement specifically. Use the certificates as gates; assess the substance on the points behind them.

04Data HandlingResidency, retention, training opt-out.

Data handling is the third domain — ten points covering data residency, retention policy, deletion process, training opt-out, output-rights ownership, and the chain of custody from prompt submission to model output. This is the domain where AI vendor risk diverges most sharply from generic SaaS risk: the data submitted to an AI vendor is not just data being processed, it is data that may be used to train future models, may be retained as part of a fine-tuning dataset, and may surface in outputs to other customers.

The training opt-out clause is the single most-skipped control in 2026 AI procurement. Without an explicit opt-out in the contract — not just in the marketing copy — every prompt and document submitted to the vendor is potentially training data for future model versions. The clause has to be enumerated in the contract, named per data category, and verifiable.

Point 21
01
Data residency options

Multiple geographic regions for storage and processing — EU, UK, US, APAC at minimum. Customer chooses region at provisioning, vendor commits in contract.

Sovereignty
Point 22
02
Retention policy named

Per data category (prompts, outputs, logs, fine-tuning data, audit data), retention duration named in days, deletion process documented, deletion verifiable on request.

Lifecycle
Point 23
03
Training opt-out

Customer data must not be used for model training without explicit opt-in. Enumerated per data category. Verifiable. The single most-skipped clause in 2026.

Non-negotiable
Point 24
04
Output rights

Outputs owned by customer; vendor licence to operate the service only. No vendor right to use outputs for marketing, case studies, or training without separate opt-in.

IP gate
Point 25
05
Cross-tenant isolation

Customer prompts and outputs cannot surface to other tenants. Isolation model documented (logical / physical), tested annually, results shareable.

Leakage prevention
Point 26
06
Deletion verifiable

Customer right to delete on request, deletion completed within named SLA (30 days typical), deletion verifiable via signed attestation. Includes backup expiry.

GDPR Article 17
Point 27
07
Sub-processor data flow

Where customer data flows to sub-processors (cloud infrastructure, fine-tuning vendors, evaluation services), the path is documented and the same opt-out applies downstream.

Chain of custody
Point 28
08
Logging discipline

Vendor's own logs of customer data access are minimised, scoped to support and security functions only, retention bounded, access audit trail available.

Internal access
Point 29
09
Backup geography

Backup storage geography matches primary residency commitment, encryption keys customer-managed where required, restoration tested annually.

Disaster recovery
Point 30
10
Export tooling

Customer right to export all submitted data in standard formats on demand, no per-export fees, no data-format lock-in. The migration insurance policy.

Exit ramp

Two operational notes on this domain. First, the training opt-out clause has to be enumerated rather than blanket. A single line saying "customer data will not be used for model training" is weaker than a clause that enumerates the data categories (prompts, documents, outputs, feedback signals, fine-tuning data, telemetry) and applies the opt-out to each named category. The enumerated version is harder to argue around at renewal and produces clearer audit evidence.

Second, deletion verifiability is more important than the deletion SLA itself. A vendor who deletes within 30 days but cannot produce a signed deletion attestation is functionally indistinguishable from a vendor who deletes within 90 days with attestation — and worse than a vendor who deletes within 45 days with attestation. The attestation is the artefact that satisfies the GDPR Article 17 right-to-erasure obligation; without it, the customer cannot demonstrate compliance to their own regulators.

"Every training-data clause you skip at signing becomes a hostage negotiation at renewal. Enumerate the data categories, name the opt-out, verify the implementation."— Standard advice from our AI procurement engagements

05Model ProvenanceTraining data, RLHF, evals, watermarking.

Model provenance is the fourth domain — ten points covering the model's lifecycle from training corpus through fine-tuning, evaluation, deployment, and ongoing monitoring. Provenance assessment is the domain that most distinguishes serious AI vendors from less serious ones; the points below cannot be bluffed in an RFP response, and the vendors who answer them substantively are the vendors whose models you can defend in an audit.

The four signals that matter most in this domain are training data composition disclosure, RLHF process documentation, eval harness coverage, and output watermarking support. Together they tell you whether the vendor treats the model as a serious engineered artefact or as a black-box capability bolted on for the demo.

Point 31
01
Training corpus disclosed

Public summary of training-data composition (web crawl scope, licensed corpora, synthetic data proportion, language coverage), with cutoff date.

Foundation transparency
Point 32
02
IP-clean training data

Affirmative statement that training corpus excludes copyrighted material licensed expressly for training, with indemnification against third-party IP claims on outputs.

IP exposure
Point 33
03
RLHF process documented

Reinforcement learning from human feedback process described: rater pool composition, rubric, inter-rater agreement, refusal behaviour calibration.

Alignment maturity
Point 34
04
Eval harness public

Public eval results on standard benchmarks (MMLU, GPQA, HumanEval, MMLU-Pro, MATH, AIME) plus AI-safety evals (HHH, TruthfulQA, BBQ). Methodology documented.

Capability honesty
Point 35
05
Red-team programme

Internal red-teaming programme with documented test categories (jailbreaks, prompt injection, data extraction), bug bounty pathway, annual external red-team.

Adversarial resistance
Point 36
06
Output watermarking

Cryptographic watermark in generated text or images where applicable, watermark verification tool published, watermark resilience tested against paraphrasing.

Attribution
Point 37
07
Model card per release

Model card published per model release with intended use, limitations, training-data summary, eval results, and known failure modes. Updated with each version.

Documentation discipline
Point 38
08
Hallucination calibration

Documented hallucination rate on standard benchmarks, calibration evaluation, confidence-score availability in API responses. The honesty axis.

Output reliability
Point 39
09
Bias evaluation

Published bias evaluation across protected categories (gender, ethnicity, age, geography, language), with documented mitigation steps and residual-bias acknowledgement.

Fairness signal
Point 40
10
Capability scope statement

Explicit statement of capabilities the model is fit for and capabilities it is not — including domain restrictions and high-risk use cases (medical advice, legal, financial).

Scope discipline

The training-data disclosure question is where vendors most often retreat into vague generalities — "diverse web corpus plus licensed datasets" is the typical answer. Push for the specifics: roughly what proportion is licensed corpora versus web crawl, what is the cutoff date per data source, what is the language distribution, what synthetic-data proportion was used, and what filtering pipeline was applied. Vendors who can answer those questions clearly are vendors with a serious data governance programme; vendors who cannot are vendors whose training pipeline is either undocumented or considered too sensitive to share. Neither of those is acceptable for AI procurement at scale.

On watermarking, the technology is maturing rapidly but adoption is uneven. The 2026 frontier is cryptographic watermarks that survive light paraphrasing on text and standard image transformations on visuals — and that publish a verification tool so the watermark can be checked outside the vendor's infrastructure. Vendors without any watermarking story are signalling that they have not engaged seriously with content-attribution challenges, which is a category of risk that is rising fast as regulators push for AI-generated content disclosure obligations.

The model-card test

Ask the vendor for the model card for the model version you will be deployed against. A current, detailed model card with named limitations and known failure modes is the single best signal of a mature AI vendor. A vendor who refuses, who produces a card with no failure modes listed, or who produces a card more than six months out of date for a currently-shipping model — is a vendor whose model-governance programme is below the standard you should accept.

06OperationalSLA, incident response, deprecation policy.

Operational resilience is the fifth and final domain — ten points covering service availability SLAs, incident response, change management, customer support, capability deprecation policy, and the financial-stability signals that predict whether the vendor will be around at the end of your contract term. The under-discussed point in this domain is deprecation policy: the written commitment around what happens when the vendor decides to stop supporting a capability you have built against.

Use the modes grid below to assess each of the five operational sub-domains; sub-points within each cover the more granular questions (named TAM availability, business-review cadence, credit policy details, executive escalation paths).

Op 01
Service SLA
99.9% / 99.95% uptime, credits

Named uptime SLA (typically 99.9% for general workloads, 99.95% for production-critical), service-credit formula documented, exclusions enumerated. Per-region SLAs where applicable.

Availability gate · 2 points
Op 02
Incident response
P1 ack < 30 min, status page, RCA

P1 acknowledgement within 30 minutes, P1 resolution target within 4 hours, public status page, post-incident root-cause analysis within 5 business days. Named escalation contact.

Operational maturity · 2 points
Op 03
Change management
Notice 30 days, rollback path

Material changes (model updates, API changes, behaviour drift) require 30 days advance notice, rollback path documented, deprecation timelines published. Versioned APIs.

Stability requirement · 2 points
Op 04
Deprecation policy
12mo notice, migration support

Written deprecation policy: minimum 12-month notice for capability deprecation, migration tooling, dedicated migration support, refund-or-credit pathway for contracted capabilities removed before contract end.

Exit insurance · 2 points
Op 05
Financial stability
Funding, customer concentration, audited financials

Funding situation transparent (latest round, runway), customer concentration disclosed (no single customer above 40% of revenue), audited financials available under NDA for enterprise deals.

Survival signal · 2 points

The deprecation policy is the under-discussed risk control in most procurement processes. The standard SaaS contract assumes the vendor will continue to support the capabilities you signed for through the contract term; the AI vendor contract has to handle a different reality, where a capability may be deprecated mid-contract because the underlying model has been retired, the vendor has pivoted strategy, or the capability has been folded into a more expensive product tier. A 12-month notice period with named migration support is the baseline; less than that and the contract is exposed to mid-term operational disruption with no recourse.

Financial-stability assessment matters more for AI vendors than for typical SaaS because the AI vendor market is going through a consolidation phase. Vendor acquisitions, strategic pivots, and outright shutdowns are happening more frequently than in mature SaaS categories, and the cost of a vendor going away mid-contract is materially higher because of the integration depth required for production AI workloads. Assess the financial-stability signals seriously; a vendor with eighteen months of runway and a single customer above 40% of revenue is a vendor whose contract you should price down significantly because of the risk premium.

"The deprecation policy is the operational clause your future self will thank you for. Negotiate it when you have signing leverage — at renewal there is none."— Procurement engagements, on operational resilience

Two operational discipline notes. First, the SLA service-credit formula has to produce real consequences. A vendor offering "service credit of 5% of monthly fees" for a multi-day outage is offering a token rather than a remedy; the credit needs to scale with outage duration and severity, with a cap that is high enough to incentivise actual operational investment. A sensible structure caps credits at 30% of monthly fees for severe outages, which is enough to be felt by the vendor's P&L without being existential. Second, the status page has to be public and historically searchable. A status page that only shows the current state and clears history is a status page designed to make the vendor look good rather than to be useful to customers; insist on a 12-month history at minimum.

07ScorecardSeverity weighting and go/no-go gates.

The 50 points roll up into a single severity-weighted scorecard per vendor, with explicit go/no-go gates applied independently of the weighted total. The severity weighting scheme below uses four tiers — Critical, High, Medium, Informational — with weights that produce a maximum theoretical score of 100 across the 50 points. Vendors that clear the shortlist threshold (typically 75 of 100) and pass every gate move to contract; vendors that fall below the threshold or fail any gate are disqualified.

Severity weighting matters because not every point on the assessment carries equal procurement consequence. A vendor failing on watermarking support is a different category of risk from a vendor failing on cross-tenant data isolation; the scoring scheme has to reflect that asymmetry.

Severity weighting · 50-point AI vendor assessment

Source: Digital Applied AI vendor risk engagements, 2025-2026
Critical · 4 points eachAuth, encryption, cross-tenant isolation, training opt-out, IP-clean training data — items that compound across the portfolio
32 pts
High · 2.5 points eachAudit logs, incident response, ISO 42001, deletion verifiable, model card, deprecation policy — items that gate operational reality
30 pts
Medium · 1.5 points eachSecrets handling, sub-processor register, RLHF documentation, change-management notice, status-page history — items that gate operational quality
21 pts
Informational · 0.7 points eachBias evaluation, capability scope statement, watermarking, financial-stability disclosures — items that gate maturity signals
17 pts
Shortlist threshold · 75/100Vendor must clear 75% of weighted total AND pass every gate to proceed to contract — threshold-only or gate-only is not enough
75 pts
Hard floor · any Critical failedAny single Critical-tier point scored at the floor (1/5) is a hard disqualifier regardless of weighted total — no commercial discount compensates
DISQUALIFY

The threshold matters as much as the weighting. The default shortlist threshold is 75 of 100 on the weighted total — high enough to filter aggressively, low enough that one or two shortlisted vendors usually clear it. Set the threshold before scoring begins; setting it after produces a moving target that rationalises whatever shortlist the team wanted in the first place. Document the threshold and the weighting scheme in the procurement file as the second artefact after the 50-point scorecard itself.

The gate list applies independently of the score. Any single Critical-tier point scored at the floor (1 out of 5) is a hard disqualifier — auth, encryption, cross-tenant isolation, training opt-out, IP-clean training data. A vendor failing any of those at the floor is disqualified regardless of how well they score on the other 49 points. The gates are non-negotiable because the risks they cover compound across the rest of the portfolio; a single gate failure makes every other strength irrelevant.

Re-running the assessment

The 50-point assessment is annual plus on-event. Run the full assessment at every contract renewal, after any vendor acquisition or strategic pivot, after any material model change (version bump, architecture change, behaviour drift), and after any reported security or AI-safety incident at the vendor. Between assessments, monitor the published trust-page and status-page signals; the assessment is the deliberate gate, the monitoring is the early warning.

The discipline that keeps the assessment honest over time is documenting the baseline and rerunning against it. Store the scored scorecard alongside the contract; on the renewal review, rerun the assessment and diff the new scores against the baseline. Where scores have moved — up or down — note the change, capture the evidence, and update the contract red-line for the renewal cycle. Teams that compound advantage in AI procurement are the teams that rerun the assessment; teams that compound regret are the ones that treat the original signing decision as final.

For teams running their first 50-point assessment, the practical starting point is to score the incumbent vendor first — the one already in production — to calibrate the scoring discipline before the shortlisted alternatives. The incumbent assessment becomes the baseline the alternatives are scored against, and the gaps in the incumbent often reveal which of the 50 points carry the most operational weight in your specific context. From there, the template scales cleanly to the rest of the procurement portfolio. Our AI transformation engagements include the full 50-point assessment as part of the procurement-and-renewal track.

"The signing-day scorecard is the cheapest insurance in the procurement portfolio. Skip it and every clause becomes a hostage negotiation; run it and the renewal conversation falls out cleanly."— Digital Applied procurement engagements, on the 50-point template
Conclusion

Vendor risk is decided at signing — the template prevents regret at renewal.

The 50-point AI vendor risk template is a signing-day artefact first, an audit artefact second, and a renewal artefact third. Five domains — security posture, compliance certifications, data handling, model provenance, operational resilience — ten points per domain, four severity tiers, a clear set of gates, and a single severity-weighted scorecard at the end. The artefact is what survives the executive transition, the renewal cycle, and the inevitable board audit that arrives sometime in between.

The discipline that makes the template work is the willingness to disqualify on a single Critical-tier failure regardless of how strong the rest of the assessment looks. Compliance certs are a floor, not a ceiling. Training opt-out is the easy-to-miss control that becomes a hostage negotiation at renewal if skipped. Model provenance distinguishes serious vendors from less serious ones, and capability deprecation policy is the operational clause your future self will thank you for negotiating at signing. Casual application of this template produces a procurement decision that ages badly; deliberate application produces a scorecard that compounds advantage through the rest of the AI portfolio.

The next concrete step is short. Pick the AI vendor your team is closest to signing — or the AI vendor whose contract is closest to renewal — and run the 50-point assessment against them in the next two days. Score severity-weighted, apply the gates, document the evidence per point, and produce the scorecard. By the time the contract goes to legal, the artefact set is documented, the gates are checked, and the renewal conversation eighteen months from now has a baseline to work against. That is the cadence the template is designed to enable.

Defend the vendor decision

AI vendor risk is decided at signing — the template prevents regret at renewal.

Our team runs AI vendor risk assessments — 50-point template, severity-weighted scorecard, go/no-go gates — for procurement and renewal cycles.

Free consultationExpert guidanceTailored solutions
What we work on

AI vendor risk engagements

  • 50-point assessment across five risk domains
  • Compliance-cert verification per region
  • Training opt-out contract clauses
  • Model-provenance evaluation framework
  • Deprecation policy red-flag detection
FAQ · Vendor risk

The questions procurement teams ask before signing.

For most teams the right answer is hybrid — internal evaluators own the assessment, an external assessor reviews the output. The internal evaluators carry the institutional knowledge of your data classifications, regulatory posture, and integration depth, and they are the ones who will live with the procurement decision. The external assessor brings cross-vendor calibration (they have seen the same questionnaires answered by twenty vendors, so they can flag when a vendor's answer is below market standard) and arms-length credibility for the audit trail. The two failure modes to avoid: fully internal with no external review tends to anchor on the most recent successful procurement, which produces drift over time; fully external with no internal ownership tends to produce generic assessments that miss the specific operational realities of your stack. The hybrid model — internal team runs the 50-point assessment, external assessor reviews the scorecard, gates, and evidence file — typically takes three to four weeks per vendor and produces the most defensible procurement artefact for the time invested.