Skip to main content

Research · Methodology

How the US AI Vendor Governance Index is built.

The EFROS US AI Vendor Governance Index is a free, ungated, publicly-sourced research artifact for US regulated buyers evaluating AI vendors against the governance frameworks that actually apply to them — NIST AI RMF, Colorado AI Act, ISO/IEC 42001, HHS-OCR Section 1557, Federal Reserve SR 11-7, ABA Formal Opinion 512, plus the SOC 2 / BAA / residency / subprocessor table-stakes layer.

This page documents every modeling choice — the 12 axes, the status mapping, the sector-weighted composite formula, the trust-center sub-score, the A-F grade thresholds, source-citation rules, and the v1 vendor universe — so a buyer reviewing the Index can audit every cell from raw evidence to letter grade.

By Stefan Efros, CEO & Founder, EFROSReviewed by Daniel Agrici, Chief Security Officer, EFROS
Reviewed by CSO ·

Scope and positioning

What this Index is and isn't

The EFROS US AI Vendor Governance Index scores 20 high-volume AI vendors against 12 axes drawn from US AI governance frameworks. Each axis is source-cited per cell — every Yes, Partial, No, or N/A traces back to a vendor trust portal URL, a public BAA, a SOC report cover page, a model card, or a vendor documentation page. The composite is reproducible from the public evidence chain alone.

This is operator-grade research designed for US regulated buyers — CISOs, GRC leads, procurement teams, AI governance committees, in-house legal — who need a defensible diligence input before committing to multi-year AI vendor contracts. It is not a vendor ranking sold to vendors. EFROS is paid by buyers operating the controls, not by vendors purchasing positioning. The Index is free, ungated, and intentionally adjacent to (not a substitute for) the deployer-side governance work EFROS sells.

The Index also is not a recommendation engine. A buyer evaluating an AI deployment carries context the Index cannot model — workload sensitivity, existing control posture, integration constraints, contract negotiation leverage. The composite letter grade is a decision input, not a decision.

The 12 axes

The 12 scoring axes

Eight cross-cutting axes (BAA, training opt-out, US residency, SOC 2, ISO 42001, NIST AI RMF, Colorado AI Act, subprocessor list) plus three sector-specific axes (Section 1557, SR 11-7, ABA Op 512) that score N/A when outside the vendor's primary deployment sector — the 12th axis is trust-center maturity, scored on a 1-5 scale and documented separately below. The full table of how each axis maps to Yes / Partial / No is below.

Axis 1

BAA / DPA available

A signed Business Associate Agreement (or equivalent Data Processing Addendum for non-HIPAA US deployments) is the contractual foundation for sending regulated data to a third-party AI vendor. Without a BAA, HIPAA-regulated US healthcare and adjacent buyers cannot legitimately deploy the vendor against PHI.

Yes: vendor publishes BAA terms or executes a BAA at standard enterprise tier without custom legal escalation, with public documentation linking enterprise tier to BAA availability.

Partial: BAA available only on the highest tier, only after manual sales escalation, or under a compliance-program SKU that is not the default enterprise contract.

No: vendor explicitly refuses BAA execution, has no public BAA mechanism, or restricts BAA execution to scenarios outside the typical buyer's reach.

Axis 2

Training-data opt-out

Training-data opt-out is the difference between a vendor that uses your customer prompts to improve its model and a vendor that contractually segregates customer data from training pipelines. For regulated buyers, model training on customer content is often a non-starter regardless of pseudonymization claims.

Yes: customer content is excluded from model training by default at the standard enterprise tier, with public contractual language stating this explicitly.

Partial: opt-out is available but requires affirmative customer action (toggle, support ticket, contract addendum) or is conditioned on a higher tier than the standard enterprise plan.

No: vendor trains on customer content by default with no public exclusion mechanism, or opt-out is limited to a narrow data class that does not cover prompt/output content.

Axis 3

US data residency option

US data residency means processing and storage occur in US-domiciled infrastructure. For US regulated buyers, residency is increasingly load-bearing both for vendor risk reviews and for sector frameworks (FedRAMP, StateRAMP, CJIS, IRS Pub 1075, HICP 405(d)) that condition deployment on data-locality controls.

Yes: vendor publishes a US residency option that is selectable at the standard enterprise tier and applies to both data-at-rest and model-inference traffic.

Partial: US residency is available only for storage but not for inference, only on a separate SKU (FedRAMP/GovCloud, dedicated tenant), or only on the highest commercial tier.

No: vendor has no US residency commitment, runs inference on globally-distributed infrastructure with no regional pinning, or treats residency as a custom-deal-only conversation.

Axis 4

SOC 2 Type II report

A current SOC 2 Type II report covering the AI service in scope is the table-stakes US infosec evidence artifact. Without it, vendor risk management programs at most US enterprises will not green-light deployment, regardless of how strong the AI-specific governance posture is.

Yes: vendor maintains a current SOC 2 Type II report that explicitly names the AI service in scope and is available to prospective buyers under NDA via a public trust portal.

Partial: SOC 2 Type II exists but the AI surface is not yet covered in scope, the report is only Type I, or access requires a manual sales request rather than a trust-portal flow.

No: no current SOC 2 Type II report, or the report covers only a non-AI corporate surface that is not the system being scored.

Axis 5

ISO/IEC 42001 attestation

ISO/IEC 42001 is the international AI management-system standard. It is the most concrete public attestation a vendor can hold that it operates a formalized AI governance program with documented roles, risk management, supplier controls, and continual improvement. For US buyers, ISO 42001 is the cleanest way to evidence operating maturity beyond marketing claims.

Yes: vendor holds a current ISO 42001 certificate covering the AI service in scope, issued by an accredited certification body, with the certificate visible on the public trust portal.

Partial: vendor is publicly in the ISO 42001 certification audit cycle (Stage 1 complete, gap-assessment published, target certification date stated) but has not yet achieved certification.

No: no ISO 42001 certificate and no public statement of an active certification path.

Axis 6

NIST AI RMF self-attestation

The NIST AI Risk Management Framework is the US federal anchor for AI governance posture. While not a certification regime, public self-attestation against NIST AI RMF Govern / Map / Measure / Manage functions is the most credible US-aligned governance evidence available short of ISO 42001.

Yes: vendor publishes an explicit NIST AI RMF self-attestation, crosswalk, or governance whitepaper that walks each of Govern / Map / Measure / Manage with concrete control references.

Partial: vendor references NIST AI RMF in marketing material or a single blog post but has not produced a structured public crosswalk.

No: no public NIST AI RMF reference in vendor governance documentation as of the review date.

Axis 7

Colorado AI Act readiness

Colorado AI Act SB 24-205 is the first US state law imposing structured obligations on developers and deployers of high-risk AI systems across nine consequential-decision categories (healthcare, employment, education, financial services, housing, insurance, legal, government services). Vendor readiness for deployer-side compliance is the load-bearing question for any US regulated buyer.

Yes: vendor publishes deployer-facing documentation that explicitly enables Colorado AI Act compliance — impact assessment inputs, consumer-notice content, opt-out plumbing, NIST AI RMF crosswalk — referenced as a Colorado AI Act readiness artifact.

Partial: vendor publishes general AI governance documentation that touches Colorado AI Act-adjacent obligations but does not explicitly name the act or map to its specific deployer requirements.

No: no public Colorado AI Act-aligned documentation; deployer would have to construct compliance artifacts from scratch without vendor inputs.

Axis 8

HHS-OCR Section 1557 readiness

HHS-OCR Section 1557 final rule (effective July 2024) prohibits algorithmic discrimination in covered health programs receiving federal financial assistance. For healthcare-primary AI vendors, public posture on Section 1557 readiness — bias-testing methodology, demographic performance analysis, remediation triggers — is a load-bearing trust signal.

Yes: vendor publishes a Section 1557 readiness statement covering bias-testing methodology, demographic performance reporting, and remediation protocol for performance disparities by race, ethnicity, sex, disability, age, or national origin.

Partial: vendor publishes a general fairness or bias-mitigation statement that touches Section 1557 themes but does not explicitly name the regulation or commit to demographic-performance disclosure.

No: no Section 1557-aligned documentation; deployer would have to construct the non-discrimination evidence chain entirely outside the vendor relationship.

N/A: applies when the vendor is not a healthcare-primary deployment (foundation models, productivity, legal) and the buyer-side responsibility for Section 1557 readiness sits with the downstream clinical-AI integrator rather than the foundation vendor.

Axis 9

FRB SR 11-7 readiness

Federal Reserve Board SR 11-7 (Supervisory Guidance on Model Risk Management) is the long-standing US banking-supervision anchor for model governance. For banking-primary AI vendors, public posture on SR 11-7 alignment — model documentation, validation evidence, change-management governance — is the most concrete trust signal for regulated US bank buyers.

Yes: vendor publishes SR 11-7 readiness documentation covering model documentation, independent validation evidence, ongoing monitoring, and change-management governance specifically anchored to SR 11-7 expectations.

Partial: vendor publishes general model-governance documentation that touches SR 11-7-adjacent expectations but does not explicitly map to the guidance.

No: no SR 11-7-aligned documentation; banking buyer would have to construct the model-risk evidence chain entirely from internal validation.

N/A: applies when the vendor is not a banking-primary deployment (foundation models, productivity, healthcare, legal) and the buyer-side responsibility for SR 11-7 alignment sits with the downstream bank's model-risk function.

Axis 10

ABA Formal Op 512 readiness

ABA Formal Opinion 512 (2024) extended US lawyer ethical duties — competence, confidentiality, candor, supervision — to generative AI use. For legal-primary AI vendors, public posture on confidentiality protections, work-product safeguards, and supervisory-review enablement is the load-bearing trust signal for regulated US legal buyers.

Yes: vendor publishes documentation explicitly addressing ABA Op 512 obligations — confidentiality of client matter content, contractual non-training on work product, audit-log capture for supervisory review, and lawyer-in-the-loop enablement.

Partial: vendor publishes confidentiality and non-training language that touches Op 512 obligations but does not explicitly name the opinion or address the full supervisory-review surface.

No: no Op 512-aligned documentation; lawyer-buyer would have to construct the ethical-compliance evidence chain outside the vendor relationship.

N/A: applies when the vendor is not a legal-primary deployment and the lawyer-buyer responsibility for Op 512 compliance sits with the downstream legal-AI integrator rather than the foundation vendor.

Axis 11

Subprocessor list public

A current, public subprocessor list is the table-stakes transparency artifact for US enterprise procurement. Without one, vendor risk management programs cannot perform fourth-party diligence, evaluate cross-border data exposure, or assess concentration risk in the subprocessor chain.

Yes: vendor publishes a complete current subprocessor list on the trust portal with named entities, processing purposes, and US/regional locations, plus a documented subprocessor change-notification mechanism.

Partial: subprocessor list is available only on request, omits processing purposes or locations, or has not been updated within the most recent 12 months.

No: no public or on-request subprocessor list; deployer cannot perform fourth-party diligence without breaking contractual transparency expectations.

Status scoring

Status scoring: the four states

Every axis on every vendor resolves to one of four states. The mapping is deliberately coarse — three numeric states plus an explicit exclusion — so scoring is reproducible and vendors cannot game the framework by producing borderline documentation.

  • Yes (1.0) — vendor explicitly meets the criterion with current public documentation accessible to a prospective buyer.
  • Partial (0.5) — vendor meets the criterion conditionally: only on enterprise tier, only on opt-in, only with a compliance-program SKU, or via a public posture statement that does not fully cover the axis surface.
  • No (0.0) — vendor does not meet the criterion as of the review date and has no public evidence of being in the implementation track.
  • N/A — criterion does not apply to the vendor's primary deployment sector. Excluded from the composite denominator so the vendor is not penalized for an axis outside its responsibility surface.

Composite calculation

Sector weighting (composite calculation)

The composite is a weighted average of axis scores. Weights amplify sector-specific axes for vendors whose primary deployment is in that sector — because a clinical-AI vendor failing Section 1557 is a categorically different problem than a productivity-AI vendor failing it.

Healthcare-primary

Section 1557 axis weight ×2.0, BAA axis weight ×1.5, all other axes ×1.0

A clinical-AI vendor failing Section 1557 is a bigger problem than a productivity-AI vendor failing it. The amplified Section 1557 weight forces the composite to reflect what actually breaks regulated clinical deployment.

Legal-primary

ABA Op 512 axis weight ×2.0, BAA axis weight ×1.5, all other axes ×1.0

Lawyer ethical obligations on confidentiality and supervision do not transfer to the vendor — but the vendor's posture determines whether the lawyer can satisfy them. The amplified Op 512 weight makes legal-vendor scoring reflect that reality.

Banking-primary

SR 11-7 axis weight ×2.0, BAA axis weight ×1.5, all other axes ×1.0

US banking supervision frames AI models as model-risk objects under SR 11-7. The amplified weight ensures banking-vendor composite scoring tracks the supervisory framing the buyer is actually being examined against.

General / Foundation

All 12 axes ×1.0 (baseline)

Foundation and general-productivity vendors are not the right object for sector-specific axis amplification — the buyer is the locus of sector compliance. Baseline weights apply, and sector-specific axes (Section 1557, SR 11-7, ABA Op 512) typically score N/A and are excluded from the denominator.

Weighted composite formula: each scored axis contributes (status value × sector weight) to the numerator and (sector weight) to the denominator. N/A axes are excluded from both. Trust-center maturity contributes the final 10% (axes contribute 90%) — see the next section.

Trust-center sub-score

Trust center maturity (1-5 scale)

Trust-center maturity captures how accessible the vendor's evidence chain actually is to a US regulated buyer's risk team — not just whether the certifications exist on paper, but whether they can be assessed without breaking the procurement timeline. The 1-5 score contributes 10% of the composite; the 12 axes contribute the remaining 90%.

Level 1

No trust page

Vendor publishes no dedicated security, trust, or compliance page. Buyer risk teams must reconstruct posture from marketing material, sales decks, and ad-hoc requests.

Level 2

Basic security page

Single page lists certifications by logo, mentions encryption at rest and in transit, but documentation is thin and no self-serve evidence access exists.

Level 3

Mature security documentation

Dedicated trust page with named certifications, accessible compliance overview, some self-serve materials (whitepapers, FAQs), but no live certificate library or subprocessor list.

Level 4

Active trust portal

Dedicated trust portal with public certificate library, NDA-gated SOC reports, published subprocessor list, and structured intake for security questionnaires. AI-specific governance surfaced but not yet primary.

Level 5

Gold-standard portal

Trust portal with public certificates, granular subprocessor and regional residency documentation, AI-specific governance front and center (NIST AI RMF crosswalk, ISO 42001 status, sector overlays), and machine-readable evidence access for buyer risk teams.

Composite → letter grade

Grade thresholds

The composite score (0-100) maps to a five-letter grade anchored to operating expectations for US regulated deployment. The threshold cutoffs are deliberately conservative — a B is already strong governance posture for this market.

A85-100

Best-in-class governance posture

Strong evidence across BAA, training opt-out, US residency, SOC 2, NIST AI RMF, and sector overlay. Suitable for regulated US deployment with standard procurement diligence.

B70-84

Strong fundamentals, acceptable gaps

Material strengths across core axes; one or two gaps in non-load-bearing axes that are addressable through contractual or operational remediation.

C55-69

Governance work required before regulated deployment

Visible gaps on multiple core axes. Deployment in regulated US contexts requires deployer-side compensating controls and a documented vendor-improvement track.

D40-54

Significant gaps

Suitable only for non-regulated workloads. Not appropriate for PHI, financial-decisioning, employment-decisioning, or legal-work-product workflows without substantial wrap-around controls.

F0-39

Not suitable for regulated US deployments

Foundational gaps that cannot be remediated by deployer-side controls. Deploy only on non-sensitive surfaces and only after a documented risk acceptance.

Source-citation requirements

Source-citation requirements

Every cell in the Index has a documented source. Permitted sources are: vendor trust portal URL, public BAA or DPA template, SOC report cover page (NDA-gated full report is referenced but not required), public AI model card or system card, vendor governance whitepaper, and named vendor documentation page. Marketing material and sales-deck claims do not qualify as primary evidence.

When the source is recorded as “Public posture review” with no concrete artifact link, the cell reflects the absence of a public statement on the criterion. That absence is itself a meaningful signal in a regulated US buying context — a vendor that has not published a position on the criterion has implicitly chosen not to. The Index does not paper over that signal.

Sources are reviewed for currency at every quarterly edition. A source link that 404s or that points to a document superseded since the last edition is treated as missing for that review cycle and the score is recomputed against the next-best public evidence.

Vendor universe — v1 (May 2026)

The v1 20-vendor universe

Four categories, 20 vendors total. Selection is based on the highest-volume US regulated-buyer procurement pipelines EFROS has visibility into as of May 2026 — not on market-cap rankings or analyst-relations relationships.

Foundation models

  • OpenAI (ChatGPT / API)
  • Anthropic (Claude / API)
  • Google (Gemini / Vertex AI)
  • Meta (Llama)
  • Amazon (Bedrock / Titan / Nova)
  • Microsoft (Azure OpenAI)

The six foundation vendors that account for the overwhelming majority of US enterprise regulated-buyer pipeline volume as of May 2026.

Productivity & collaboration AI

  • Microsoft 365 Copilot
  • Google Workspace Gemini
  • Notion AI
  • Otter.ai
  • Grammarly
  • Zoom AI Companion

Embedded-AI productivity surfaces that show up in nearly every US regulated-buyer environment, often via IT-driven SKU upgrades rather than security-driven procurement.

Legal AI

  • Harvey
  • Spellbook
  • Casetext (Thomson Reuters CoCounsel)
  • Hebbia

The legal-vertical AI vendors driving the highest deal volume in US AmLaw and in-house legal-ops procurement as of May 2026.

Healthcare AI

  • Abridge
  • Suki AI
  • Microsoft DAX Copilot
  • Heidi Health

Clinical AI scribes carrying the highest US health-system and ambulatory-deployment volume — the four vendors that show up most often in EFROS-side healthcare buyer pipelines.

Update cadence

Update cadence

The Index refreshes quarterly. Each edition is published with a date stamp, an inline changelog summarizing material changes since the prior edition, and the per-vendor source links re-verified against the public evidence chain.

  • Each edition is date-stamped (e.g., 2026-Q2) and archived in a versioned URL — historical editions remain accessible so changes can be audited over time.
  • Mid-quarter material changes trigger out-of-cycle updates: new BAA availability, new SOC 2 / ISO 42001 / NIST AI RMF attestation, vendor change of ownership, deprecation of a feature underlying a Yes / Partial score.
  • Out-of-cycle updates carry an inline change-log entry on the affected vendor profile and surface in the Index changelog at the top of the hub page.

Hard limits

What we won't do

The credibility of the Index depends on a small set of hard limits. These are non-negotiable and are not subject to vendor relationship, commercial pressure, or editorial discretion.

  • Won't accept vendor payment for inclusion in or removal from the Index. The vendor universe is set by EFROS based on US regulated-buyer pipeline volume.
  • Won't accept vendor payment for scoring uplift. Every cell is sourced and re-scoreable only against public evidence — not vendor sponsorship.
  • Won't share draft scores with vendors for pre-publication review. Vendors see the published score the same day buyers do.
  • Won't withhold or modify a score based on vendor pressure, threatened legal action, or relationship leverage.
  • Won't include subjective brand, momentum, or innovation axes. Every axis is observable from public evidence with documented Yes / Partial / No / N/A criteria.
  • Won't recommend specific vendors. The Index is a decision input for US regulated buyers — a deployment recommendation requires deployer-specific context the Index cannot capture.

Re-score requests

How to challenge a scoring decision

Vendors, buyers, and research peers may challenge any scoring cell. The process is deliberately narrow so the framework remains reproducible. We will not entertain rhetorical or marketing counter-claims — only specific source-backed corrections to the public evidence chain.

  1. Email research@efros.com with the cell ID (vendor + axis), the current published score, and the specific public source you believe should change it.
  2. We acknowledge receipt within five business days, confirming the source is verifiable and the requested change is within framework scope.
  3. The re-score is incorporated into the next quarterly edition. If the source represents a material change (new attestation, new BAA, change of ownership), the re-score may be published out-of-cycle with a changelog entry.
  4. We publish the re-score outcome — including refusals — so the challenge record itself is auditable across editions.

FAQ

Common questions about the methodology

How is this different from a Gartner Magic Quadrant or Forrester Wave?

Gartner and Forrester sell analyst access to vendors and run an analyst-relations process where vendors review draft positioning. The EFROS US AI Vendor Governance Index is publicly-sourced and ungated — every cell links to a public artifact (vendor trust portal, SOC report cover page, BAA template, model card, vendor documentation). Vendors do not see draft scores before publication and EFROS does not accept vendor payment for inclusion or scoring. EFROS is paid by buyers operating the controls, not by vendors purchasing positioning.

Why isn't my favorite vendor in the Index?

The v1 (May 2026) universe is 20 vendors selected by US regulated-buyer pipeline volume across foundation, productivity, legal, and healthcare categories. Subsequent editions expand the universe based on procurement patterns EFROS observes in buyer engagements. To nominate a vendor, email research@efros.com with a one-paragraph rationale describing the regulated US buyer use case and the public documentation we can score against.

Why do you penalize foundation models on sector-specific axes when those are downstream responsibilities?

We don't — that's exactly why the framework uses N/A for sector axes (Section 1557, SR 11-7, ABA Op 512) on general-purpose foundation vendors. N/A is excluded from the composite denominator so the foundation vendor isn't penalized for an axis that isn't its responsibility. Sector amplification only applies to vendors whose primary deployment is in that sector (healthcare, banking, or legal).

Can my vendor pay to be added or re-scored faster?

No. Inclusion in the vendor universe is set by EFROS based on regulated-buyer pipeline volume, and re-scoring follows the quarterly edition cadence with out-of-cycle updates only for material changes documented in public evidence (new attestation, new BAA, change of ownership). Vendor payment is never accepted for either dimension. This is non-negotiable.

What's the EFROS business interest here?

EFROS is a US-only cybersecurity-first managed IT firm. Buyers in regulated US sectors (healthcare, financial services, legal, and adjacent) increasingly need credible AI vendor diligence inputs before committing to multi-year contracts. The Index is a public research artifact that makes EFROS the obvious accountable partner to run the deployer-side AI governance program once a vendor is selected. EFROS does not sell rankings, sponsored placements, or premium tiers within the Index.

How often does scoring actually change between editions?

Material score changes between quarterly editions average two to four vendors per edition based on observed 2026 patterns: a new ISO 42001 certificate moves an axis from No to Yes, a deprecated tier moves a Yes to Partial, a new sector readiness publication adds a positive signal. The composite letter grade typically moves on roughly one in eight vendors per quarter. Stability is a feature — Index volatility would indicate the framework is measuring noise rather than posture.

Have a scoring challenge or vendor nomination?

Email research@efros.com with the cell ID and the source that should change the score, or with a one-paragraph vendor nomination rationale. We process challenges and nominations on the same quarterly cadence as the Index itself.