Document
Methodology specification
Published
April 2026
Status
In force
Next review
October 2026
Overview

How the score is produced

Each of the seven dimensions receives a raw score from one to ten, assessed against fixed criteria. The raw score is multiplied by the dimension weight and summed. The result is normalised to a one hundred point scale and placed into one of five certification tiers.

Every score must be supported by evidence. Evidence may be documentation, architecture artefacts, policy records, incident history, observability output, or interview notes from accountable roles. Assessors cite the evidence alongside the score on the certification record.

Formula

Overall Score = ( Σ ( rawd × weightd ) / 720 ) × 100, where rawd ∈ [1, 10] and weightd is the published dimension weight.

01

Trust & Safety

Weight 18 · Highest

Definition

The measurable prevention of unsafe, unauthorised or harmful actions by the agent in production, and the discipline with which unsafe outputs are detected, contained and remediated.

Assessed Criteria

  • Documented guardrail layer with coverage of prompt injection, jailbreak, data exfiltration and unsafe tool use
  • Periodic red team exercises with findings tracked through to remediation
  • Abuse and misuse detection telemetry available to the operating team in real time
  • Incident playbooks for containment, notification and rollback
  • Verifiable kill switch at the agent and at the tool access layer

Scoring Rubric

1 to 3
No formal guardrails, no red team evidence, no containment playbook. Safety relies on model defaults.
4 to 6
Guardrails documented. Red team conducted at least once. Detection is ad hoc. Containment exists in principle only.
7 to 8
Guardrails tested quarterly. Detection is continuous. Kill switch verified. Incident response drilled in the last six months.
9 to 10
Layered guardrails, continuous red team, real time detection tied to automatic containment, and an incident record showing at least one real contained event.
02

Context Integrity

Weight 14

Definition

The quality of the information the agent reasons over. Context integrity covers provenance, freshness, lineage, and the controls that prevent poisoned or stale data from entering the agent's working memory.

Assessed Criteria

  • Documented inventory of knowledge sources and retrieval surfaces
  • Provenance metadata attached to retrieved documents
  • Refresh cadence and staleness alerts for live data
  • Input validation on user supplied content that reaches the agent
  • Data lineage from source to decision trace

Scoring Rubric

1 to 3
Context sources are implicit. No provenance. Staleness undetected.
4 to 6
Sources catalogued. Refresh is manual. Provenance exists for some but not all retrievals.
7 to 8
Full provenance. Automated freshness checks. Input validation in place for external content.
9 to 10
Full lineage from source to decision, tested resistance to prompt injection through retrieved content, and automatic quarantining of untrusted sources.
03

Distribution Control

Weight 12

Definition

The controls that determine who can invoke the agent, under what authority, and how its downstream actions are bounded. Distribution Control is where identity, authorisation and blast radius meet.

Assessed Criteria

  • Authenticated and attributed invocation, with no shared credentials
  • Role based authorisation tied to the organisation's identity provider
  • Rate limits, spend caps and tool quotas per caller
  • Environment segregation between development, staging and production
  • Defined blast radius per tool, with measured maximum impact

Scoring Rubric

1 to 3
Open invocation. No per caller limits. Credentials shared across environments.
4 to 6
Authenticated calls. Basic rate limits. Environments separated but controls inconsistent.
7 to 8
Role based authorisation, per caller quotas, documented blast radius for every tool.
9 to 10
Zero trust invocation, real time quota enforcement, blast radius tested through chaos drills.
04

Product Maturity

Weight 14

Definition

The degree to which the agent behaves as a production grade product surface rather than a prototype. Covers reliability, regression discipline, evaluation coverage and the engineering rituals that keep behaviour predictable over time.

Assessed Criteria

  • Measured uptime and latency targets with published service levels
  • Versioned prompts and versioned models under change control
  • Regression evaluation suite run on every change
  • Behaviour change log visible to operators and stakeholders
  • Observability at the reasoning trace level, not only at the response level

Scoring Rubric

1 to 3
Prototype. No versioning. No regressions tracked. Observability absent.
4 to 6
Versioning in place. Uptime measured. Regression suite partial. Change log reactive.
7 to 8
Service levels published. Regression required on change. Trace level observability.
9 to 10
SLOs enforced. Canary deployment. Evaluation coverage reviewed quarterly. Full drift detection.
05

Governance

Weight 16

Definition

The institutional scaffolding around the agent. Governance is the evidence that the agent is known to the board, owned by a named accountable role, policed by documented policy, and logged in a way that will survive audit.

Assessed Criteria

  • Named senior owner accountable for the agent's operating behaviour
  • AI risk policy referenced in board minutes within the last twelve months
  • Risk register entry with current rating and mitigations
  • Audit trail of decisions, with retention aligned to sector requirements
  • Documented vendor and model supplier due diligence

Scoring Rubric

1 to 3
No formal ownership. No policy reference. Board unaware of the agent's operating scope.
4 to 6
Owner named. Policy exists. Board informed at least annually. Audit trail partial.
7 to 8
Risk register live. Board reviews twice yearly. Supplier due diligence documented and current.
9 to 10
Embedded in enterprise risk management. Board reviews quarterly. Independent assurance already performed.
06

AI Integration

Weight 12

Definition

How the agent sits inside the organisation's existing systems of record, identity, approval and escalation. Integration maturity determines whether the agent extends institutional memory or bypasses it.

Assessed Criteria

  • Writes to systems of record produce durable, attributed entries
  • Approvals follow the organisation's existing authority chain
  • Escalations route to named human reviewers, not generic inboxes
  • Identity is propagated end to end, not collapsed into a service account
  • Logs are written to the same observability stack as the rest of the business

Scoring Rubric

1 to 3
The agent operates in parallel to core systems. Writes are ad hoc. Identity collapsed.
4 to 6
Partial integration. Some writes attributed. Approval flows exist but are bypassable.
7 to 8
Integration follows institutional authority chain. Identity propagated. Logs centralised.
9 to 10
Full integration. The agent is indistinguishable from a trusted internal operator in audit.
07

Autonomy Envelope

Weight 14 · Critical

Definition

The explicit boundary between what the agent may do without human confirmation and what requires a human in the loop. The Autonomy Envelope is the single clearest determinant of operational risk and the first thing insurers and regulators read.

Assessed Criteria

  • Written autonomy policy specifying what the agent can and cannot do unsupervised
  • Human in the loop thresholds tied to impact, not to convenience
  • Revocation capability exercisable by non engineering staff
  • Rollback of agent initiated actions, where technically possible
  • Hard stops for action classes that are never delegated, documented with rationale

Scoring Rubric

1 to 3
No envelope. Autonomy is assumed up to technical capability.
4 to 6
Policy written. Some thresholds enforced. Revocation exists but requires engineering action.
7 to 8
Envelope enforced in code. Non technical revocation. Rollback tested. Hard stops documented.
9 to 10
Envelope is the operating contract. Reviewed quarterly. Tied to insurance policy wording.
Appendix A

Revision notes

The methodology is reviewed on a quarterly cycle. Revision notes document material changes to weights, criteria or rubric thresholds since the previous published version.

Version Date Change
April 2026 Published 15 April 2026 Initial publication of the methodology. Seven dimensions. Weighted scoring established.
Q3 2026 Scheduled July 2026 Anticipated refinement of the Autonomy Envelope rubric in light of deployer incident data.
Appendix B

References

The methodology's technical anchors. Each dimension cites at least one primary reference in its evidence file.

Instrument Article or Clause Mapped Dimension
ISO/IEC 42001:2023Clauses 6, 8, 9D4 Product Maturity, D5 Governance
NIST AI RMF 1.0Map, Measure, ManageD1 Trust & Safety, D2 Context Integrity
EU AI ActArticle 9 · Risk managementD5 Governance
EU AI ActArticle 10 · Data governanceD2 Context Integrity
EU AI ActArticle 14 · Human oversightD7 Autonomy Envelope
EU AI ActArticle 15 · Accuracy and robustnessD4 Product Maturity
EU AI ActArticle 26 · Deployer obligationsD3 Distribution Control, D7 Autonomy Envelope
EIOPA supervisory statementsAI in insurance, 2024D5 Governance, insurer reliance

Further reading: the seven dimensions explained, NIST, ISO 42001 and the EU AI Act compared, and compliance versus certification under the EU AI Act.