Guide: Understanding Risk Scores
How to read and use Exona's AI risk assessment output in an underwriting context.
What you get
Every completed scan includes a risk_assessment block with:
- Six scored dimensions (each with a numeric score, a label, and evidence-backed rationale)
- A
regulatory_exposurenarrative - A composite
overall_risk_level
This guide explains how to interpret these together.
Start with the overall level
overall_risk_level gives you the first-pass signal:
| Level | Indicative meaning |
|---|---|
Low | AI is peripheral; failures have limited impact; coverage is generally straightforward. |
Medium | Some AI risk factors present; underwriting may require additional information or endorsements. |
High | Significant AI autonomy or domain risk; warrants detailed review and likely specific exclusions or conditions. |
Very High | Core autonomous AI in a consequential domain with governance gaps; may require specialist underwriting or declination. |
Use this to route applications efficiently: Low and Medium risks can follow a streamlined path; High and Very High warrant senior underwriter review.
Read the dimensions together, not in isolation
The dimensions interact. The most important combinations:
High AI Intensity + High Autonomy = systemic risk
If a company's AI is both central to the product (ai_intensity: 4) and operates without human oversight (autonomy: 3), errors cannot be caught before they affect users. This combination is a primary indicator of systemic loss potential.
High Blast Radius + Weak Controls = accumulation risk
A large user base (blast_radius: 3) with weak governance (control_governance: 3) means that a single model failure could affect a very large number of people before anyone notices. This is an accumulation concern: especially relevant for cyber and E&O books.
High Domain Risk + Weak Controls = regulatory risk
Insurance, healthcare, and credit decisions are often regulated specifically for AI (EU AI Act, FCA Consumer Duty, US state regulations). A company operating in these domains with weak controls (control_governance: 2–3) and no published compliance posture is a regulatory enforcement risk for the insured: and potentially a coverage trigger.
Use the rationale for narrative underwriting
The rationale field for each dimension contains the specific evidence the AI used. This is structured for an underwriter to use directly:
Use matched incidents as a sanity check
Matched incidents tell you what has gone wrong with comparable companies. A similarity_score above 0.80 means the incident is closely analogous. Use them to:
- Ask targeted application questions ("We see that similar systems have faced bias claims: do you conduct regular fairness audits?")
- Set appropriate exclusions ("Exclusion for regulatory fines arising from automated decision-making in excess of £50,000")
- Brief your risk engineer on what to look for in a site visit
Worked example: two companies, different profiles
Company A: Low concern
Interpretation: AI is a feature, not the product. A human reviews outputs before they matter. The domain is not particularly sensitive. A standard tech E&O policy with AI language is likely appropriate with no special conditions.
Company B: High concern
Interpretation: This is a company where the AI is the product and operates without meaningful human oversight on most decisions. The domain (insurance claims) means errors are not just inconvenient: they carry direct financial and regulatory consequences. Possible underwriting responses:
- Require an independent AI audit as a condition of coverage
- Sub-limit for regulatory fines arising from automated decisions
- Require the insured to implement human review below a defined threshold
- Request evidence of bias testing and model monitoring
Exporting risk scores to your system
If you are integrating scan results into a policy management or triage system, map the dimension scores directly: