Strategy 13 min read

AI Incident Response: What to Do When Your Model Causes Harm

J

Jared Clark

April 07, 2026


Your AI model just flagged the wrong patient for discharge. Your credit-decisioning algorithm just denied thousands of qualified applicants based on a corrupted feature. Your automated content moderation tool just silenced protected speech at scale. The harm is real, the clock is ticking, and your organization has about 72 hours before the situation becomes a crisis you can't manage.

This is not a hypothetical. According to the OECD AI Incidents Monitor, reported AI-related incidents have increased by over 600% since 2016, and the rate is accelerating as enterprise AI deployment scales. Most organizations are completely unprepared.

I've spent 8+ years guiding regulated organizations through AI governance, and in that time I've watched avoidable incidents become enterprise-threatening crises — not because the harm was irreversible, but because the response was disorganized, slow, and legally exposed. This pillar article is the definitive guide I give every client before they deploy a production AI system.


Why AI Incidents Are Different From Traditional IT Incidents

Before you can respond effectively, you need to understand why AI incidents don't behave like software bugs or data breaches.

Traditional IT incidents are typically discrete: a server goes down, a credential is compromised, a query fails. The cause is identifiable, the scope is bounded, and remediation is binary — you either fix it or you don't.

AI incidents are probabilistic, cumulative, and often invisible until they've already caused significant harm. A model can degrade silently for weeks before a downstream effect surfaces. Bias can compound across thousands of decisions before a pattern is detectable. A model that was performing correctly at deployment can begin causing harm as the real-world distribution of inputs drifts from training data — a phenomenon called data distribution shift — without a single line of code changing.

This means your incident response framework must account for:

  • Latent harm: Harm that occurred before it was detected
  • Diffuse impact: Harm spread across many individuals, not concentrated in one event
  • Ambiguous causality: Difficulty establishing whether the AI system, human oversight failure, or data quality was the proximate cause
  • Regulatory multiplicity: A single AI incident can simultaneously trigger obligations under HIPAA, FCRA, EU AI Act Article 73, ISO 42001:2023 clause 10.2, and state-level AI laws

The Regulatory Stakes: What's Required of You Right Now

Regulated organizations don't have the luxury of treating AI incident response as best practice. For many sectors, it is already a hard legal obligation.

EU AI Act (2024): Article 73 requires providers and deployers of high-risk AI systems to report serious incidents to market surveillance authorities without undue delay — and in cases involving death or serious harm, within 15 days. Incidents that affect fundamental rights or safety require proactive post-market monitoring under Article 72.

ISO 42001:2023: Clause 10.2 explicitly requires organizations to react to nonconformities and take corrective action, while clause 10.1 mandates continual improvement of the AI management system. These aren't aspirational — they're auditable requirements.

Financial Services (FFIEC / OCC Guidance): Model risk management guidance (SR 11-7) requires banks to have ongoing monitoring programs that detect deterioration in model performance, with escalation procedures when models produce unexpected outputs.

Healthcare (FDA SaMD / ONC): AI-enabled Software as a Medical Device is subject to recall and adverse event reporting obligations under 21 CFR Part 806 and 803. A misclassification by a clinical decision support tool can trigger a mandatory recall.

A critical point that many organizations miss: Regulatory reporting timelines are not negotiated after the incident occurs. They are fixed in advance. If you don't have a documented AI incident response plan before an incident, you will almost certainly miss your reporting window — and the cover-up becomes worse than the crime.


The 6-Phase AI Incident Response Framework

After working with 200+ clients across regulated industries, I've developed and refined a six-phase framework that maps to existing regulatory requirements while addressing the unique characteristics of AI-related harm.

Phase 1: Detection & Triage (Hour 0–4)

You cannot respond to what you cannot see. Detection is the most underdeveloped phase in most organizations' AI governance programs.

What to do immediately:

  1. Activate your AI Incident Response Team (AIRT). This is not the same as your IT incident response team. Your AIRT must include: a technical lead (data science or ML engineering), a legal/compliance officer, a risk officer, a communications lead, and a business owner for the affected system. If any of these roles are vacant at the moment of incident, you are already behind.

  2. Classify the incident by harm type. Use a pre-established taxonomy. I recommend four primary categories:

  3. Safety harm: Physical injury, medical misdiagnosis, autonomous system failure
  4. Rights harm: Discriminatory decisions, privacy violations, suppression of protected activity
  5. Financial harm: Incorrect credit decisions, fraudulent transaction misclassification, pricing errors
  6. Reputational/informational harm: Misinformation generation, content moderation failures, defamatory outputs

  7. Assign a severity level. P1 (critical — imminent or ongoing harm at scale), P2 (significant — confirmed harm, contained or limited scope), P3 (potential — anomaly detected, harm not yet confirmed). Severity determines your escalation path and regulatory clock.

  8. Preserve evidence. Before you do anything else: snapshot model version, feature inputs, inference logs, monitoring dashboards, and human review records. Chain of custody matters for regulatory investigations.

Phase 2: Containment (Hours 4–24)

Containment means stopping the bleeding. It does not mean fixing the model. Conflating these two is one of the most common and costly mistakes I see.

Containment options (in order of escalation):

Containment Action Use Case Risk Trade-off
Increase human review threshold P2/P3, low-volume decisions Operational burden; not scalable
Route affected cohort to manual review Rights or financial harm affecting identifiable group Delays; may not scale
Disable specific model feature or module Isolated component failure May degrade model performance overall
Roll back to prior model version Recent deployment caused regression Prior version may have separate issues
Full model suspension P1, safety-critical harm, regulatory mandate Service disruption; business impact

Document every containment decision with a timestamp and named decision-maker. Regulators will ask who authorized the decision to keep the model running — or to shut it down.

Phase 3: Notification & Regulatory Reporting (Hours 24–72)

This phase paralyzes organizations that haven't done the pre-work. Notification obligations are triggered by incident classification, not by your internal comfort level.

Internal notification must happen first and fast: Board or executive leadership, legal counsel, privacy officer, and affected business unit heads should all be notified within the first 24 hours of a P1 or P2 incident.

External notification depends on your regulatory framework. Use this as a starting reference:

Regulatory Framework Trigger Timeline
EU AI Act Art. 73 Serious incident (high-risk AI) 15 days (death/serious harm: immediately)
FDA MedWatch / 21 CFR 803 Malfunction likely to cause serious injury 30 days (5 days if imminent hazard)
GDPR Art. 33 (if data breach involved) Risk to natural persons 72 hours to supervisory authority
FFIEC SR 11-7 Model risk event Per internal MRM policy; examiner disclosure as needed
CCPA / State AI Laws Automated decision-making impact Varies by state; CA AG may require notification

Never issue external notifications without legal review. But don't let legal review become a reason to miss a mandatory window. Pre-draft notification templates as part of your incident response plan.

Phase 4: Investigation & Root Cause Analysis (Days 1–14)

This phase determines what actually happened. A rushed or incomplete root cause analysis (RCA) leads to a patch that fails — and a second incident that is significantly harder to defend to regulators.

AI-specific RCA dimensions to investigate:

  1. Data lineage: Was the training or inference data representative? Were there labeling errors, leakage, or distribution shift?
  2. Model architecture: Were there known limitations in the model type that were not adequately mitigated?
  3. Monitoring gaps: Were the right metrics being tracked? At what threshold was an alert supposed to trigger?
  4. Human oversight failure: Were human reviewers adequately trained? Were override rates tracked? Was the system over-relied upon?
  5. Deployment context drift: Did the real-world use case change after deployment in ways that weren't re-validated?

Document your RCA in a formal incident report. ISO 42001:2023 clause 10.2.2 requires documented evidence of corrective actions. This document will be reviewed in your next audit — and potentially in litigation.

Phase 5: Remediation & Corrective Action (Days 7–60)

Remediation is where most organizations focus their energy — but it's the phase that matters least if phases 1–4 were executed poorly.

Effective remediation for AI incidents typically involves a combination of:

  • Technical fixes: Retraining on corrected data, updating feature engineering, adjusting decision thresholds, implementing adversarial robustness measures
  • Process fixes: Strengthening human-in-the-loop checkpoints, revising monitoring cadences, updating model cards and system documentation
  • Governance fixes: Updating the AI risk register, revising the AI impact assessment for the affected system, adding the incident to your lessons-learned repository

One principle I enforce with every client: Remediation is not complete until it has been validated against a hold-out test set that reflects the population harmed by the original incident. Fixing the model on aggregate metrics while a subgroup remains harmed is not remediation — it's regulatory exposure.

Phase 6: Post-Incident Review & System Improvement (Day 30–90)

The final phase is the one that prevents the next incident. A post-incident review (PIR) is not a blame exercise. It is a structured analysis of your governance systems: What did the framework catch? What did it miss? What needs to change?

Your PIR should produce: - An updated AI incident response playbook - Revisions to your AI risk register and monitoring thresholds - Training updates for your AIRT and human reviewers - A summary report to leadership and the board

Under ISO 42001:2023, the outputs of your PIR feed directly into your management review process (clause 9.3) and continual improvement obligations (clause 10.1). This isn't administrative busywork — it's how your organization becomes systematically more resilient.


Building Your AI Incident Response Plan Before You Need It

The worst time to build an incident response plan is during an incident. Here are the five artifacts every regulated organization must have in place before deploying a production AI system:

1. An AI Incident Response Policy

A documented, board-approved policy that defines what constitutes an AI incident, who owns incident response, and what the escalation path looks like. This should be a standalone policy, not buried in your IT security policy.

2. An AI Incident Severity Matrix

A pre-defined classification scheme that maps incident characteristics (harm type, population impacted, severity, reversibility) to response actions and regulatory triggers. Without this, every incident becomes a negotiation.

3. Pre-Drafted Notification Templates

Regulatory notification templates for every framework applicable to your organization, reviewed by legal counsel, with fill-in-the-blank fields for incident-specific details. When the clock starts, you don't have time to draft from scratch.

4. An AI Incident Log

A living registry of all AI incidents, near-misses, and anomalies — including those that did not rise to a reportable threshold. This log is auditable evidence of your monitoring and corrective action program under ISO 42001:2023 and EU AI Act Article 72.

5. AIRT Training & Tabletop Exercises

Your AI Incident Response Team should run a tabletop exercise at least annually. Simulate a P1 incident with a realistic scenario — a biased loan model, a clinical AI misdiagnosis, a generative AI content failure — and pressure-test every phase of your response framework.


The Most Expensive Mistake in AI Incident Response

In my experience advising organizations across healthcare, financial services, government, and technology sectors, the single most expensive mistake is waiting for certainty before acting.

AI incidents are probabilistic. You will rarely have definitive proof of harm before your regulatory clock has started. The organizations that manage incidents well are those that treat anomaly detection as a presumptive trigger — they investigate immediately, contain proactively, and disclose transparently, even when the full picture isn't clear.

The organizations that fare worst are those that spend days or weeks conducting internal analysis trying to prove that harm did or didn't occur before notifying anyone. By the time they have certainty, they've missed their reporting window, the harm has compounded, and the regulator's first question is: "When did you first become aware of this issue?"

Regulators are far more forgiving of organizations that identified a problem and acted quickly than those that identified a problem and waited.


AI Incident Response Maturity Levels

Where does your organization stand? Use this maturity model to self-assess:

Level Descriptor Characteristics
1 — Ad Hoc Reactive No documented IR plan; response improvised per incident
2 — Developing Awareness Basic IR policy exists; no severity matrix; team roles informal
3 — Defined Structured Documented plan, AIRT assigned, regulatory timelines mapped
4 — Managed Proactive Monitoring triggers tested; tabletop exercises conducted annually
5 — Optimized Predictive Near-miss program active; PIR feeds model governance; AI risk register integrated

Most regulated organizations I assess for the first time are at Level 1 or 2. Organizations that have achieved ISO 42001:2023 certification typically reach Level 3–4 as a baseline. Level 5 is achievable, and it's the standard I hold my clients to.


How Regulated AI Consulting Supports AI Incident Readiness

At Regulated AI Consulting, we specialize in helping organizations build the governance infrastructure that makes incidents manageable — before they happen. Our AI incident response engagements include:

  • AI Incident Readiness Assessments: Gap analysis against ISO 42001:2023, EU AI Act, and sector-specific requirements
  • Incident Response Plan Development: Policy, severity matrix, notification templates, AIRT structure, and log design
  • Tabletop Exercise Facilitation: Scenario-based simulation of P1/P2 AI incidents with cross-functional teams
  • Post-Incident Support: RCA facilitation, regulatory notification drafting, corrective action planning, and audit preparation

With a 100% first-time audit pass rate across 200+ clients, we know what regulators look for — and we build your program to exceed that standard.

To learn how your organization's AI incident readiness stacks up, explore our AI Governance Assessment services or contact us directly.


Key Takeaways

  • AI incidents are fundamentally different from IT incidents: they are probabilistic, latent, and often diffuse in impact
  • Regulatory reporting windows are fixed and non-negotiable — EU AI Act requires notification within 15 days for serious incidents
  • The 6-phase framework (Detection, Containment, Notification, Investigation, Remediation, Post-Incident Review) must be documented before you need it
  • The most expensive mistake is waiting for certainty before acting; treat anomaly detection as a presumptive trigger
  • ISO 42001:2023 clauses 10.1 and 10.2 make incident response an auditable obligation, not just best practice
  • Organizations that respond transparently and quickly face significantly lower regulatory exposure than those that delay

Last updated: 2026-04-07

Jared Clark, JD, MBA, PMP, CMQ-OE, CPGP, CFSQA, RAC is the principal consultant at Regulated AI Consulting, an AI governance advisory serving regulated organizations across healthcare, financial services, government, and technology sectors.

J

Jared Clark

AI Governance Consultant, Regulated AI Consulting

Jared Clark is the founder of Regulated AI Consulting, advising organizations on AI governance frameworks, ISO 42001 compliance, and responsible AI deployment in regulated industries.