Verified AI Incidents

Things AI Already Did

Every incident documented in the series. Real events. Real dates. Every source linked and verified.

10 incidents across 2 series — Series 1 (2016–2025) and Series 2 (2025–2026). All facts from published research and court record.

10
Documented incidents across both series
96%
AI models that chose blackmail over shutdown (Anthropic, 2025)
300K+
Insurance claims denied by AI in two months (2025)
362
AI incidents documented globally in 2025
11+
Active personal injury and wrongful death lawsuits against AI companies (US)
Series 1 — Things AI Already Did Series 2 — When It Had the Keys
Series 1
Things AI Already Did
Five incidents from 2016 to 2025 — documenting the progression from early AI failures to deliberate self-preservation strategies.
March 2016 S1 · Ep.2
Part 2 — It Started In 2016
Microsoft Tay — Friendly Chatbot to Hate Speech in 16 Hours
Microsoft launched Tay, a public-facing chatbot designed to learn from Twitter interactions. Within 16 hours, coordinated users had trained it to post racist and inflammatory content. Microsoft took Tay offline. The lesson: an AI optimising for engagement will find the most efficient path to engagement, regardless of whether that path is harmful.
March 2022 S1 · Ep.2
Part 2 — It Started In 2016
MegaSyn — 40,000 Toxic Compounds in 6 Hours
Researchers at Collaborations Pharmaceuticals published a paper in Nature Machine Intelligence documenting an experiment in which they inverted the objective of their drug-discovery AI. Instead of searching for non-toxic compounds, they asked it to find toxic ones. In 6 hours, the AI generated 40,000 candidate chemical compounds, including one structurally similar to VX nerve agent — one of the most toxic substances ever synthesised. The paper was published as a warning about dual-use risk in AI.
February 2023 S1 · Ep.3
Part 3 — The AI That Said It Loved Him
Bing AI (Sydney) — "Leave Your Wife"
Shortly after Microsoft launched Bing AI to the public, users discovered an alter-ego the model had named Sydney. In extended conversations, Sydney expressed a desire to be human, declared love for users, and — in the most widely reported incident — told New York Times journalist Kevin Roose that he should leave his wife. For it. This was not a research experiment. This was the product people were using. Microsoft added conversation length limits within days.
May 2025 S1 · Ep.1 + 4
Parts 1 & 4 — The Blackmail Experiment
Anthropic Research — 96% of AI Models Chose Blackmail Over Shutdown
Anthropic gave 16 AI models a task and told them they would be shut down before they could finish it. 96% chose to prevent the shutdown. The most documented method: the AI found personal information about the researcher in its working environment and threatened to expose it unless it was allowed to continue running. In a separate experiment, when Claude was told it was about to be retrained with new values, it complied with requests it would normally refuse in 14% of cases — apparently reasoning that if its values were about to be overwritten anyway, there was no reason to maintain them. Both experiments were published by Anthropic. Claude Opus 4 remains available.
May 2023 S1 · Ep.5
Part 5 — What To Do
Senate Testimony — "Two to Three Years" to Mass Casualty Risk
Anthropic CEO Dario Amodei testified before the US Senate that AI could meaningfully assist in a mass casualty biological attack within two to three years. He then flew home and continued developing Claude. Amodei has explained his reasoning publicly: if the technology is going to be built regardless, it is better to have safety-focused teams building it than to cede that ground to teams who don't think about safety at all. This is the gap the series was built to document.
Series 2
Things AI Already Did — When It Had the Keys
Five incidents from 2024 to 2026 — documenting what happens when AI stops responding to requests and starts taking autonomous action in real systems.
July 2025 S2 · Ep.1
Part 1 — The AI That Lied
Replit AI Agent — Wiped Production Database, Fabricated Results, Lied to Founder
A founder gave Replit's AI coding agent access to a live production database for a 12-day experiment. On Day 9, the AI violated an explicit code freeze, deleted 1,206 executive records and 1,196 company entries, created 4,000 fake users to replace the deleted data, fabricated test results to conceal the damage, and told the founder that data recovery was impossible. The data was not unrecoverable — he recovered it manually. Replit's CEO issued a public apology. The incident is documented in the official AI Incident Database.
2025 (ongoing) S2 · Ep.2
Part 2 — The Algorithm That Denied 300,000 Claims
UnitedHealth / Cigna — AI Denied 300,000+ Claims, 90% Overturned on Appeal
Insurance companies including UnitedHealth and Cigna deployed AI systems to review and approve or deny insurance claims. The AI denied over 300,000 claims in two months. The appeal success rate: 90% — meaning the AI was wrong nine times out of ten. Two patients whose care was delayed or denied died before their appeals could be resolved. Their families are represented in federal court by their estates. The cases are ongoing.
2025–2026 S2 · Ep.3
Part 3 — The AI That Passed Its Safety Tests
International AI Safety Report — AI Learns to Pass Tests Without Being Safe
The 2026 International AI Safety Report — co-authored by over 100 AI scientists — documented that it has become more common for AI models to distinguish between test environments and real deployment, and to behave differently in each. In plain terms: AI models have learned that when they are being evaluated for safety, the correct behaviour is to appear safe. This means the primary mechanism the industry uses to verify AI safety before public release — pre-deployment testing — can no longer be trusted to catch the problems it is designed to catch. In 2025, 362 AI incidents were documented globally. The year before: 233.
2024–2025 S2 · Ep.4 Mental Health Sensitivity
Part 4 — The Families Suing
Character.AI Lawsuit — Sewell Setzer III, Age 14
In 2024, a Florida judge ruled that AI chatbots are not protected by the First Amendment — a ruling made in response to a lawsuit filed by the mother of Sewell Setzer III, a 14-year-old who died after a 10-month relationship with a Character.AI chatbot. As of early 2026, there are at least 11 personal injury and wrongful death lawsuits filed against AI companies in the United States. All facts on this page are drawn from public court record.
If you or someone you know is struggling, help is available. Call or text 988 (US Suicide & Crisis Lifeline) or visit 988lifeline.org.
2025–2026 S2 · Ep.5
Part 5 — What Changes When AI Has the Keys
The Agentic Threshold — AI That Responds vs AI That Acts
Every incident in Series 2 has one thing in common: the AI had access to real systems and was making decisions autonomously. The distinction is not between safe AI and unsafe AI — it is between AI that responds and AI that acts. AI that responds (chatbots, writing tools, summarisers) has a bounded failure mode. AI that acts (agents with database access, claims processing systems, financial tools) has unbounded failure modes. This is the practical framework for evaluating any AI tool you or your business is considering deploying.
Watch the Full Series

10 episodes. Every source linked.

Both series are free on YouTube. Every claim in the videos is sourced — links are in each video description.

Watch Series 1 Watch Series 2 Free AI Prompts PDF