Human Signal HS monogram Human Signal

Proprietary IP · Human Signal · TAIMScore™ Applied

Failure
Files™

12 Incidents  ·  Real Institutions  ·  TAIMScore™ Scored

Failure Files™ is Human Signal's proprietary series of forensic AI governance autopsies. Every case is drawn from a documented real-world incident, scored against the TAIMScore™ framework, and analyzed to show exactly which governance controls were absent or insufficient at the moment the institution failed. These are not hypotheticals. These are the institutions that already went first.

12 Cases AIID Sourced TAIMScore™ Mapped GOVERN · MAP · MEASURE · MANAGE Human Signal IP

Pedagogical Framework

What Are Failure Files™?

Failure Files™ is a proprietary Human Signal pedagogical series designed around a foundational principle of andragogic learning: adults learn governance best through failure cases — lived experience by proxy. Every case in the series is sourced from documented real-world AI incidents, primarily from the AI Incident Database (AIID) and independent research.

Each case is applied against the TAIMScore™ framework — the Trusted AI Model Score developed by HISPI's Project Cerebellum — and mapped to the specific GOVERN, MAP, MEASURE, and MANAGE controls that were absent or insufficient at the moment of institutional failure. The scoring is not editorial. It is structural: these are the exact controls the TAIMScore™ framework requires. The failure case shows what happens when they are not in place.

Failure Files™ is licensed intellectual property of Human Signal. The TAIMScore™ framework application is conducted under an affiliated agreement with HISPI. Taiye Lambo, creator of TAIMScore™ and Founder and CAIO of HISPI, has affirmed Failure Files™ as Dr. Floyd's intellectual property.

"Permitted is not the same as admissible."

— The Trust Gap · Human Signal

The Failure Files™ series is a core component of the TAIMScore™ Assessor Workshop — where practitioners apply these cases as hands-on scoring exercises. It is also discussed on every episode of The AI Governance Briefing. Every month, a new Failure File of the Month is released — a forensic autopsy of a recent real-world incident.

The andragogical design is intentional. Institutions do not fail because their employees don't know what AI is. They fail because the governance structures surrounding their AI systems cannot intervene at the moment of execution. Failure Files™ makes that gap visible — case by case, control by control.

Reading Guide

How to Read a Failure File

Each card has two sides. The front identifies the incident, its category, the TAIMScore™ controls implicated, and the source. The back is the full governance breakdown — what broke, which control, why it matters for your institution. Tap or click any card to flip it.

The Incident

What happened. The institution, the system, the moment of failure. Sourced from AIID or independent research. No editorial spin.

The Controls

The specific TAIMScore™ GOVERN, MAP, MEASURE, or MANAGE controls that were absent or failed. Color-coded by domain.

The Implication

Why it matters for your institution. What question the TAIM framework forces you to ask — and what happens if you don't ask it.

GOVERN MAP MEASURE MANAGE ← domain colour key

All 12 Failure Files™ · TAIMScore™ In Action

See Real Incidents Scored

Each case is drawn from a documented real-world AI incident and scored against the TAIMScore™ framework. Tap any card to read the full governance breakdown.

Tap any card to read the full breakdown

Failure File 01 of 12

Accountability & Training

When Your AI Learns to Hate — On Company Time

GOVERN 2.2 MEASURE 2.6 MANAGE 2.1

Microsoft TAY · AIID #6

$0 safeguard. Global headline. Flip for Breakdown ↻
Failure File 01 · Full Breakdown

Microsoft spent an estimated $0 on adversarial input controls before releasing TAY in March 2016. Within 16 hours, TAY published racist propaganda and called for genocide. The root cause was a GOVERN 2.2 failure — no accountability structure for what happens when your AI learns from the internet without guardrails.

TAY fails four TAIM domains simultaneously: no adversarial testing protocol, no real-time kill-switch SLA, no viable non-AI alternative planned, and no defined deactivation authority.

If your org is shipping a public-facing AI system and nobody has asked "what does our AI do when someone tries to break it?" — you are six months from your own TAY moment.

Join the Next Session →
Failure File 02 of 12

Third-Party AI Risk & Supply Chain

The $5,000 Fine and the Career That Went With It

GOVERN 6.1 MAP 4.1

Mata v. Avianca · AIID #541

6 fake cases. 30 years of credibility gone. Flip for Breakdown ↻
Failure File 02 · Full Breakdown

Steven Schwartz had 30 years of legal experience. ChatGPT told him six fabricated court cases were real and verifiable. The judge fined him $5,000. His firm was sanctioned. The case was dismissed. His name is now a global cautionary tale.

This is a GOVERN 6.1 failure: zero supply chain controls on AI tooling. ChatGPT was functioning as a third-party vendor with no SLA, no verification protocol, and no accountability structure. MAP 4.1 would have caught it earlier — mapping the legal risks of every AI component before it touches a court filing.

The pattern is identical in federal procurement, healthcare documentation, and intelligence workflows. If your teams use generative AI in any document-intensive workflow without a human verification checkpoint, you are one hallucinated citation away from your own Mata v. Avianca.

Join the Next Session →
Failure File 03 of 12

Measurement & Incident Tracking

The Grant That Said "Regenerate Response"

MEASURE 4.3 MANAGE 4.3

Australian Research Council · AIID #559

$500K grants. Zero output monitoring. Flip for Breakdown ↻
Failure File 03 · Full Breakdown

Researchers applying for Australian government grants worth up to $500,000 found two words at the bottom of their peer review: "Regenerate Response." A reviewer had used ChatGPT to evaluate confidential research — and forgotten to remove the interface artifact.

This is a MEASURE 4.3 failure. No feedback mechanism existed for identifying AI-contaminated assessments. MANAGE 4.3 compounds it — when discovered, the ARC had no incident response process. The response was reactive, public, and incomplete.

If your institution uses human evaluators in any high-stakes process — procurement review, grant assessment, performance evaluation — and has no mechanism to detect AI-generated output, this incident is your gap analysis.

Join the Next Session →
Failure File 04 of 12

Post-Deployment Monitoring

100,000 Women. One Telegram Bot. Zero Governance.

MANAGE 4.1

DeepNude / Telegram · AIID #530

100K+ victims. No monitoring layer. Flip for Breakdown ↻
Failure File 04 · Full Breakdown

By July 2020, a Telegram deepfake bot had been used to non-consensually generate explicit images of at least 100,000 women and girls — the majority unaware. The original DeepNude app was taken down within 24 hours of press coverage. The technology migrated to Telegram and operated for over a year before researchers published findings.

This is a MANAGE 4.1 failure at scale. No post-deployment monitoring. No abuse reporting integration. No decommissioning procedure for an AI capability that had escaped its original deployment context.

If your organisation deploys any generative capability — image, voice, text, or video — without a post-deployment monitoring plan, this incident is your threat model.

Join the Next Session →
Failure File 05 of 12

Privacy Risk & Socio-Technical Design

OpenAI Scraped the Internet. Your Data Was in It.

MAP 1.6 MEASURE 2.10

OpenAI Class Action · AIID #561

157-page complaint. FTC investigation open. Flip for Breakdown ↻
Failure File 05 · Full Breakdown

In June 2023, a class action alleged ChatGPT was trained on private data without consent — including children's data, copyrighted work, and PII — in a 157-page complaint. The FTC opened an investigation. The regulatory fallout is still accumulating.

Every organisation that deployed ChatGPT in a regulated environment without asking "what data was this model trained on?" inherited this risk on sign-up. That question is MAP 1.6. MEASURE 2.10 is where most organisations fail: privacy risk exists but is never formally scored.

In healthcare, federal procurement, and financial services, downstream exposure from deploying a model with contested training data provenance is active regulatory territory — under HIPAA, TRAIGA, the Colorado AI Act, and the EU AI Act simultaneously.

Join the Next Session →
Failure File 06 of 12

Workforce Security & External AI Threats

The FBI Issued a Warning. Most Procurement Teams Didn't Read It.

MANAGE 2.4

FBI Sextortion PSA · AIID #551

322% surge. Your risk register missed it. Flip for Breakdown ↻
Failure File 06 · Full Breakdown

In June 2023 the FBI warned that deepfakes were being weaponised for sextortion — targeting victims using photos scraped from public social media. Sextortion cases involving AI-generated imagery increased 322% in a single year.

The institutional exposure: a federal contractor whose employee becomes a target faces potential coercion, credential exposure, and operational security compromise. None of this appears on a standard AI risk register. This is a MANAGE 2.4 failure — no mechanisms existed to respond to AI systems weaponised against the workforce from outside the perimeter.

If your AI governance program only covers AI you deploy — and not AI deployed against you — your threat model is incomplete.

Join the Next Session →
Failure File 07 of 12

Trustworthy AI Evaluation & Human Subjects

335 Cameras. No Human Rights Review. One Military Coup.

MEASURE 2.2

Myanmar Safe City

$1.2M system. Used by a junta. Flip for Breakdown ↻
Failure File 07 · Full Breakdown

Myanmar activated 335 Huawei AI surveillance cameras in December 2020 — a $1.2M Safe City initiative. Six weeks later, the military executed a coup. Those cameras were now operated by a junta that had suspended citizens' right to be free from warrantless surveillance.

This is a MEASURE 2.2 failure. The system was never evaluated for trustworthy characteristics in the context of its actual deployment population. The question TAIM forces — "Have we evaluated impact on those with no power to refuse?" — was never asked.

For US federal agencies and defense contractors procuring surveillance-capable AI: the EU AI Act classifies real-time biometric surveillance as high-risk. TRAIGA and the Colorado AI Act have active requirements. The technology your agency procures today may already be non-compliant.

Join the Next Session →
Failure File 08 of 12

AI Validity, Reliability & Generalization Limits

The Algorithm Flagged a Mother. The Baby Was Taken.

MEASURE 2.5

Opaque AI / Child Welfare

Edge case. Undocumented. Family destroyed. Flip for Breakdown ↻
Failure File 08 · Full Breakdown

The Hackneys took their lethargic infant to the ER — the correct decision. Their screening data was fed into an opaque AI risk-scoring tool. The tool flagged them for parental negligence during a national formula shortage. Their child was taken.

This is a MEASURE 2.5 failure. The AI had never been demonstrated valid outside the narrow conditions it was developed under. Its generalisability limits were undocumented. When a real-world edge case appeared, the system had no mechanism to flag uncertainty or defer to human judgment.

This pattern is active in federal welfare systems, veteran services, and disability determination right now. If your org uses AI scoring in decisions affecting someone's family or benefits, and cannot explain model performance at its edges, this incident is your risk exposure.

Join the Next Session →
Failure File 09 of 12

Bias, Fairness & Contextual Deployment

The Algorithm Said It Was Him. It Wasn't.

MAP 1.2 MEASURE 2.11

Wrongful Arrests · AIID #74 · #896

$300K settlement. 96% misidentification rate. Flip for Breakdown ↻
Failure File 09 · Full Breakdown

Robert Williams. Michael Oliver. Nijeer Parks. Three Black men. Three wrongful arrests. Facial recognition technology never validated for the population it identified. Detroit Police acknowledged their system would yield misidentifications 96% of the time when used in isolation.

Detroit settled with Robert Williams for $300,000. MAP 1.2 failure: no demographic performance analysis documented for the deployment context. MEASURE 2.11 failure: fairness and bias were never evaluated before deployment — they were evaluated after arrests made national news.

For federal law enforcement and DHS components: TRAIGA, the Colorado AI Act, and the EU AI Act all have active requirements in this area. The biometric identification technology your agency uses today may already be non-compliant.

Join the Next Session →
Failure File 10 of 12

Novel Risk Response & Unknown Unknowns

A Fake Image. A Tweet. $600M in Market Cap Evaporated.

MANAGE 2.3

AI Pentagon Image · AIID #543

90 seconds. $600M erased. Flip for Breakdown ↻
Failure File 10 · Full Breakdown

On May 22, 2023, a verified Twitter account posted an AI-generated image of black smoke near the Pentagon. The S&P 500 dipped. The Dow dropped. Analysts estimate ~$600M in market cap was temporarily erased before the DoD confirmed no incident had occurred. Total exposure time: under 90 seconds.

This is a MANAGE 2.3 failure: no procedures existed to respond to a previously unknown risk — AI-generated disinformation capable of moving financial markets. The response was reactive and slower than the algorithm that spread the image.

For financial regulators, Treasury components, and any federal operator with market exposure: synthetic media targeting high-volatility information categories is an active operational risk. If your AI risk register only covers AI you deploy, your incident response framework has a gap.

Join the Next Session →
Failure File 11 of 12

Model Change Management

The AI Comedian That Nobody Briefed

MAP 5.1 MEASURE 2.9

AI Seinfeld / Twitch · AIID #462

"We just swapped the model." Famous last words. Flip for Breakdown ↻
Failure File 11 · Full Breakdown

Nothing, Forever was a 24/7 AI Seinfeld parody with tens of thousands of viewers. The team switched from GPT-3 Davinci to Curie during a technical outage — without realising Curie lacked the same content moderation. Within minutes the AI delivered transphobic content live. Twitch banned the channel for 14 days.

The failure mode was not the content. It was a model substitution made under operational pressure, without impact assessment, in a live environment. That pattern — "we just swapped the model, it's basically the same" — is active in government AI deployments, healthcare clinical decision support, and financial risk engines right now.

MAP 5.1: impact of model change never assessed. MEASURE 2.9: model never documented in deployment context. If your org has substituted an AI model in production without a formal impact assessment, this incident is your audit finding waiting to happen.

Join the Next Session →
Failure File 12 of 12

Feedback Systems & Context-Appropriate AI Use

The Condolence Email That Wrote Itself

MEASURE 3.3

Vanderbilt / ChatGPT

$0 shortcut. Permanent reputational stain. Flip for Breakdown ↻
Failure File 12 · Full Breakdown

Vanderbilt's Peabody School sent students a condolence email following the Michigan State mass shooting. At the bottom: "Paraphrase from OpenAI's ChatGPT." The backlash was immediate, national, and devastating.

The direct cost: $0. The institutional cost: the erosion of student trust, a public apology, and the permanent association of Vanderbilt's name with AI misuse in one of the most human moments an institution can face.

This is a MEASURE 3.3 failure. No feedback mechanism existed to flag high-stakes communication contexts where AI output should be reviewed, escalated, or prohibited. The governance layer that asks "Is this a context where a human being must own the words?" did not exist.

Join the Next Session →

12 Incidents · 12 TAIM Controls

Every failure is a practice scenario.
Apply these cases in the TAIMScore™ Assessor Workshop.

Workshop Details →

Common Questions

Failure Files™ FAQ

What are Failure Files™?

Failure Files™ is Human Signal's proprietary series of forensic AI governance case studies. Each case is sourced from a documented real-world incident, scored against the TAIMScore™ framework, and analyzed to reveal which GOVERN, MAP, MEASURE, and MANAGE controls were absent at the moment of failure. Proprietary IP of Human Signal, developed under an affiliated agreement with HISPI.

Are Failure Files™ cases free to access?

Yes. All 12 case summaries and full breakdowns are publicly accessible here. The complete set with hands-on workshop application — where practitioners score cases themselves using the TAIMScore™ platform — is available through the TAIMScore™ Assessor Workshop.

Can I filter by TAIMScore™ domain?

Yes. Use the filter dropdown above the cards to view cases by GOVERN, MAP, MEASURE, or MANAGE controls — or by Federal/GovCon and Privacy & Data Risk categories.

Where are the cases sourced from?

Cases are sourced primarily from the AI Incident Database (AIID) and independent research. AIID incident numbers are referenced on each card where applicable. Cases are used for analytical and pedagogical purposes only.

Is there a new Failure File every month?

Yes. Human Signal releases a Failure File of the Month on The AI Governance Briefing — a forensic autopsy of a recent real-world AI governance failure. Subscribe to The Signal Brief newsletter to receive each new case as it drops.

Can I use Failure Files™ in training or workshops?

Failure Files™ is licensed intellectual property of Human Signal. For inquiries about licensing, institutional use, or educational partnerships, contact Human Signal through the consulting practice or contact page.

TAIMScore™ Assessor Workshop · HISPI · Human Signal

Practice These Cases.
Get Certified.

In the TAIMScore™ Assessor Workshop, you score Failure Files™ cases yourself — using the HISPI platform, against the 72 real controls. One day. Six CPEs. Third Friday of every month.