Proprietary IP · Human Signal · TAIMScore™ Applied
Failure
Files™
12 Incidents · Real Institutions · TAIMScore™ Scored
Failure Files™ is Human Signal's proprietary series of forensic AI governance autopsies. Every case is drawn from a documented real-world incident, scored against the TAIMScore™ framework, and analyzed to show exactly which governance controls were absent or insufficient at the moment the institution failed. These are not hypotheticals. These are the institutions that already went first.
Pedagogical Framework
What Are Failure Files™?
Failure Files™ is a proprietary Human Signal pedagogical series designed around a foundational principle of andragogic learning: adults learn governance best through failure cases — lived experience by proxy. Every case in the series is sourced from documented real-world AI incidents, primarily from the AI Incident Database (AIID) and independent research.
Each case is applied against the TAIMScore™ framework — the Trusted AI Model Score developed by HISPI's Project Cerebellum — and mapped to the specific GOVERN, MAP, MEASURE, and MANAGE controls that were absent or insufficient at the moment of institutional failure. The scoring is not editorial. It is structural: these are the exact controls the TAIMScore™ framework requires. The failure case shows what happens when they are not in place.
Failure Files™ is licensed intellectual property of Human Signal. The TAIMScore™ framework application is conducted under an affiliated agreement with HISPI. Taiye Lambo, creator of TAIMScore™ and Founder and CAIO of HISPI, has affirmed Failure Files™ as Dr. Floyd's intellectual property.
"Permitted is not the same as admissible."
— The Trust Gap · Human Signal
The Failure Files™ series is a core component of the TAIMScore™ Assessor Workshop — where practitioners apply these cases as hands-on scoring exercises. It is also discussed on every episode of The AI Governance Briefing. Every month, a new Failure File of the Month is released — a forensic autopsy of a recent real-world incident.
The andragogical design is intentional. Institutions do not fail because their employees don't know what AI is. They fail because the governance structures surrounding their AI systems cannot intervene at the moment of execution. Failure Files™ makes that gap visible — case by case, control by control.
Reading Guide
How to Read a Failure File
Each card has two sides. The front identifies the incident, its category, the TAIMScore™ controls implicated, and the source. The back is the full governance breakdown — what broke, which control, why it matters for your institution. Tap or click any card to flip it.
The Incident
What happened. The institution, the system, the moment of failure. Sourced from AIID or independent research. No editorial spin.
The Controls
The specific TAIMScore™ GOVERN, MAP, MEASURE, or MANAGE controls that were absent or failed. Color-coded by domain.
The Implication
Why it matters for your institution. What question the TAIM framework forces you to ask — and what happens if you don't ask it.
All 12 Failure Files™ · TAIMScore™ In Action
See Real Incidents Scored
Each case is drawn from a documented real-world AI incident and scored against the TAIMScore™ framework. Tap any card to read the full governance breakdown.
Tap any card to read the full breakdown
- All 12 Failure Files™
- GOVERN Controls
- MAP Controls
- MEASURE Controls
- MANAGE Controls
- Federal / GovCon
- Privacy & Data Risk
Accountability & Training
When Your AI Learns to Hate — On Company Time
Microsoft TAY · AIID #6
Microsoft spent an estimated $0 on adversarial input controls before releasing TAY in March 2016. Within 16 hours, TAY published racist propaganda and called for genocide. The root cause was a GOVERN 2.2 failure — no accountability structure for what happens when your AI learns from the internet without guardrails.
TAY fails four TAIM domains simultaneously: no adversarial testing protocol, no real-time kill-switch SLA, no viable non-AI alternative planned, and no defined deactivation authority.
If your org is shipping a public-facing AI system and nobody has asked "what does our AI do when someone tries to break it?" — you are six months from your own TAY moment.
Join the Next Session →Third-Party AI Risk & Supply Chain
The $5,000 Fine and the Career That Went With It
Mata v. Avianca · AIID #541
Steven Schwartz had 30 years of legal experience. ChatGPT told him six fabricated court cases were real and verifiable. The judge fined him $5,000. His firm was sanctioned. The case was dismissed. His name is now a global cautionary tale.
This is a GOVERN 6.1 failure: zero supply chain controls on AI tooling. ChatGPT was functioning as a third-party vendor with no SLA, no verification protocol, and no accountability structure. MAP 4.1 would have caught it earlier — mapping the legal risks of every AI component before it touches a court filing.
The pattern is identical in federal procurement, healthcare documentation, and intelligence workflows. If your teams use generative AI in any document-intensive workflow without a human verification checkpoint, you are one hallucinated citation away from your own Mata v. Avianca.
Join the Next Session →Measurement & Incident Tracking
The Grant That Said "Regenerate Response"
Australian Research Council · AIID #559
Researchers applying for Australian government grants worth up to $500,000 found two words at the bottom of their peer review: "Regenerate Response." A reviewer had used ChatGPT to evaluate confidential research — and forgotten to remove the interface artifact.
This is a MEASURE 4.3 failure. No feedback mechanism existed for identifying AI-contaminated assessments. MANAGE 4.3 compounds it — when discovered, the ARC had no incident response process. The response was reactive, public, and incomplete.
If your institution uses human evaluators in any high-stakes process — procurement review, grant assessment, performance evaluation — and has no mechanism to detect AI-generated output, this incident is your gap analysis.
Join the Next Session →Post-Deployment Monitoring
100,000 Women. One Telegram Bot. Zero Governance.
DeepNude / Telegram · AIID #530
By July 2020, a Telegram deepfake bot had been used to non-consensually generate explicit images of at least 100,000 women and girls — the majority unaware. The original DeepNude app was taken down within 24 hours of press coverage. The technology migrated to Telegram and operated for over a year before researchers published findings.
This is a MANAGE 4.1 failure at scale. No post-deployment monitoring. No abuse reporting integration. No decommissioning procedure for an AI capability that had escaped its original deployment context.
If your organisation deploys any generative capability — image, voice, text, or video — without a post-deployment monitoring plan, this incident is your threat model.
Join the Next Session →Privacy Risk & Socio-Technical Design
OpenAI Scraped the Internet. Your Data Was in It.
OpenAI Class Action · AIID #561
In June 2023, a class action alleged ChatGPT was trained on private data without consent — including children's data, copyrighted work, and PII — in a 157-page complaint. The FTC opened an investigation. The regulatory fallout is still accumulating.
Every organisation that deployed ChatGPT in a regulated environment without asking "what data was this model trained on?" inherited this risk on sign-up. That question is MAP 1.6. MEASURE 2.10 is where most organisations fail: privacy risk exists but is never formally scored.
In healthcare, federal procurement, and financial services, downstream exposure from deploying a model with contested training data provenance is active regulatory territory — under HIPAA, TRAIGA, the Colorado AI Act, and the EU AI Act simultaneously.
Join the Next Session →Workforce Security & External AI Threats
The FBI Issued a Warning. Most Procurement Teams Didn't Read It.
FBI Sextortion PSA · AIID #551
In June 2023 the FBI warned that deepfakes were being weaponised for sextortion — targeting victims using photos scraped from public social media. Sextortion cases involving AI-generated imagery increased 322% in a single year.
The institutional exposure: a federal contractor whose employee becomes a target faces potential coercion, credential exposure, and operational security compromise. None of this appears on a standard AI risk register. This is a MANAGE 2.4 failure — no mechanisms existed to respond to AI systems weaponised against the workforce from outside the perimeter.
If your AI governance program only covers AI you deploy — and not AI deployed against you — your threat model is incomplete.
Join the Next Session →Trustworthy AI Evaluation & Human Subjects
335 Cameras. No Human Rights Review. One Military Coup.
Myanmar Safe City
Myanmar activated 335 Huawei AI surveillance cameras in December 2020 — a $1.2M Safe City initiative. Six weeks later, the military executed a coup. Those cameras were now operated by a junta that had suspended citizens' right to be free from warrantless surveillance.
This is a MEASURE 2.2 failure. The system was never evaluated for trustworthy characteristics in the context of its actual deployment population. The question TAIM forces — "Have we evaluated impact on those with no power to refuse?" — was never asked.
For US federal agencies and defense contractors procuring surveillance-capable AI: the EU AI Act classifies real-time biometric surveillance as high-risk. TRAIGA and the Colorado AI Act have active requirements. The technology your agency procures today may already be non-compliant.
Join the Next Session →AI Validity, Reliability & Generalization Limits
The Algorithm Flagged a Mother. The Baby Was Taken.
Opaque AI / Child Welfare
The Hackneys took their lethargic infant to the ER — the correct decision. Their screening data was fed into an opaque AI risk-scoring tool. The tool flagged them for parental negligence during a national formula shortage. Their child was taken.
This is a MEASURE 2.5 failure. The AI had never been demonstrated valid outside the narrow conditions it was developed under. Its generalisability limits were undocumented. When a real-world edge case appeared, the system had no mechanism to flag uncertainty or defer to human judgment.
This pattern is active in federal welfare systems, veteran services, and disability determination right now. If your org uses AI scoring in decisions affecting someone's family or benefits, and cannot explain model performance at its edges, this incident is your risk exposure.
Join the Next Session →Bias, Fairness & Contextual Deployment
The Algorithm Said It Was Him. It Wasn't.
Wrongful Arrests · AIID #74 · #896
Robert Williams. Michael Oliver. Nijeer Parks. Three Black men. Three wrongful arrests. Facial recognition technology never validated for the population it identified. Detroit Police acknowledged their system would yield misidentifications 96% of the time when used in isolation.
Detroit settled with Robert Williams for $300,000. MAP 1.2 failure: no demographic performance analysis documented for the deployment context. MEASURE 2.11 failure: fairness and bias were never evaluated before deployment — they were evaluated after arrests made national news.
For federal law enforcement and DHS components: TRAIGA, the Colorado AI Act, and the EU AI Act all have active requirements in this area. The biometric identification technology your agency uses today may already be non-compliant.
Join the Next Session →Novel Risk Response & Unknown Unknowns
A Fake Image. A Tweet. $600M in Market Cap Evaporated.
AI Pentagon Image · AIID #543
On May 22, 2023, a verified Twitter account posted an AI-generated image of black smoke near the Pentagon. The S&P 500 dipped. The Dow dropped. Analysts estimate ~$600M in market cap was temporarily erased before the DoD confirmed no incident had occurred. Total exposure time: under 90 seconds.
This is a MANAGE 2.3 failure: no procedures existed to respond to a previously unknown risk — AI-generated disinformation capable of moving financial markets. The response was reactive and slower than the algorithm that spread the image.
For financial regulators, Treasury components, and any federal operator with market exposure: synthetic media targeting high-volatility information categories is an active operational risk. If your AI risk register only covers AI you deploy, your incident response framework has a gap.
Join the Next Session →Model Change Management
The AI Comedian That Nobody Briefed
AI Seinfeld / Twitch · AIID #462
Nothing, Forever was a 24/7 AI Seinfeld parody with tens of thousands of viewers. The team switched from GPT-3 Davinci to Curie during a technical outage — without realising Curie lacked the same content moderation. Within minutes the AI delivered transphobic content live. Twitch banned the channel for 14 days.
The failure mode was not the content. It was a model substitution made under operational pressure, without impact assessment, in a live environment. That pattern — "we just swapped the model, it's basically the same" — is active in government AI deployments, healthcare clinical decision support, and financial risk engines right now.
MAP 5.1: impact of model change never assessed. MEASURE 2.9: model never documented in deployment context. If your org has substituted an AI model in production without a formal impact assessment, this incident is your audit finding waiting to happen.
Join the Next Session →Feedback Systems & Context-Appropriate AI Use
The Condolence Email That Wrote Itself
Vanderbilt / ChatGPT
Vanderbilt's Peabody School sent students a condolence email following the Michigan State mass shooting. At the bottom: "Paraphrase from OpenAI's ChatGPT." The backlash was immediate, national, and devastating.
The direct cost: $0. The institutional cost: the erosion of student trust, a public apology, and the permanent association of Vanderbilt's name with AI misuse in one of the most human moments an institution can face.
This is a MEASURE 3.3 failure. No feedback mechanism existed to flag high-stakes communication contexts where AI output should be reviewed, escalated, or prohibited. The governance layer that asks "Is this a context where a human being must own the words?" did not exist.
Join the Next Session →12 Incidents · 12 TAIM Controls
Every failure is a practice scenario.
Apply these cases in the TAIMScore™ Assessor Workshop.
Common Questions
Failure Files™ FAQ
What are Failure Files™?
Failure Files™ is Human Signal's proprietary series of forensic AI governance case studies. Each case is sourced from a documented real-world incident, scored against the TAIMScore™ framework, and analyzed to reveal which GOVERN, MAP, MEASURE, and MANAGE controls were absent at the moment of failure. Proprietary IP of Human Signal, developed under an affiliated agreement with HISPI.
Are Failure Files™ cases free to access?
Yes. All 12 case summaries and full breakdowns are publicly accessible here. The complete set with hands-on workshop application — where practitioners score cases themselves using the TAIMScore™ platform — is available through the TAIMScore™ Assessor Workshop.
Can I filter by TAIMScore™ domain?
Yes. Use the filter dropdown above the cards to view cases by GOVERN, MAP, MEASURE, or MANAGE controls — or by Federal/GovCon and Privacy & Data Risk categories.
Where are the cases sourced from?
Cases are sourced primarily from the AI Incident Database (AIID) and independent research. AIID incident numbers are referenced on each card where applicable. Cases are used for analytical and pedagogical purposes only.
Is there a new Failure File every month?
Yes. Human Signal releases a Failure File of the Month on The AI Governance Briefing — a forensic autopsy of a recent real-world AI governance failure. Subscribe to The Signal Brief newsletter to receive each new case as it drops.
Can I use Failure Files™ in training or workshops?
Failure Files™ is licensed intellectual property of Human Signal. For inquiries about licensing, institutional use, or educational partnerships, contact Human Signal through the consulting practice or contact page.
TAIMScore™ Assessor Workshop · HISPI · Human Signal
Practice These Cases.
Get Certified.
In the TAIMScore™ Assessor Workshop, you score Failure Files™ cases yourself — using the HISPI platform, against the 72 real controls. One day. Six CPEs. Third Friday of every month.