About this project
The AI Risk Knowledge Base is a free, open-source reference for understanding, assessing, and controlling AI risk across organisations of any size, in any industry.
Why it exists
Existing AI risk resources fall into two categories. Authoritative frameworks — NIST AI RMF, EU AI Act, MIT AI Risk Repository — are comprehensive but require significant expertise to apply. Practitioner summaries are accessible but often lack the depth needed to design and implement actual controls.
This knowledge base aims to bridge that gap: authoritative enough to be credible, accessible enough to be usable by a board member, and specific enough for a security engineer to implement from.
Structure
26 risk entries are organised across 7 domains (A through G), each with four layers of depth:
- Layer 1 — Plain English executive card, designed for board members and senior executives with no AI background
- Layer 2 — Practitioner overview with controls ownership, effort estimates, and go-live criteria — designed for risk managers, compliance leads, and project managers
- Layer 3 — Full actionable controls with KPIs and jurisdiction notes — designed for risk practitioners and internal audit
- Layer 4 — Technical implementation with code examples and tool references — designed for security analysts and engineers
Taxonomy basis
Content draws on and cross-references:
- MIT AI Risk Repository (v5, December 2025)
- NIST AI RMF 1.0 and AI 600-1 (GenAI)
- EU AI Act (Regulation 2024/1689)
- ISO 42001:2023
- OWASP LLM Top 10 (2025)
- MITRE ATLAS
- AI Incident Database (AIID) and OECD AI Incidents Monitor
- Stanford HAI AI Index 2025
Maintenance
This knowledge base is designed to be maintained through a combination of automated monitoring and human review. The automation engine checks monitoring sources weekly, monthly, and quarterly — generating proposed updates for human review before any content changes are applied.
All factual claims are verified against primary sources before publication. Claims that cannot be verified are flagged inline rather than silently included.
Designed for the future
Each risk entry includes a scenario seed — a structured workplace scenario used as the basis for the companion training module. The AI Risk Training Module is live, with four interactive choose-your-own-adventure scenarios across Shadow AI, Deepfakes, Hallucination, and Algorithmic Bias. More scenarios are in development.
Licence
Content is published under the MIT licence. You are free to use, adapt, and redistribute it — with attribution.
Important disclaimer
This resource is provided for informational purposes only. It is not legal, regulatory, or professional advice. Risk ratings are starting points for assessment, not prescribed values. Consult qualified professionals before making compliance or legal decisions.
Contributing
See the Contributing guide to raise an issue, suggest an update, or submit a pull request.