The Definitive Quick Reference for Product Managers & Security Experts
AI Governance is the system of rules, practices, processes, and technological tools that ensure an organization's use of AI technologies aligns with its strategies, objectives, legal obligations, and societal values.
It encompasses policies, oversight structures, accountability mechanisms, and technical guardrails across the entire AI lifecycle — from ideation to decommissioning.
AI systems should treat all people equitably, avoiding discrimination based on race, gender, age, disability, or other protected characteristics. Includes equal quality of service and equitable impact assessment.
AI systems must perform reliably and safely under expected and unexpected conditions. Includes robustness testing, fail-safes, graceful degradation, and continuous monitoring.
AI systems should be designed to engage and empower everyone, addressing potential barriers and ensuring accessibility across diverse populations, abilities, and contexts.
AI systems must protect user data, ensure confidentiality, prevent unauthorized access, and resist adversarial attacks. Data minimization and purpose limitation are core.
People should understand how AI systems work, what data they use, and how decisions are made. Includes documentation, model cards, and clear communication to all stakeholders.
Organizations and individuals must be answerable for AI systems they design, develop, and deploy. Requires clear ownership, audit trails, redress mechanisms, and governance committees.
| Dimension | 🇺🇸 NIST AI RMF 1.0 | 🌐 ISO/IEC 42001 | 🇪🇺 EU AI Act |
|---|---|---|---|
| Type | Voluntary framework / guidelines | Certifiable international standard | Binding legislation (regulation) |
| Scope | All AI systems; any organization | AI management system (AIMS) for any org | AI systems placed on or used in EU market |
| Structure | 4 Functions: Govern → Map → Measure → Manage | PDCA cycle: Plan → Do → Check → Act (ISO management system) | Risk-based tiers: Prohibited → High → Limited → Minimal |
| Risk Approach | Context-dependent risk profiles | Formal risk assessment aligned to ISO 31000 | Pre-defined risk categories with obligations |
| Enforcement | None (voluntary), but referenced by regulators | Third-party certification audits | Fines: up to €35M or 7% global turnover |
| Best For | Building internal AI risk culture & processes | Demonstrating maturity to clients & regulators | Legal compliance for EU market |
| Relationship | NIST maps to ISO 42001 — official crosswalk exists | ISO 42001 + NIST help satisfy EU AI Act conformity | |
| Obligation | What It Requires | Who Is Responsible |
|---|---|---|
| Risk Management System | Continuous identification, analysis, estimation & treatment of risks throughout lifecycle | Provider |
| Data Governance | Training data must be relevant, representative, free of errors; documented data lineage | Provider |
| Technical Documentation | Detailed docs on system design, development, capabilities, limitations before market placement | Provider |
| Record-Keeping | Automatic logging of events for traceability (minimum retention periods apply) | Provider + Deployer |
| Transparency | Clear instructions for deployers; users informed they are interacting with AI | Provider + Deployer |
| Human Oversight | Human-in-the-loop or human-on-the-loop controls; ability to override/shut down | Deployer |
| Accuracy & Robustness | Appropriate levels of accuracy; resilience to errors, faults, adversarial attacks | Provider |
| Cybersecurity | Protection against unauthorized access, data poisoning, model manipulation | Provider |
| Conformity Assessment | Self-assessment or third-party audit before market placement | Provider |
| Post-Market Monitoring | Ongoing surveillance; serious incident reporting to authorities | Provider |
| Risk Category | Examples | Impact | Mitigation Strategy |
|---|---|---|---|
| Technical | Model drift, data quality degradation, adversarial attacks, hallucinations | Wrong decisions, safety failures | Continuous monitoring, retraining triggers, adversarial testing |
| Ethical | Bias, discrimination, unfair outcomes, manipulation | Legal liability, reputational damage, societal harm | Fairness audits, diverse datasets, ethics review boards |
| Legal & Regulatory | Non-compliance with EU AI Act, GDPR violations, IP infringement | Fines, lawsuits, market bans | Compliance mapping, legal review, regulatory monitoring |
| Operational | Over-reliance on AI, loss of human expertise, single point of failure | Business disruption, skill atrophy | Human-in-the-loop, fallback procedures, skills training |
| Reputational | Public controversy, biased outputs reported by media, user backlash | Brand damage, customer churn | Crisis communication plans, proactive transparency |
| Security | Data poisoning, model theft, prompt injection, data exfiltration | Breach, data loss, compromised integrity | Security testing, access controls, input validation |
| Societal | Job displacement, misinformation at scale, concentration of power | Social instability, inequality | Impact assessments, stakeholder engagement, policy advocacy |
| Bias Type | Description | Real-World Example | Stage |
|---|---|---|---|
| Selection Bias | Training data not representative of real-world population | Medical AI trained mostly on Caucasian patient data misdiagnoses for darker skin | Data |
| Historical Bias | Data reflects past societal prejudices | Amazon recruiting tool penalized resumes with "women's" — trained on 10 yrs of male-dominated hiring | Data |
| Measurement Bias | Proxy variables correlate with protected attributes | Using zip code as proxy for creditworthiness discriminates by race | Data |
| Label Bias | Inconsistent or subjective labeling of training data | Annotators labeling African-American Vernacular English as "toxic" more often | Data |
| Aggregation Bias | Single model for diverse subgroups | Diabetes prediction models optimized for majority population fail for minorities | Algorithm |
| Algorithmic Bias | Model architecture or optimization amplifies disparities | COMPAS recidivism tool: higher false positive rate for Black defendants | Algorithm |
| Confirmation Bias | System reinforces existing patterns | Content recommendation creating filter bubbles and radicalization pathways | Deployment |
| Automation Bias | Humans over-trust AI decisions | Clinicians accepting AI diagnosis without independent verification | Deployment |
| Feedback Loop Bias | Biased outputs reinforce biased inputs over time | Predictive policing: more patrols → more arrests → "validates" model → more patrols | Feedback |
| Representation Bias | Underrepresentation in dev teams leads to blind spots | Voice assistants performing poorly for non-native English speakers | Process |
| Technique | Type | How It Works | Scope | Best For |
|---|---|---|---|---|
| SHAP (SHapley Additive exPlanations) |
Model-agnostic | Uses game theory (Shapley values) to assign contribution of each feature to prediction | Local + Global | Feature importance, regulatory explanations, any ML model |
| LIME (Local Interpretable Model-agnostic Explanations) |
Model-agnostic | Creates local linear approximation around a specific prediction by perturbing inputs | Local only | Explaining individual predictions to end-users |
| Attention Maps | Model-specific | Visualizes which parts of input the model "attends to" (transformers, CNNs) | Local | NLP, image classification, debugging |
| Counterfactual Explanations | Model-agnostic | "What would need to change for a different outcome?" — smallest change to flip the decision | Local | Loan denials, job rejections — actionable feedback |
| Partial Dependence Plots | Model-agnostic | Shows marginal effect of one or two features on predicted outcome | Global | Understanding feature relationships |
| Decision Trees / Rule Lists | Intrinsically interpretable | Model itself is human-readable — series of if-then rules | Global | High-stakes domains requiring full transparency |
| Model Cards | Documentation | Standardized disclosure of model purpose, performance, limitations, ethical considerations | Global | Transparency for all stakeholders |
| Confidence Scores | Output-level | Probability or certainty level attached to each prediction | Local | Enabling human override when confidence is low |
| Threat | Description | Attack Vector | Defense |
|---|---|---|---|
| Prompt Injection #1 OWASP |
Attacker crafts inputs to override system instructions or extract sensitive info | Direct (user input) or indirect (poisoned content the model reads) | Input sanitization, instruction hierarchies, output filtering, canary tokens, system prompt hardening |
| Data Poisoning | Corrupting training data to introduce backdoors or degrade model performance | Compromised training pipelines, web scraping of adversarial content | Data provenance tracking, anomaly detection on training data, robust training methods |
| Adversarial Examples | Subtle perturbations to inputs that cause misclassification | Pixel-level image changes, word substitutions in text | Adversarial training, input preprocessing, ensemble methods, certified defenses |
| Model Inversion | Extracting training data or sensitive info from model outputs | Querying model repeatedly to reconstruct private training examples | Differential privacy, output perturbation, rate limiting, membership inference detection |
| Model Theft / Extraction | Replicating a model's behavior through API queries | Systematic querying to build a surrogate model | Rate limiting, watermarking, query monitoring, response perturbation |
| Supply Chain Attack | Compromised pre-trained models, libraries, or datasets | Trojanized models on Hugging Face, malicious pip/npm packages | Model scanning, SBOM (Software Bill of Materials), verified model sources |
| Data Exfiltration | LLM reveals PII, credentials, or proprietary info from training data or context | Crafted prompts, jailbreaks, context window exploitation | PII detection/masking, output scanning, DLP integration, context isolation |
| Excessive Agency | AI system takes unauthorized actions beyond intended scope | Tool-calling without proper constraints, autonomous code execution | Least privilege, action approval gates, sandboxing, human-in-the-loop |
| Phase | Governance Activities | Key Artifacts | Responsible Roles |
|---|---|---|---|
| 1. Problem Formulation & Design | Purpose assessment, stakeholder impact analysis, risk classification (EU AI Act tier), ethics review, necessity & proportionality check, defining fairness criteria | AI Use Case Assessment, Ethics Review Form, Risk Classification Report | Product Manager, Ethics Board, Legal |
| 2. Data Collection & Preparation | Data lineage documentation, consent verification, bias auditing of datasets, DPIA (if high-risk), data quality assessment, representativeness check | Data Card / Datasheet, DPIA Report, Consent Records, Data Lineage Map | Data Engineer, DPO, Data Scientist |
| 3. Model Development & Training | Fairness-aware training, adversarial testing, privacy-preserving techniques, explainability method selection, experiment tracking, IP/copyright compliance | Model Card, Training Report, Fairness Assessment, Experiment Logs | ML Engineer, Data Scientist, Security Engineer |
| 4. Testing & Validation | Red-teaming, bias testing across subgroups, robustness evaluation, security penetration testing, performance benchmarking, human evaluation | Test Reports, Red Team Findings, Conformity Assessment (EU AI Act) | QA, Security Team, Ethics Reviewer |
| 5. Deployment | Human oversight mechanisms, user disclosure (AI transparency), monitoring setup, rollback procedures, incident response readiness, access controls | Deployment Checklist, Monitoring Dashboard, Incident Response Plan, User Disclosure | DevOps/MLOps, Product Manager, Legal |
| 6. Monitoring & Maintenance | Performance drift detection, fairness metric tracking, adversarial monitoring, user feedback analysis, incident logging, periodic re-assessment | Monitoring Reports, Drift Alerts, Incident Logs, Periodic Review Reports | MLOps, Data Scientist, AI Risk Owner |
| 7. Decommissioning | Sunsetting plan, data deletion/archival (GDPR compliance), user notification, documentation archival, model artifact disposal, knowledge transfer | Decommissioning Report, Data Disposal Certificate, Archive Records | Product Manager, DPO, Engineering |
| Role | Key Responsibilities | Governance Focus |
|---|---|---|
| Chief AI Officer (CAIO) | Overall AI strategy, governance oversight, board reporting | Strategic direction, risk appetite, resource allocation |
| AI Ethics Board | Review high-risk use cases, set ethical guidelines, handle escalations | Ethics policies, use case approvals, societal impact |
| Product Manager | Define AI product requirements, manage stakeholder expectations, ensure responsible use cases | Use case justification, user impact, fairness requirements, labeling/transparency |
| Data Protection Officer (DPO) | GDPR/privacy compliance, DPIAs, data subject rights | Privacy by design, consent, data minimization, erasure |
| AI Risk Manager | Identify, assess, and track AI risks; maintain risk register | Risk taxonomy, mitigation tracking, incident management |
| Security Engineer | Secure AI pipelines, adversarial defense, access controls | Threat modeling, penetration testing, supply chain security |
| ML Engineer / Data Scientist | Model development, fairness testing, explainability implementation | Bias mitigation, model cards, experiment reproducibility |
| Legal / Compliance | Regulatory mapping, contractual obligations, liability assessment | EU AI Act classification, GDPR compliance, IP rights |
| Internal Audit | Independent verification of governance controls | Audit trails, conformity assessments, effectiveness reviews |
| Region / Country | Key Framework | Approach | Status (2026) |
|---|---|---|---|
| 🇪🇺 European Union | EU AI Act + GDPR + AI Liability Directive | Risk-based, comprehensive regulation with strict enforcement | Enforcing Full high-risk rules from Aug 2026 |
| 🇺🇸 United States | NIST AI RMF + Executive Orders + State laws (CO, IL, CA) | Voluntary frameworks + sector-specific rules + state-level legislation | Evolving Patchwork of federal + state regulations |
| 🇬🇧 United Kingdom | Pro-Innovation AI Framework + AI Safety Institute | Principles-based, sector regulators apply existing rules to AI | Active AI Safety Institute operational |
| 🇨🇳 China | Generative AI Measures + Algorithm Recommendation Rules + Deep Synthesis Rules | Specific regulations per AI type; state content control; mandatory registration | Enforcing Multiple specific laws active |
| 🇨🇦 Canada | AIDA (Artificial Intelligence and Data Act) — Bill C-27 | Risk-based classification similar to EU; criminal penalties for reckless AI | Pending Still in legislative process |
| 🇮🇳 India | Digital India Act (proposed) + DPDP Act 2023 | Balancing innovation with data protection; no dedicated AI law yet | Developing Advisory-based approach |
| 🇧🇷 Brazil | AI Bill (PL 2338/2023) | Risk-based framework inspired by EU AI Act | Advancing Expected passage soon |
| 🇯🇵 Japan | AI Guidelines for Business + Hiroshima AI Process | Soft law / voluntary guidelines; industry self-regulation | Active Voluntary guidelines in effect |
| 🌐 International Bodies | OECD AI Principles, UNESCO Recommendation, G7 Hiroshima Process, ISO/IEC 42001 | Setting global norms, principles, and standards for harmonization | Active Influencing national laws globally |
GenAI models confidently generate false, fabricated, or nonsensical information. This is especially dangerous in healthcare, legal, and financial domains.
Models trained on copyrighted content raise legal questions about training data use and generated output ownership.
AI-generated video, audio, and images that are indistinguishable from real content.
Users craft inputs to bypass safety controls, extract system prompts, or make the model produce harmful content.
Training large models consumes enormous energy and water resources.
| Category | Tool / Platform | Provider | Key Capabilities |
|---|---|---|---|
| Fairness & Bias | AI Fairness 360 | IBM (Open Source) | 70+ fairness metrics, 11 bias mitigation algorithms |
| Fairlearn | Microsoft (Open Source) | Fairness assessment, bias mitigation for classification & regression | |
| What-If Tool | Google (Open Source) | Visual interface for exploring model fairness without code | |
| Explainability | SHAP Library | Open Source | Shapley-based explanations for any ML model |
| LIME Library | Open Source | Local interpretable explanations via surrogate models | |
| AI Explainability 360 | IBM (Open Source) | 8 explainability algorithms + tutorials | |
| Governance Platforms | ModelOp Center | ModelOp | Enterprise AI governance, model inventory, risk scoring |
| Credo AI | Credo AI | AI governance platform with policy packs for EU AI Act, NIST | |
| Holistic AI | Holistic AI | Auditing, risk management, compliance automation | |
| Security | Garak | NVIDIA (Open Source) | LLM vulnerability scanner — prompt injection, data leakage |
| Rebuff | Open Source | Prompt injection detection framework | |
| Monitoring | Evidently AI | Open Source | ML model monitoring — drift, performance, data quality |
| Fiddler AI | Fiddler | Model performance management + explainability | |
| Privacy | PySyft | OpenMined (Open Source) | Federated learning, differential privacy, encrypted computation |
| Opacus | Meta (Open Source) | Differential privacy for PyTorch model training | |
| Documentation | Model Card Toolkit | Google (Open Source) | Standardized model documentation generation |
| Term | Definition |
|---|---|
| AI System | A machine-based system that generates outputs (predictions, decisions, recommendations, content) that can influence environments. The EU AI Act defines it as a system designed to operate with varying levels of autonomy. |
| Algorithmic Impact Assessment (AIA) | A systematic evaluation of the potential impacts of an AI system on individuals, groups, and society before deployment. |
| Adversarial Attack | Intentionally crafted inputs designed to cause an AI model to make errors or behave unexpectedly. |
| Bias (Algorithmic) | Systematic and repeatable errors in AI output that create unfair outcomes for particular groups. |
| Conformity Assessment | The process of verifying that an AI system complies with regulatory requirements (EU AI Act requires this for high-risk systems). |
| Data Lineage | The complete tracking of data from its origin through all transformations to its use in model training and inference. |
| Deepfake | AI-generated synthetic media (video, audio, images) designed to appear authentic. Must be labeled under EU AI Act. |
| Differential Privacy | A mathematical framework that provides provable guarantees that individual data points cannot be identified in aggregate datasets. |
| DPIA | Data Protection Impact Assessment — required under GDPR Art. 35 for processing likely to result in high risk to individuals. |
| Explainability (XAI) | The degree to which the internal mechanics and outputs of an AI system can be understood by humans. |
| Federated Learning | A machine learning approach where models are trained across decentralized data sources without exchanging raw data. |
| GPAI | General-Purpose AI — AI models capable of performing a wide range of tasks (e.g., GPT-4, Claude, Gemini). Subject to specific EU AI Act obligations. |
| Hallucination | When a generative AI model produces content that is factually incorrect, fabricated, or inconsistent with training data, presented with false confidence. |
| Human-in-the-Loop (HITL) | A system design where a human must approve or intervene in AI decisions before they take effect. |
| Human-on-the-Loop (HOTL) | A system design where a human monitors AI decisions and can intervene or override, but AI acts autonomously by default. |
| Model Card | A standardized document describing a model's intended use, performance metrics, limitations, and ethical considerations. |
| Model Drift | The degradation of model performance over time as real-world data distributions change from training data. |
| Prompt Injection | An attack where malicious instructions are embedded in inputs to override an LLM's intended behavior. |
| RAG | Retrieval-Augmented Generation — technique that grounds LLM responses in retrieved factual documents to reduce hallucinations. |
| Red-Teaming | Adversarial testing where a team actively tries to find failures, biases, and security vulnerabilities in an AI system. |
| Responsible AI | The practice of designing, developing, deploying, and operating AI systems in a manner consistent with ethical principles, legal requirements, and societal values. |
| Shadow AI | Unauthorized or ungoverned use of AI tools by employees outside of IT/governance oversight — a major organizational risk. |
| Systemic Risk (GPAI) | Under EU AI Act, a GPAI model has systemic risk if trained with >10²⁵ FLOPs or designated by EU AI Office, triggering additional obligations. |
AI Governance & Responsible AI — Quick Reference Guide
Prepared for Product Managers & Security Experts · Last Updated: March 2026
Frameworks covered: EU AI Act · NIST AI RMF · ISO/IEC 42001 · OECD AI Principles · GDPR
This document is for educational reference. Consult legal professionals for specific compliance advice.