CompTIA Security+ Exam Notes

CompTIA Security+ Exam Notes
Let Us Help You Pass

Friday, March 20, 2026

SCEP Explained: How Devices Securely Enroll and Renew Certificates at Scale

 SCEP (Simple Certificate Enrollment Protocol)

SCEP (Simple Certificate Enrollment Protocol) is a protocol used to automate the enrollment, distribution, and renewal of digital certificates in large-scale environments.

It enables devices, such as laptops, mobile devices, network hardware, and servers, to request and receive certificates from a Certificate Authority (CA) securely without manual intervention.

Originally created by Cisco, SCEP is widely used in:

  • Network infrastructure (routers, switches, firewalls)
  • Mobile Device Management (MDM) (Microsoft Intune, MobileIron, Workspace ONE)
  • VPN and Wi-Fi authentication
  • Zero-trust and identity-based security models
  • IoT devices that need certificates

What Problem Does SCEP Solve?

In enterprise networks, certificates are used for:

  • Device authentication
  • User authentication
  • TLS encryption
  • Wi-Fi 802.1X
  • VPN access
  • Secure email (S/MIME)

Without SCEP, certificates would need to be installed manually, which is:

  • Time-consuming
  • Error-prone
  • Impossible at scale

SCEP enables devices to automatically generate keys, submit certificate requests, and obtain certificates securely.

How SCEP Works (Step-by-Step)

Below is the simplified SCEP workflow.

1. Device generates a key pair

The device creates:

  • A private key (stored securely)
  • A public key used in the certificate request

2. Device creates a Certificate Signing Request (CSR)

The CSR includes:

  • Public key
  • Device identity info
  • Requested certificate type

3. Request is sent to the SCEP server

The device communicates with an SCEP endpoint, typically hosted on:

  • Microsoft NDES (Network Device Enrollment Service)
  • Cisco IOS
  • Cloud PKI systems

4. Authentication (to prevent rogue requests)

Because SCEP is simple, authentication options include:

  • SCEP challenge password (shared secret)
  • One-time passwords
  • Device identity validation via MDM
  • Pre-authentication by Intune or Cisco ISE

5. CA reviews and issues the certificate

The Certificate Authority:

  • Verifies the request
  • Signs the certificate
  • Sends it back to the device

6. Device installs the certificate

The device stores:

  • The certificate
  • The private key
  • Intermediate CA chain

7. Automatic renewal

Before expiration, SCEP allows seamless renewal.

SCEP in Microsoft Intune

In Microsoft Intune, SCEP is used to deploy certificates to:

  • Windows devices
  • iOS/iPadOS
  • Android
  • macOS

Intune uses something called NDES (Network Device Enrollment Service) to bridge the gap between Intune and your internal Microsoft ADCS certificate authority.

The flow looks like this:

1. Intune tells the device: “Here’s where to get your certificate (SCEP URL).”

2. The device generates a key pair.

3. The device sends a CSR to NDES.

4. NDES forwards it to the CA.

5. CA issues a certificate.

6. Intune enforces renewal before expiration.

This enables:

  • Wi-Fi authentication with EAP-TLS
  • VPN authentication
  • Zero-trust, certificate-based access

Security Considerations

SCEP is functional but old, so it has some limitations.

Issues:

  • Weak authentication method (shared secret)
  • No strong device identity validation unless enforced by MDM
  • Limited cryptographic flexibility in early implementations

Mitigations:

  • Always pair SCEP with an MDM (E.g., Intune).
  • Use strong challenge passwords or one-time passwords
  • Use network controls to restrict access to the SCEP URL
  • Prefer modern alternatives when available

SCEP vs Modern Certificate Enrollment Options

SCEP remains common because it is:

  • Lightweight
  • Supported by nearly all devices
  • Easy to integrate

When Should You Use SCEP?

SCEP is best when you need:

  • Automated certificate deployment at scale
  • Support across mixed OS environments
  • Device-based certificate authentication
  • Compatibility with older network equipment or IoT devices
  • Integration with Intune or Cisco ISE

Summary

SCEP (Simple Certificate Enrollment Protocol) is a widely used protocol for automating certificate issuance and renewal across large networks. It allows devices to securely generate key pairs, submit certificate requests, and receive certificates from a CA with minimal manual involvement.

It is essential for:

  • Wi-Fi and VPN authentication
  • Mobile device certificate deployment
  • Zero-trust security models
  • Network infrastructure authentication

Thursday, March 19, 2026

The E‑Discovery Process (EDRM) Made Simple: A Practical Overview

 What Is E‑Discovery? 

E‑Discovery (electronic discovery) is the process of identifying, collecting, preserving, and producing electronic information that is relevant to a legal case, compliance investigation, audit, or regulatory request.

It applies in litigation, HR investigations, cybersecurity events, FOIA/public‑records requests, internal compliance probes, and more.

E‑Discovery focuses specifically on ESI (Electronically Stored Information), which includes:

  • Emails and attachments
  • Documents, spreadsheets, presentations
  • Chat messages (Teams, Slack, SMS, WhatsApp)
  • Databases and logs
  • Cloud data (Microsoft 365, Google Workspace, Salesforce, AWS, etc.)
  • Mobile device data
  • Social media content
  • Audio and video recordings
  • Metadata (timestamps, authorship, access logs, etc.)

The E‑Discovery Process (The EDRM Model)

Most organizations follow the EDRM (Electronic Discovery Reference Model), which outlines 9 stages:

1. Information Governance

Policies and procedures for how data is created, stored, and retained. Good governance reduces e‑discovery costs later.

2. Identification

Determining what ESI might be relevant:

  • Which users?
  • Which devices?
  • Which cloud services?
  • What date ranges?
  • What communication channels?

3. Preservation

Preventing deletion or modification of potentially relevant data.

Tools:

  • Litigation hold
  • Legal hold notifications
  • Retention locks
  • Snapshot backups

4. Collection

Gathering the preserved data in a forensically sound way (without altering metadata).

May include:

  • Exporting mailboxes
  • Collecting Teams/Slack chats
  • Imaging hard drives
  • Exporting logs or cloud records

5. Processing

Reducing data volume and preparing files for review.

Includes:

  • De‑duplication
  • Text extraction
  • Metadata normalization
  • Filtering by date or keyword

6. Review

Attorneys or reviewers examine data for:

  • Relevance
  • Privilege (attorney–client, work product)
  • Confidentiality

Often uses AI tools for efficiency:

  • Predictive coding
  • Technology Assisted Review (TAR)
  • Machine learning relevance ranking

7. Analysis

Deep examination of evidence:

  • Communication patterns
  • Timelines
  • Topic clustering
  • Financial or transactional patterns

8. Production

Providing the requested material to opposing counsel or regulators in an agreed‑upon format (PDF, TIFF, native files, load files, etc.).

9. Presentation

Using selected documents as evidence in court or internal proceedings.

How E‑Discovery Works in Microsoft 365 (high-level)

If you're working in an enterprise environment, e‑discovery is commonly performed using:

Microsoft Purview eDiscovery Standard

For basic cases:

  • Search content across M365
  • Place holds
  • Export results

Microsoft Purview eDiscovery Premium

Advanced, defensible workflows:

  • Legal hold notifications
  • Custodian management
  • Review sets
  • Processing & de-duping
  • Near-duplicate detection
  • Machine learning–based review

Common workloads collected:

  • Exchange Online (email)
  • SharePoint / OneDrive
  • Teams chats (including private & shared channels)
  • Viva Engage/Yammer
  • Purview Audit logs
  • Third‑party data via connectors

Legal and Compliance Considerations

E‑Discovery is heavily governed by legal requirements such as:

  • FRCP (Federal Rules of Civil Procedure) — U.S. federal litigation
  • GDPR — data protection & subject access requests
  • HIPAA — healthcare data
  • SOX — financial records
  • SEC/FINRA — regulated communications

Organizations must ensure:

  • Data preservation is defensible
  • Chain of custody is documented
  • No spoliation (losing or altering evidence)
  • Proper retention schedules exist

Common Technical Challenges in E‑Discovery

  • Massive data volumes
  • Data stored in many systems (cloud, mobile, personal devices)
  • Ephemeral messaging (Teams private channels, Slack DMs, WhatsApp)
  • Encryption and BYOD devices
  • Metadata integrity
  • Cross‑border privacy and data sovereignty

Summary

E‑Discovery is the end‑to‑end process of managing electronic evidence for legal or compliance purposes. It covers:

  • Finding relevant data
  • Preserving it defensibly
  • Collecting it without altering metadata
  • Reviewing and analyzing it
  • Producing it in a legal context


Friday, March 13, 2026

Key Risk Indicators: What They Are and Why They Matter

 Key Risk Indicators (KRIs)

1. What Are KRIs?

Key Risk Indicators (KRIs) are measurable metrics that help an organization detect rising risk exposure before problems occur. They function like the early‑warning sensors of a business, flagging conditions that might lead to operational, financial, strategic, or compliance failures.

Think of KRIs as the smoke detectors in an organization’s risk‑management system, alerting you before the fire spreads.

2. Why KRIs Are Important

KRIs provide:

  • Early detection of risks: They monitor patterns or changes that may indicate rising risk, giving time to take corrective action.
  • Proactive decision-making; KRIs shift organizations from being reactive (fixing problems after damage) to proactive (preventing them).
  • Quantifiable, trackable data: They turn risk into numbers, allowing trends, comparisons, thresholds, and analysis over time.
  • Alignment with business objectives: KRIs help ensure risks are monitored in line with strategic goals, operations, and compliance requirements.

3. Key Characteristics of Effective KRIs

A. Predictive: 

  • KRIs should provide advance warning, not report events that have already occurred.
  • Example: Increase in failed login attempts as an indicator of possible credential‑theft attempts.

B. Measurable and reliable:

  • The data source must be consistent, objective, and accessible.
  • Example: Number of critical system patches not yet applied.

C. Relevant:

  • KRIs must correlate directly with meaningful risks affecting organizational goals.
  • Example: Supplier defect rate for manufacturing quality risk.

D. Threshold-based: 

KRIs usually include:

  • Normal range
  • Warning level
  • Critical level

This allows automated prioritization and escalation.

E. Comparable over time

Good KRIs show trends: increasing, decreasing, or stabilizing risk.

4. Types of KRIs (by risk category)

1. Operational KRIs

Monitor processes, systems, and internal failures.

  • System downtime hours
  • Number of customer complaints
  • Failed backups

2. Financial KRIs

Track financial health and exposure.

  • Days' sales outstanding (DSO)
  • Liquidity ratios
  • Percentage of overdue invoices

3. Compliance KRIs

Identify exposure to legal/regulatory risk.

  • Number of policy violations
  • Percentage of compliance training completed
  • Audit findings

4. Cybersecurity KRIs

Track threats and control effectiveness.

  • Number of phishing attempts detected
  • Patch compliance rate
  • Average time to detect/respond to incidents

5. Strategic KRIs

Linked to long-term organizational goals.

  • Market‑share change
  • Product development delays
  • Customer churn rates

5. How KRIs Fit into Risk Management

KRIs are part of a broader ecosystem:

KPI (Key Performance Indicator)

  • Measures performance (Are we achieving our goals?)

KCI (Key Control Indicator)

  • Measures whether risk controls are working.

KRI (Key Risk Indicator)

  • Measures potential future risk exposure.

These three together form a balanced risk–performance monitoring system.

6. How KRIs Are Developed

Step 1 — Identify critical risks

Start with a risk assessment:

  • "What events could hurt the organization most?"

Step 2 — Determine causes and triggers

  • KRIs should measure the root causes of risk events.

Step 3 — Select measurable indicators

  • Choose metrics directly linked to the risk.

Step 4 — Set thresholds and escalation rules

Define:

  • Normal range
  • Warning level
  • Critical level

Step 5 — Assign ownership

Define who monitors, reviews, and responds to KRI deviations.

Step 6 — Track, report, and refine

  • KRIs must evolve with business strategy and changing risk environments.

7. Examples of Strong KRIs (with explanations)

Example 1: Cybersecurity Risk

  • KRI: Number of systems with overdue critical patches
  • Why: Rising numbers indicate increased vulnerability to attacks.

Example 2: Financial Risk

  • KRI: Ratio of debt to equity
  • Why: High debt levels increase insolvency risk.

Example 3: Operational Risk

  • KRI: Defect rate in manufacturing
  • Why: High defect rates indicate process failures and future financial loss.

Example 4: Compliance Risk

  • KRI: Percent of employees overdue for mandatory compliance training
  • Why: Direct indicator of potential regulatory violations.

8. Benefits of Using KRIs

  • Reduced surprises: Early detection helps avoid catastrophic failures.
  • Better resource allocation: KRIs highlight where controls are truly needed.
  • Increased stakeholder confidence: Boards, regulators, and investors value transparency.
  • Stronger governance: KRIs integrate risk into day-to-day management practices.

9. Common Pitfalls to Avoid

  • Too many indicators (“information overload”)
  • KRIs that measure symptoms, not root causes
  • Poor quality or unreliable data
  • Ignoring threshold breaches due to alert fatigue
  • Setting thresholds too high or too low
  • KRIs are not aligned with the business strategy

In Summary

Key Risk Indicators are measurable, predictive metrics that alert organizations to rising risks.

They help prevent failures, support strategic decision-making, and strengthen the organization’s risk management framework.

Wednesday, March 11, 2026

Expansionary Risk Appetite: What It Is and When It Makes Sense

 What “Expansionary” Means in Risk Appetite

In risk management, risk appetite refers to the amount and type of risk an organization is willing to accept in pursuit of its objectives. It ranges from risk-averse (very low appetite) to risk-seeking (very high appetite).

An expansionary risk appetite sits on the higher end of that spectrum.

Definition: Expansionary Risk Appetite

An expansionary risk appetite means the organization is willing to accept higher-than-normal levels of risk in order to pursue growth, innovation, competitive advantage, or aggressive strategic goals.

It is typically chosen by organizations that want to:

  • Enter new markets
  • Launch new products
  • Rapidly scale operations
  • Invest heavily in innovation or R&D
  • Take bold strategic initiatives

This approach assumes that taking on more risk can bring higher returns, and leadership is consciously choosing this path.

Characteristics of an Expansionary Risk Appetite

1. High Tolerance for Uncertainty

The organization is comfortable operating in areas with unknown outcomes, such as:

  • Emerging technologies
  • Untested business models
  • Rapidly changing markets

2. Acceptance of Higher Financial Risk

Examples include:

  • Large capital investments
  • Reduced reliance on guaranteed returns
  • Higher debt or leverage to fuel growth

3. Proactive, Not Defensive

Instead of protecting its current position, the organization aims to push boundaries, even if failure is possible.

4. Fast Decision-Making

Expansionary organizations accept the risk of imperfect information to maintain speed:

  • Decisions made quickly
  • Shorter project evaluation cycles
  • Willingness to pivot rapidly

5. Innovative and Adaptive Culture

They encourage:

  • Experimentation
  • Creative problem-solving
  • Trial-and-error learning

Failure is treated as a learning opportunity, not grounds for punishment.

Examples of Expansionary Risk Appetite in Practice

Business expansion

  • Opening offices in foreign countries
  • Acquiring competitors or start-ups

Technology adoption

  • Using cutting-edge tools before industry-wide maturity
  • Investing in AI, automation, or IoT aggressively

Product innovation

  • Creating new product lines with uncertain demand
  • Entering high-risk, high-reward markets

Financial decisions

  • Borrowing capital to invest in growth
  • Accepting volatile revenue streams for future potential
Benefits of an Expansionary Risk Appetite

  • Faster innovation
  • Competitive advantage
  • High potential returns
  • Market leadership opportunities
  • Ability to capitalize on emerging trends before others

Organizations with this appetite often grow quickly when successful.

Downsides / Risks

With greater reward comes greater potential downside:

  • Higher chance of financial losses
  • Operational strain due to rapid scaling
  • Higher likelihood of project failure
  • Potential compliance oversights
  • Increased security or privacy exposure (if not managed carefully)

Thus, strong risk controls, monitoring, and contingency planning must accompany expansionary strategies.

Where Expansionary Sits in a Risk Appetite Scale

Expansionary is proactive and growth-oriented, but not reckless.

When an Expansionary Risk Appetite Makes Sense

Organizations tend to adopt an expansionary stance when:

  • The market is full of opportunities
  • They seek rapid scale-up
  • They want to outpace competitors
  • Leadership culture values innovation
  • They have financial stability to absorb potential losses

It is common in:

  • Technology firms
  • Start-ups
  • Companies entering a new market
  • Organizations undergoing digital transformation

Sunday, March 8, 2026

What Is VPN Split Tunneling and How Does It Work

 What Is VPN Split Tunneling?

Split tunneling is a VPN feature that lets you decide which network traffic goes through the encrypted VPN tunnel and which traffic goes directly to the internet without the VPN.

Think of it as creating two separate “paths” for your device’s traffic:

  • Path A: Encrypted → Goes through the VPN to a remote network
  • Path B: Direct → Uses the normal internet connection (no VPN encryption)

Without split tunneling, all your traffic normally flows through the VPN tunnel.

Why Split Tunneling Exists

Split tunneling solves a common problem:

When you connect to a work VPN, you often don’t need everything (Netflix, personal browsing, software updates) to go through the corporate network. Doing so can:

  • Slow your internet connection
  • Overload the VPN
  • block services (e.g., streaming, gaming)
  • increase latency for apps like Zoom or Teams

Split tunneling lets you use the VPN only when needed.

How Split Tunneling Works (Technical Deep Dive)

A VPN creates an encrypted tunnel between your device and the VPN gateway. Split tunneling modifies system routing so that:

  • Selected IP ranges or applications are routed through the VPN gateway
  • Everything else uses the standard network gateway (your ISP router)

Two Types of Split Tunneling

Inclusive Split Tunneling

Only selected traffic uses the VPN.

You choose what to send over the tunnel, e.g.:

  • Only apps like Outlook, SAP, and SSH
  • Only traffic to a corporate IP range
  • Only a specific browser window

Everything else bypasses the VPN.

Exclusive Split Tunneling

Everything uses the VPN EXCEPT specific traffic.

Example exclusions:

  • Streaming services
  • Gaming services
  • Banking websites
  • Local LAN devices (printers, NAS)

Practical Examples

Example 1: Corporate Environment

You're working from home, connected to a company VPN.

Traffic that goes through the VPN:

  • Internal servers (10.x.x.x or 172.16.x.x)
  • Corporate tools like SharePoint or Teams
  • Intranet pages

Traffic that bypasses the VPN:

  • YouTube
  • Personal browsing
  • OS updates
  • Smart home devices

Example 2: Using a VPN for Privacy

You want your web browsing to be private, but want local apps (like printers or smart TVs) to be accessible.

  • Browser traffic → through VPN
  • Local device traffic → bypass VPN

How It’s Implemented (Routing Behavior)

When split tunneling is enabled, the OS routing table is modified:

  • Routes to corporate subnets → next-hop = VPN gateway
  • Routes to local LAN and most public traffic → next-hop = local gateway

This is done using:

  • Windows Routing Table
  • Linux ip route / iptables
  • macOS network routing
  • Mobile OS VPN APIs (Android VpnService, iOS NEPacketTunnelProvider)

VPN clients apply these rules dynamically when the tunnel is established.

Benefits of Split Tunneling

Risks and Considerations

When You Should Not Use Split Tunneling

  • When working with sensitive financial or government data
  • On untrusted public Wi-Fi networks
  • When full anonymity is required
  • If your organization uses zero-trust principles

In these cases, force all traffic through the VPN ("full tunneling").

Summary

Split tunneling = selectively routing traffic through or outside a VPN.

  • Gives performance, flexibility, and reduced load
  • BUT also introduces security trade-offs
  • Can be inclusive (only certain traffic goes through VPN)
  • Or exclusive (everything except selected traffic goes through VPN)

Thursday, February 26, 2026

The NIST AI RMF Explained: A Lifecycle Approach to Managing AI Risk

 NIST AI Risk Management Framework

The NIST Artificial Intelligence Risk Management Framework (AI RMF) is a voluntary, sector‑agnostic, and consensus‑driven framework released by the U.S. National Institute of Standards and Technology on January 26, 2023. Its purpose is to help organizations identify, assess, manage, and reduce risks associated with AI systems across their entire lifecycle. The framework remains a living document and is updated periodically.

It is intended to support:

  • Trustworthy AI development and deployment
  • Decision-making about AI risks
  • Continual monitoring and governance
  • Cross-functional collaboration across technical, operational, and executive teams 

To help organizations operationalize the framework, NIST provides companion resources, including the AI RMF Playbook, Crosswalks, Roadmap, and specialized profiles (including the Generative AI Profile, released July 26, 2024). 

1. Purpose and Philosophy of the AI RMF

Unlike rigid compliance checklists, the AI RMF:

  • Supports AI governance through a flexible, lifecycle-based approach.
  • Addresses socio‑technical risks rather than just technical risks.
  • Encourages continuous, not one‑time, risk management.
  • Helps align AI development with organizational values, ethical constraints, and societal well‑being.

Its core goal is to build trustworthy AI, characterized by reliability, safety, security/resilience, explainability, transparency, privacy enhancement, and fairness, with bias mitigated.

2. AI RMF Structure

The AI RMF consists of two major parts:

1. Governance and Risk Principles

2. The AI RMF Core, based on four high‑level functions:

  • GOVERN
  • MAP
  • MEASURE
  • MANAGE

These functions are continuous and iterative, not linear. A governance foundation informs all other functions. [airc.nist.gov]

3. The Four Core Functions (GOVERN–MAP–MEASURE–MANAGE)

A. GOVERN — Establish organizational governance for AI risk

This is the foundational function.

It ensures:

  • Clear policies, processes, and procedures for AI risk governance
  • Defined roles, responsibilities, and accountability
  • A culture supporting ethics, transparency, DEIA (diversity, equity, inclusion, and accessibility)
  • Strong stakeholder engagement, internal and external
  • Supply‑chain and third‑party risk management processes
  • Ongoing communication and risk awareness across teams 

GOVERN aligns leadership, legal, engineering, data science, compliance, and external stakeholders.

B. MAP — Understand the context and scope of the AI system

MAP focuses on defining what the AI system is, how it will be used, and who and what it will affect.

Key MAP activities include:

  • Identify the context, purpose, and environment of the AI system
  • Categorize the AI system (e.g., safety‑critical vs. low‑impact)
  • Benchmark AI capabilities against alternatives
  • Assess risks across the ecosystem, including data sources, APIs, and third‑party components
  • Identify impacts on individuals, communities, and society
  • Determine risk tolerance and organizational constraints 

MAP ensures organizations define potential harms, foreseeable misuse, dependencies, and assumptions early.

C. MEASURE — Assess and analyze AI risks

MEASURE provides quantitative and qualitative risk evaluations.

Typical MEASURE activities include:

  • Pre-deployment and post‑deployment testing, such as:
    • robustness testing
    • bias and fairness assessments
    • performance and drift monitoring
    • privacy evaluations
  • Verification and validation (V&V)
  • Measuring alignment of AI outputs with intended use
  • Logging, benchmarking, and documentation for risk evidence
  • Independent audit or challenge mechanisms 

MEASURE helps ensure claims about an AI system’s behavior are evidence‑based.

D. MANAGE — Actively manage risks throughout the AI lifecycle

MANAGE implements decisions based on the MAP and MEASURE functions.

Common MANAGE activities:

  • Deploying mitigation strategies for identified risks
  • Implementing risk controls, guardrails, and monitoring plans
  • Incident response planning
  • Lifecycle management: updates, retraining, tuning, or decommissioning
  • Communication procedures for adverse events or misuse
  • Continuous feedback loops between operational teams and leadership 

MANAGE is where organizations convert analysis into action.

4. Trustworthiness Characteristics Embedded in the AI RMF

NIST highlights several key attributes of trustworthy AI:

  • Valid and Reliable
  • Safe
  • Secure and Resilient
  • Accountable and Transparent
  • Explainable and Interpretable
  • Privacy‑Enhanced
  • Fair with Harmful Bias Managed

These characteristics guide organizations in evaluating AI risks and making balanced tradeoffs. 

5. Profiles and Extensions — Including Generative AI

To support specific use cases, NIST publishes Profiles, which tailor the RMF.

The Generative AI Profile (NIST AI 600‑1), released July 26, 2024, identifies unique GAI‑specific risks, including:

  • Hallucinations
  • Intellectual property leakage
  • Toxic or abusive content
  • Security vulnerabilities
  • Misalignment or unexpected model behavior
  • Sensitive data leakage
  • Information integrity threats

These profiles help organizations apply the AI RMF to evolving AI technologies.

6. Implementation Support — The AI RMF Playbook

The AI RMF Playbook provides:

  • Implementation checklists
  • Tactical actions aligned with GOVERN, MAP, MEASURE, MANAGE
  • Practical examples and templates
  • Guidance for aligning risk controls with organization‑specific needs

It is designed to help operationalize the AI RMF, not replace it. 

7. How organizations commonly use the AI RMF

Organizations adopt the AI RMF to:

  • Build internal AI governance systems
  • Address regulator or stakeholder expectations
  • Benchmark their AI risk maturity
  • Avoid ad hoc AI decision‑making pitfalls
  • Harmonize with ISO/IEC 42001 and global AI standards
  • Support compliance with legal regimes such as GDPR and emerging U.S. regulatory guidance

8. Summary

The NIST AI RMF is a flexible, lifecycle‑oriented, risk‑based approach to managing AI systems.

It helps organizations:

  • Establish governance (GOVERN)
  • Understand context and impacts (MAP)
  • Analyze risk (MEASURE)
  • Mitigate and monitor (MANAGE)

Tuesday, February 24, 2026

The MIT AI Risk Repository: A Detailed Guide to the World’s Largest AI Risk Database

 MIT AI Risk Repository 

The MIT AI Risk Repository is a major research initiative created to provide the world’s most comprehensive, structured, and unified resource on risks posed by artificial intelligence. It functions as a living, continuously updated database of AI risks, taxonomies, and documented sources, developed by the MIT AI Risk Initiative / MIT FutureTech Group.

It is publicly accessible at airisk.mit.edu.

1. What the MIT AI Risk Repository Is

According to MIT, the AI Risk Repository is:

  • A centralized, living database of AI-related risks, currently listing 700–1700+ risks depending on the version referenced (MIT's web version lists 1700+, while the academic paper documents 777 risks).
  • Compiled from dozens of academic, government, and industry AI frameworks (43–74 frameworks, depending on the version).
  • Designed to create a shared vocabulary for researchers, policymakers, auditors, and companies when discussing AI risks.
  • Open-access and designed to be extensible, meaning new risks can be added as the field evolves. 

The repository aims to unify a fragmented AI governance landscape and support future policy, regulation, audits, and safe AI development practices.

2. Core Components of the Repository

MIT describes the repository as having three primary components:

A. The AI Risk Database

Contains:

  • 700–1700+ documented AI risks
  • Direct links to source material (papers, frameworks, reports)
  • Quotes and page numbers verifying each risk 

This database enables:

  • Filtering risks by type, cause, domain, or scenario
  • Downloading risks in formats like Google Sheets or OneDrive
  • Reviewing evidence and citations for each risk

B. The Causal Taxonomy of AI Risks

This taxonomy classifies how a risk arises based on three dimensions:

1. Entity

  • Human
  • AI
  • Other/ambiguous

2. Intentionality

  • Intentional
  • Unintentional
  • Undefined

3. Timing

  • Pre-deployment
  • Post-deployment
  • Unspecified

This answers:

Who caused the risk?

Was it intentional?

When does it arise?

C. The Domain Taxonomy of AI Risks

This organizes risks into 7 major domains and 23–24 subdomains.

The seven high-level domains are:

1. Discrimination & toxicity

2. Privacy & security

3. Misinformation

4. Malicious actors & misuse

5. Human-computer interaction issues

6. Socioeconomic & environmental impacts

7. AI system safety, failures & limitations

MIT notes, for example, that privacy and security risks appear in 70%+ of the reviewed frameworks, while risks such as AI rights and welfare appear in <1%.

3. How the Repository Was Created

The repository was built via a systematic meta-review of existing AI risk frameworks.

Researchers: Peter Slattery, Neil Thompson, and a multi-disciplinary MIT team. [ide.mit.edu], [arxiv.org]

The process involved:

1. Reviewing 43–74 AI governance documents

2. Extracting every explicit AI risk described

3. An expert consultation process

4. Creating high-level and mid-level taxonomies

5. Publishing the database and taxonomies openly

The academic paper describing this process is titled:

“The AI Risk Repository: A Comprehensive Meta‑Review, Database, and Taxonomy of Risks From Artificial Intelligence” (2024–2025). 

4. Why the MIT AI Risk Repository Matters

A. Establishes a Shared Language

The AI governance ecosystem is fragmented. Different industries, researchers, and governments use inconsistent terminology. The MIT repository unifies them under one standard. [mitsloan.mit.edu]

B. Improves AI Safety and Compliance

Organizations can use the repository to:

  • Identify relevant risks
  • Prioritize risk mitigation
  • Build audits and assessments
  • Improve AI governance frameworks

C. Helps Policymakers

Regulators can more clearly understand:

  • Where risks occur
  • How common they are
  • How they compare across industries 

D. Tracks Underexplored Risk Categories

For example, MIT found:

  • Privacy & security risks appear in >70% of risk frameworks
  • Misinformation risks appear in only ~40%
  • AI welfare/rights appear in <1% 

This highlights research gaps.

E. Supports Research, Education, and Standardization

The repository is used for:

  • Academic research
  • Policy development
  • Corporate risk audits
  • Curriculum design

5. Examples of Risks Found in the Repository

The repository documents risks across many categories, including:

  • Bias and discrimination in model outputs
  • Privacy breaches/data leakage
  • Deepfake misinformation
  • AI-enabled cyberattacks
  • Model hallucinations
  • Autonomous system failures
  • Socioeconomic displacement
  • Environmental resource consumption 

Each risk is paired with:

  • Citations
  • Exact quotes
  • Evidence
  • Categorization by taxonomy

6. How to Use the MIT AI Risk Repository

MIT suggests several uses:

  • Search for risks relevant to a specific AI system
  • Explore causal and domain factors to build risk models
  • Build governance frameworks and compliance plans
  • Teach AI risk management in educational settings
  • Monitor emerging risks as the database updates

7. Strengths and Limitations (Based on Research Commentary)

Strengths 

  • Open-access, transparent, regularly updated
  • Most comprehensive resource of its kind
  • Useful taxonomies (causal and domain-based)
  • Unified framework that integrates 700+ risks
  • Valuable for practical AI governance

Limitations

  • Some risks may be high-level or ambiguous
  • Interpretation depends on user expertise
  • Coverage of novel or speculative risks is still evolving
  • Some domains are underrepresented (e.g., AI rights)

8. Summary

The MIT AI Risk Repository is one of the most important AI governance tools available today. It combines:

  • A living database of 700–1700+ AI risks
  • A causal taxonomy explaining how risks arise
  • A domain taxonomy categorizing risk areas
  • Full citations and evidence
  • Open-access resources for researchers, businesses, auditors, and policymakers

Its purpose is to standardize AI risk vocabulary, support governance, and improve global understanding of AI risks in a rapidly evolving field.