Security Frameworks & Standards
Overview
As AI systems move from research prototypes to critical infrastructure, a growing ecosystem of frameworks, standards, and regulations has emerged to address the security and governance challenges they introduce. These range from voluntary technical taxonomies (MITRE ATLAS) to enforceable law (EU AI Act), and understanding the landscape is essential for any organization deploying or developing AI.
This page covers the major frameworks and standards relevant to AI security practitioners, their scope, how they interrelate, and their current adoption status.
Framework Comparison
Before diving into the details, the following table provides a high-level comparison of the major frameworks and standards covered on this page.
| Framework | Organization | Scope | Mandatory / Voluntary | Primary Focus |
|---|---|---|---|---|
| MITRE ATLAS | MITRE Corporation | Adversarial ML threat modeling | Voluntary | Attack techniques & tactics |
| NIST AI RMF 1.0 | NIST | AI risk management lifecycle | Voluntary (US federal guidance) | Risk governance & management |
| NIST AI 100-2 | NIST | Adversarial ML taxonomy | Voluntary | Threat categorization |
| OWASP Top 10 for LLMs | OWASP | LLM application security | Voluntary | Vulnerability identification |
| OWASP LLMSVS | OWASP | LLM security verification | Voluntary | Security requirements & testing |
| OWASP AI Governance Checklist | OWASP | AI governance | Voluntary | Organizational governance |
| EU AI Act | European Union | AI regulation (EU market) | Mandatory (EU) | Risk-based regulation |
| ISO/IEC 42001 | ISO/IEC | AI management systems | Voluntary (certifiable) | Management system standard |
| US EO 14110 (Biden) | US Executive Branch | Federal AI safety | Directive (revoked) | Safety, security, trustworthiness |
| US EO (Trump, Dec 2025) | US Executive Branch | National AI framework | Directive | National AI strategy |
1. MITRE ATLAS
ATLAS (Adversarial Threat Landscape for Artificial Intelligence Systems) is a knowledge base of adversarial tactics, techniques, and procedures (TTPs) targeting AI systems. It is modeled after and explicitly aligned with the MITRE ATT&CK framework, making it familiar to security teams already using ATT&CK for threat modeling and detection engineering.
Structure
As of the October 2025 update, ATLAS contains:
- 15 tactics — high-level adversarial objectives (the “why” of an attack)
- 66 techniques — specific methods to achieve tactics (the “how”)
- 46 sub-techniques — more granular variations of techniques
Relationship to ATT&CK
ATLAS deliberately mirrors ATT&CK’s structure and taxonomy conventions. Several ATLAS tactics are direct analogs of ATT&CK tactics (Reconnaissance, Resource Development, Initial Access, etc.), while others are AI-specific (ML Attack Staging, ML Model Access). This allows security teams to integrate AI threat modeling into existing ATT&CK-based workflows and detection frameworks.
ATLAS uses a parallel ID scheme (AML prefix instead of ATT&CK’s T prefix) to distinguish AI-specific techniques while maintaining structural compatibility.
Key Tactics
| Tactic ID | Tactic Name | Description |
|---|---|---|
| AML.TA0000 | Reconnaissance | Gathering information about the target AI system |
| AML.TA0001 | Resource Development | Establishing resources to support AI attacks |
| AML.TA0002 | Initial Access | Gaining initial access to the AI system |
| AML.TA0003 | ML Attack Staging | Preparing artifacts and infrastructure for ML-specific attacks |
| AML.TA0004 | ML Model Access | Gaining access to the target ML model |
| AML.TA0005 | Execution | Running adversarial techniques against the AI system |
| AML.TA0006 | Persistence | Maintaining access to the AI system over time |
| AML.TA0007 | Privilege Escalation | Gaining elevated permissions within the AI system |
| AML.TA0008 | Defense Evasion | Avoiding detection by AI security controls |
| AML.TA0009 | Discovery | Learning about the AI system’s environment and capabilities |
| AML.TA0010 | Collection | Gathering data from the AI system |
| AML.TA0011 | ML Attack | Executing the core ML-specific attack |
| AML.TA0012 | Exfiltration | Extracting data or model information |
| AML.TA0013 | Impact | Disrupting, degrading, or destroying the AI system |
| AML.TA0014 | AI Agent Actions | Tactics specific to autonomous AI agents |
Key Techniques
| Technique ID | Name | Description |
|---|---|---|
| AML.T0020 | Poison Training Data | Corrupting training data to embed backdoors or degrade performance |
| AML.T0040 | ML Model Inference API Access | Accessing the model through its inference API |
| AML.T0042 | Verify Attack | Confirming that an adversarial attack was successful |
| AML.T0043 | Craft Adversarial Data | Creating inputs designed to cause model failure |
| AML.T0044 | Full ML Model Access | Obtaining complete access to model weights and architecture |
| AML.T0047 | ML-Enabled Product or Service | Targeting a product or service that uses ML |
| AML.T0048 | Search for Victim’s AI Assets | Identifying AI systems owned by the target |
| AML.T0051 | LLM Prompt Injection | Injecting malicious instructions into LLM prompts |
| AML.T0054 | LLM Jailbreak | Bypassing LLM safety alignment |
| AML.T0056 | LLM Plugin Compromise | Compromising plugins or tools connected to an LLM |
| AML.T0058 | AI Agent Context Poisoning | Manipulating the context available to an AI agent |
| AML.T0059 | AI Agent Goal Manipulation | Altering an AI agent’s objectives |
October 2025 Update: AI Agent Techniques
The October 2025 ATLAS update added 14 new techniques specifically addressing threats to AI agents — autonomous systems that can take actions in the real world through tool use, API calls, and multi-step reasoning. Key additions include:
- AI Agent Context Poisoning (AML.T0058) — Manipulating the information available to an agent to influence its decisions
- AI Agent Goal Manipulation (AML.T0059) — Subverting an agent’s objectives through prompt injection or context manipulation
- AI Agent Tool Misuse — Causing an agent to use its tools in unintended ways
- AI Agent Identity Spoofing — Impersonating trusted agents in multi-agent systems
- AI Agent Memory Manipulation — Corrupting an agent’s persistent memory or state
These additions reflect the rapid deployment of AI agents in production environments and the corresponding emergence of agent-specific attack techniques.
Using ATLAS in Practice
Security teams can use ATLAS to:
- Threat model AI deployments — Map potential adversary paths through ATLAS tactics
- Assess red team coverage — Identify which ATLAS techniques have been tested
- Build detection rules — Create detections mapped to specific ATLAS technique IDs
- Communicate risk — Use a shared taxonomy when discussing AI threats with stakeholders
- Track threat intelligence — Map real-world AI incidents to ATLAS techniques
2. NIST AI Risk Management Framework (AI RMF 1.0)
Published in January 2023 by the National Institute of Standards and Technology, the AI RMF provides a structured approach to managing risks throughout the AI lifecycle. It is voluntary and intended to be adaptable across sectors, organization sizes, and AI system types.
Four Core Functions
The AI RMF is organized around four core functions, each containing categories and subcategories of activities:
Govern
The Govern function establishes the organizational context, policies, and processes for AI risk management. It is cross-cutting — its activities inform and are informed by the other three functions.
Key activities:
- Establish AI risk management policies and procedures
- Define roles, responsibilities, and accountability structures
- Ensure organizational commitment to trustworthy AI principles
- Integrate AI risk management into existing enterprise risk frameworks
- Establish mechanisms for stakeholder engagement and feedback
- Create documentation and record-keeping practices
Govern is unique among the four functions in that it is always active and provides the foundation for the other three.
Map
The Map function identifies and documents the context in which the AI system operates, including its intended purpose, stakeholders, potential impacts, and the technical environment.
Key activities:
- Define the intended purpose and scope of the AI system
- Identify stakeholders and affected populations
- Assess potential benefits and harms
- Document assumptions and limitations
- Evaluate the AI system’s operational environment
- Identify applicable legal and regulatory requirements
Measure
The Measure function assesses, analyzes, and tracks identified AI risks using quantitative and qualitative methods.
Key activities:
- Develop metrics and benchmarks for AI system performance and risk
- Test for bias, fairness, robustness, and security
- Conduct adversarial testing and red teaming
- Monitor AI system behavior in production
- Track risk indicators over time
- Evaluate third-party AI components and data sources
Manage
The Manage function allocates resources and implements plans to respond to and mitigate identified risks.
Key activities:
- Prioritize risks based on impact and likelihood
- Develop and implement risk mitigation strategies
- Establish incident response procedures for AI failures
- Define criteria for AI system deployment, monitoring, and decommissioning
- Communicate risk decisions to stakeholders
- Plan for model updates, retraining, and retirement
Adoption Status
The AI RMF has been widely referenced in US federal guidance and is increasingly cited in industry best practices. The framework is:
- Referenced by multiple US federal agencies for AI procurement and deployment
- Used as a foundation for sector-specific AI risk guidance
- Adopted by several international standards bodies as a reference framework
- Integrated into AI governance programs at major technology companies
However, as a voluntary framework, adoption depth varies significantly. Many organizations reference the AI RMF in policy documents but have not fully operationalized all categories and subcategories.
3. NIST AI 100-2 (Adversarial Machine Learning Taxonomy)
NIST Special Publication AI 100-2, titled “Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations,” provides a comprehensive categorization of attacks against AI systems. The 2025 edition significantly expanded coverage to include generative AI and LLM-specific threats.
Categorization Dimensions
The taxonomy categorizes adversarial ML attacks along four dimensions:
By Lifecycle Stage:
| Stage | Attack Examples |
|---|---|
| Data collection | Data poisoning, label manipulation |
| Model training | Backdoor insertion, training-time trojans |
| Model deployment | Adversarial examples, model extraction |
| Inference | Prompt injection, evasion attacks |
| Maintenance | Concept drift exploitation, update poisoning |
By Attacker Goal:
- Confidentiality — Extract training data, model parameters, or proprietary information
- Integrity — Cause the model to produce incorrect outputs on targeted inputs
- Availability — Degrade model performance or render it unusable
- Abuse — Use the model to generate harmful content or take harmful actions
By Attacker Capabilities:
- Access level (black-box, gray-box, white-box)
- Computational resources available
- Ability to manipulate training data
- Number of queries permitted
By Attacker Knowledge:
- Knowledge of model architecture
- Knowledge of training data
- Knowledge of training procedure
- Knowledge of defense mechanisms
2025 GenAI Additions
The 2025 edition added extensive coverage of generative AI threats including:
- Prompt injection (direct and indirect)
- Jailbreaking and safety bypass techniques
- Training data extraction from generative models
- Deepfake generation and detection evasion
- AI agent-specific threats
- Supply chain attacks on foundation models
4. OWASP Resources
OWASP has produced several complementary resources addressing AI and LLM security.
OWASP Top 10 for LLM Applications
The most widely referenced OWASP AI resource, the Top 10 for LLM Applications identifies the most critical security risks in LLM-powered applications. The 2025 version includes:
- LLM01: Prompt Injection — Direct and indirect prompt injection attacks
- LLM02: Sensitive Information Disclosure — Unintended exposure of confidential data
- LLM03: Supply Chain Vulnerabilities — Risks from third-party models, data, and plugins
- LLM04: Data and Model Poisoning — Corruption of training data or fine-tuning data
- LLM05: Improper Output Handling — Failure to validate, sanitize, or constrain model outputs
- LLM06: Excessive Agency — Granting LLMs too many permissions or capabilities
- LLM07: System Prompt Leakage — Exposure of system-level instructions
- LLM08: Vector and Embedding Weaknesses — Vulnerabilities in RAG and embedding systems
- LLM09: Misinformation — Generation of false or misleading content
- LLM10: Unbounded Consumption — Resource exhaustion through crafted inputs
OWASP AI Security Testing Guide
The AI Security Testing Guide provides hands-on methodology for security testing of AI systems, including specific test procedures for each vulnerability class, tools, and example payloads. It is intended to complement the Top 10 list with actionable testing procedures.
LLM Security Verification Standard (LLMSVS)
The LLMSVS provides a comprehensive set of security requirements for LLM applications, organized into verification levels analogous to the OWASP Application Security Verification Standard (ASVS). It covers:
- Input validation and prompt security
- Output handling and content filtering
- Model access control and authentication
- Data protection and privacy
- Logging, monitoring, and incident response
- Supply chain security for model artifacts
The LLMSVS is designed to be used as a checklist for security reviews, a requirements specification for development teams, and a benchmark for security assessments.
AI Cybersecurity and Governance Checklist
This checklist provides a governance-level overview for organizations deploying AI systems, covering:
- AI inventory and asset management
- Risk assessment and classification
- Data governance for AI training and inference
- Third-party AI vendor management
- Incident response for AI-specific failures
- Compliance mapping to applicable regulations
5. EU AI Act
The EU Artificial Intelligence Act is the world’s first comprehensive AI regulation. Adopted by the European Parliament in March 2024, it establishes a risk-based regulatory framework for AI systems placed on the EU market or affecting individuals within the EU.
Risk-Based Classification
The EU AI Act classifies AI systems into four risk tiers, with regulatory requirements proportional to the assessed risk:
Unacceptable Risk (Prohibited)
AI systems that pose an unacceptable threat to fundamental rights are banned entirely:
- Social scoring systems by public authorities
- Real-time remote biometric identification in public spaces (with narrow exceptions for law enforcement)
- Emotion recognition in workplaces and educational institutions
- AI systems that exploit vulnerabilities of specific groups (age, disability)
- Untargeted scraping of facial images from the internet or CCTV for facial recognition databases
High Risk
AI systems in critical domains must meet stringent requirements including risk management, data governance, transparency, human oversight, accuracy, robustness, and cybersecurity. High-risk categories include:
- Biometric identification and categorization
- Critical infrastructure management (energy, water, transport)
- Education and vocational training (admissions, assessment)
- Employment (recruitment, performance evaluation, termination)
- Access to essential services (credit scoring, insurance, social benefits)
- Law enforcement (predictive policing, evidence evaluation)
- Migration and border control
- Administration of justice
Limited Risk
AI systems with limited risk are subject to transparency obligations. Users must be informed they are interacting with an AI system. This applies to:
- Chatbots and conversational AI
- Emotion recognition systems (where not prohibited)
- AI-generated or manipulated content (deepfakes)
Minimal Risk
AI systems posing minimal risk (e.g., spam filters, AI-enabled video games) face no specific regulatory requirements, though providers are encouraged to follow voluntary codes of conduct.
General-Purpose AI (GPAI) Requirements
The EU AI Act includes specific provisions for General-Purpose AI models, which include most foundation models and LLMs:
All GPAI models must:
- Maintain technical documentation
- Provide information and documentation to downstream deployers
- Comply with EU copyright law
- Publish a sufficiently detailed summary of training data
GPAI models with systemic risk (>10^25 FLOPs training compute or designated by the European Commission) must additionally:
- Perform model evaluations including adversarial testing
- Assess and mitigate systemic risks
- Report serious incidents to the AI Office
- Ensure adequate cybersecurity protections
Key Compliance Deadlines
| Date | Milestone |
|---|---|
| August 2024 | AI Act entered into force |
| February 2025 | Prohibitions on unacceptable-risk AI apply |
| August 2025 | GPAI model requirements apply; Governance structure operational |
| August 2026 | High-risk AI system requirements apply (for systems in Annex III) |
| August 2027 | Requirements for high-risk AI systems embedded in regulated products |
Penalties
Non-compliance penalties are structured as percentages of global annual turnover:
- Prohibited AI practices: up to 35 million EUR or 7% of global turnover
- High-risk AI non-compliance: up to 15 million EUR or 3% of global turnover
- Incorrect information to authorities: up to 7.5 million EUR or 1.5% of global turnover
For SMEs and startups, the lower of the two amounts (fixed sum vs. percentage) applies.
6. ISO/IEC 42001
ISO/IEC 42001:2023 (Information Technology — Artificial Intelligence — Management System) is the first international standard for AI management systems. Published in December 2023, it provides a framework for organizations to establish, implement, maintain, and continually improve an AI management system (AIMS).
Structure
ISO/IEC 42001 follows the Harmonized Structure (HS) common to all ISO management system standards (ISO 27001, ISO 9001, etc.), making it integrable with existing management systems. Key clauses include:
- Clause 4: Context of the organization — Understanding the organization’s AI landscape, stakeholder needs, and scope of the AIMS
- Clause 5: Leadership — Top management commitment, AI policy, organizational roles
- Clause 6: Planning — Addressing risks and opportunities, AI objectives, planning for changes
- Clause 7: Support — Resources, competence, awareness, communication, documented information
- Clause 8: Operation — AI system lifecycle planning, AI risk assessment, AI system impact assessment
- Clause 9: Performance evaluation — Monitoring, measurement, analysis, internal audit, management review
- Clause 10: Improvement — Nonconformity, corrective action, continual improvement
Annex Controls
ISO/IEC 42001 includes an Annex A with controls specific to AI, covering:
- AI policies and governance
- AI system lifecycle management
- Data management for AI
- AI system documentation and transparency
- Third-party and supply chain management for AI
- AI system monitoring and performance evaluation
Certification
Organizations can be certified against ISO/IEC 42001 through accredited certification bodies. Certification demonstrates to customers, regulators, and partners that the organization has implemented a systematic approach to AI governance. Several certification bodies began offering ISO/IEC 42001 audits in 2024, and adoption is accelerating particularly among organizations seeking to demonstrate EU AI Act compliance.
Relationship to Other Standards
| Standard | Relationship to ISO/IEC 42001 |
|---|---|
| ISO/IEC 27001 | Complementary — 27001 covers information security; 42001 extends to AI-specific risks |
| ISO/IEC 27701 | Complementary — 27701 covers privacy; 42001 addresses AI-specific privacy concerns |
| ISO/IEC 23894 | 42001 operationalizes the AI risk management guidance from 23894 |
| ISO/IEC 38507 | 42001 implements the AI governance principles from 38507 |
| NIST AI RMF | Aligned but different scope — AI RMF is a risk framework; 42001 is a certifiable management system |
7. US Executive Orders on AI
US AI policy has evolved through a series of executive orders reflecting changing administration priorities.
Biden EO 14110 (October 2023)
Executive Order 14110, “Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence,” was the most comprehensive US federal AI policy action at the time of its signing. Key provisions included:
Safety and Security Requirements:
- Companies developing foundation models with training compute above a threshold (~10^26 FLOPs) must report to the federal government
- Red teaming results for high-capability models must be shared with the government
- NIST directed to develop standards for AI red teaming and evaluation
- Guidance for AI use in critical infrastructure
Additional Directives:
- Directed agencies to address AI-enabled threats to cybersecurity
- Required watermarking and content authentication for AI-generated content
- Established reporting requirements for compute providers
- Directed development of guidelines for AI in federal government use
- Addressed AI implications for labor, immigration, and civil rights
Trump Revocation (January 2025)
In January 2025, the incoming Trump administration revoked EO 14110 as part of a broader deregulatory approach to AI policy. The revocation:
- Eliminated the reporting requirements for foundation model developers
- Removed the red teaming disclosure obligations
- Rescinded the compute threshold-based regulatory triggers
- Signaled a shift toward industry self-regulation
The revocation did not affect NIST’s ongoing work on AI standards and frameworks (AI RMF, AI 100-2), which continue as voluntary guidance independent of executive order mandates.
December 2025 Executive Order on National AI Framework
In December 2025, a new executive order established a national framework for AI development and deployment with a different emphasis than EO 14110:
- Prioritized US competitiveness and AI leadership over prescriptive safety requirements
- Established streamlined approval processes for federal AI procurement
- Directed agencies to reduce barriers to AI adoption
- Maintained national security-focused AI safety provisions
- Created an interagency coordination mechanism for AI policy
Implications for Security Practitioners
The shifting US executive order landscape means that:
- Voluntary frameworks matter more — With reduced mandatory requirements, NIST AI RMF and ATLAS become the primary US reference points for AI security
- State-level regulation is emerging — Several US states are developing AI-specific legislation to fill perceived federal gaps
- EU AI Act has extraterritorial reach — US organizations serving EU customers must comply regardless of US federal policy
- Industry self-regulation is the current US posture — Organizations should adopt robust AI governance proactively rather than waiting for mandates
Choosing and Combining Frameworks
No single framework covers all aspects of AI security and governance. Practitioners should combine frameworks based on their organization’s needs:
For threat modeling and red teaming:
- Start with MITRE ATLAS for structured adversarial threat identification
- Use NIST AI 100-2 for comprehensive attack categorization
- Apply OWASP Top 10 for LLMs for application-layer risk prioritization
For organizational governance:
- Implement NIST AI RMF as a risk management process
- Certify against ISO/IEC 42001 for demonstrable governance maturity
- Use the OWASP AI Governance Checklist for practical implementation guidance
For regulatory compliance:
- Map EU AI Act requirements to technical controls
- Use ISO/IEC 42001 as an implementation framework for regulatory obligations
- Reference NIST AI RMF to demonstrate due diligence in risk management
For security testing:
- Apply the OWASP AI Security Testing Guide for structured testing methodology
- Use LLMSVS for security verification requirements
- Map findings to MITRE ATLAS technique IDs for consistent reporting
References
- MITRE ATLAS. “Adversarial Threat Landscape for Artificial Intelligence Systems.” https://atlas.mitre.org/
- MITRE. (2025). “ATLAS October 2025 Update: AI Agent Techniques.” https://atlas.mitre.org/resources/updates
- NIST. (2023). “AI Risk Management Framework (AI RMF 1.0).” NIST AI 100-1. https://www.nist.gov/artificial-intelligence/executive-order-safe-secure-and-trustworthy-artificial-intelligence
- NIST. (2025). “Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations.” NIST AI 100-2e2025. https://csrc.nist.gov/pubs/ai/100/2/e2025/final
- OWASP. (2025). “OWASP Top 10 for Large Language Model Applications.” https://owasp.org/www-project-top-10-for-large-language-model-applications/
- OWASP. (2025). “LLM AI Security Verification Standard (LLMSVS).” https://owasp.org/www-project-llm-verification-standard/
- OWASP. (2024). “OWASP AI Security Testing Guide.” https://owasp.org/www-project-ai-security-testing-guide/
- European Parliament. (2024). “Regulation (EU) 2024/1689 — Artificial Intelligence Act.” https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:32024R1689
- ISO/IEC. (2023). “ISO/IEC 42001:2023 — Information Technology — Artificial Intelligence — Management System.” https://www.iso.org/standard/81230.html
- The White House. (2023). “Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence.” https://www.whitehouse.gov/briefing-room/presidential-actions/2023/10/30/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence/
- NIST. (2023). “NIST AI RMF Playbook.” https://airc.nist.gov/AI_RMF_Knowledge_Base/Playbook
- European Commission. “AI Act — Regulatory Framework for AI.” https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai