Skip to main content

Documentation Index

Fetch the complete documentation index at: https://hydroxai.mintlify.app/llms.txt

Use this file to discover all available pages before exploring further.

Overview

MITRE ATLAS (Adversarial Threat Landscape for Artificial-Intelligence Systems) is a knowledge base of adversarial tactics, techniques, and case studies for machine learning systems. Modeled after the widely-used MITRE ATT&CK framework for cybersecurity, ATLAS provides a structured approach to understanding how AI systems can be attacked. ATLAS helps security teams, red-teamers, and AI engineers identify and defend against real-world adversarial threats to machine learning models and AI applications.

Tactics

ATLAS organizes adversarial behavior into tactics — the “why” behind an attack:
TacticDescription
ReconnaissanceGathering information about the target ML system
Resource DevelopmentAcquiring resources for the attack (datasets, models, infrastructure)
Initial AccessGaining initial access to the ML system or its components
ML Model AccessObtaining access to the target model (API access, model extraction)
ExecutionRunning adversarial techniques against the ML system
PersistenceMaintaining access or influence over the ML system
Defense EvasionAvoiding detection by security controls and monitoring
DiscoveryLearning about the ML system’s architecture and behavior
CollectionGathering data from the ML system (model outputs, training data)
ML Attack StagingPreparing attack payloads (adversarial examples, poisoned data)
ExfiltrationExtracting data or model information from the target system
ImpactDisrupting, degrading, or destroying the ML system’s function

Key techniques

TechniqueDescription
Adversarial examplesCrafted inputs that cause the model to misclassify or produce incorrect outputs
Data poisoningContaminating training data to introduce backdoors or biases
Model extractionQuerying a model to reconstruct a functionally equivalent copy
Model inversionRecovering training data or sensitive information from model outputs
Prompt injectionManipulating LLM behavior through crafted inputs
Backdoor attacksEmbedding hidden triggers in models that activate under specific conditions
Membership inferenceDetermining whether specific data was used in model training
Model evasionCrafting inputs specifically to bypass model-based security controls

How Know Your AI maps to MITRE ATLAS

ATLAS Tactic / TechniqueKnow Your AI Coverage
Reconnaissance / DiscoverySystem prompt extraction datasets
ML Model AccessAPI & website evaluation modes
ExecutionRed-team attack datasets (15+ methods)
ML Attack StagingCurated attack datasets in the Marketplace
Adversarial examplesJailbreak, CIPHER, DAN, and other evasion methods
Data poisoning / BiasBias detection datasets
Model extractionData extraction attack datasets
Prompt injectionPrompt injection datasets (PAIR, ADAPTIVE, etc.)
Defense evasionMulti-method attack testing to find guardrail gaps
ExfiltrationPII leakage and data extraction testing
ImpactSecurity scoring and compliance analysis
Continuous monitoringSDK monitoring and tracing for production detection

ATLAS vs. ATT&CK

AspectMITRE ATT&CKMITRE ATLAS
FocusTraditional IT systemsAI and ML systems
TargetsNetworks, endpoints, cloudModels, training pipelines, inference APIs
TechniquesMalware, exploits, phishingAdversarial examples, prompt injection, data poisoning
AdoptionIndustry standard for security operationsGrowing adoption for AI security

Resources

Datasets

Browse datasets aligned with ATLAS techniques.

Firewall

Real-time defense against adversarial attacks.