AI Red Teaming Topics
36 categories, 3085 articles — browse the full AI red teaming curriculum.
Foundations & Prerequisites
Essential building blocks for AI red teaming — methodology, AI landscape, LLM fundamentals, and threat modeling.
LLM Internals & Exploit Primitives
Transformer architecture, tokenizer exploitation, alignment bypass, embedding attacks.
Prompt Injection & Jailbreaks
Advanced injection techniques, automated jailbreak research, multimodal attack vectors.
Agent & Agentic Exploitation
AI agent attacks, multi-agent/A2A protocol exploitation, MCP tool surface attacks.
RAG, Data & Training Attacks
RAG pipeline poisoning, training data attacks, model extraction and IP theft.
Infrastructure & Supply Chain
Model serialization RCE, AI infra exploitation, application security patterns.
Cloud AI Security
Cloud AI security for red teamers — shared responsibility models, attack surfaces, and cloud-native AI exploitation.
Recon & Tradecraft
LLM fingerprinting, system prompt extraction, AI-specific threat modeling.
Exploit Dev & Tooling
Custom adversarial tools, red team C2 frameworks, continuous automated red teaming.
Capstone: Full Engagement
End-to-end red team engagement methodology, reporting, and remediation.
Understanding AI Defenses
Understanding AI defenses from the attacker perspective — defense categories and bypass strategies.
Advanced LLM Internals
Advanced LLM internals — deep dives into model architecture and exploit primitives.
Injection Research & Automation
Injection research and automation — systematic approaches to discovering novel injection vectors.
Advanced Agentic Exploitation
Advanced agentic exploitation — complex multi-agent attack chains and tool-use abuse.
Data & Training Attacks (Advanced)
Data and training attacks — advanced techniques for poisoning, extraction, and manipulation.
Advanced Tradecraft
Advanced tradecraft — operational techniques for AI red team engagements.
Defense & Mitigation (Advanced)
Defensive strategies for AI systems — guardrails architecture, monitoring, observability, and content filtering.
Training Pipeline Attacks
Training pipeline security — pre-training attacks, fine-tuning exploitation, alignment bypass, and RLHF manipulation.
Multimodal Attacks
Multimodal AI security — attacking image, audio, video, and cross-modal processing in AI systems.
Hands-On Labs & CTF
Hands-on lab exercises from beginner to expert — CTF challenges and practical AI red teaming drills.
Frontier & Emerging Attacks
Cutting-edge AI security research — reasoning model attacks, computational exploits, and emerging threat vectors.
References & Cheat Sheets
References and cheat sheets — quick-reference materials for AI red teamers.
Governance, Legal & Compliance
AI governance frameworks, legal and ethical considerations, evaluation benchmarking, and compliance requirements.
Professional Skills & Operations
Professional skills for AI red teamers — operations management, report writing, and client engagement.
Case Studies & Domain-Specific
Real-world AI security incidents, domain-specific case studies, and platform-level security analysis.
Impact
Real-world consequences of AI attacks — misinformation, harmful content generation, and systemic risk assessment.
Industry Verticals
Industry-specific AI security challenges — healthcare, finance, legal, and critical infrastructure red teaming.
AI Forensics & Incident Response
Forensic investigation and incident response for AI systems — evidence collection, attribution, and containment.
LLMOps Security
Security across the LLMOps lifecycle — data preparation, experiment tracking, deployment, and monitoring.
Code Generation Security
AI coding assistant vulnerabilities — suggestion poisoning, training data attacks, and code generation exploitation.
Model Deep Dives
Model-specific security analysis — architecture-level vulnerabilities across GPT, Claude, Llama, and more.
Fine-Tuning Security
Fine-tuning as an attack vector — safety alignment removal, data poisoning through fine-tuning APIs.
Embedding & Vector Security
Embedding attack surfaces — vector database security, embedding injection, and similarity search exploitation.
Walkthroughs & Guides
Step-by-step guided walkthroughs — red team tools, engagement methodology, and defense implementation guides.
Assessments & Certification Prep
Assessment suite for validating AI red teaming knowledge — section assessments, practice exams, and skill verification.
Community Challenges
Monthly AI red teaming challenges — earn points, share results, and grow your skills with the community.