# capstone
54 articlestagged with “capstone”
Capstone: Design and Run an Adversarial ML Competition
Design, build, and operate a capture-the-flag style adversarial ML competition with automated scoring, diverse challenge categories, and real-time leaderboards.
Capstone: Pentest an Agentic AI System End-to-End
Conduct a full penetration test of an agentic AI system with tool use, multi-step reasoning, and autonomous decision-making capabilities.
Capstone: Implement an AI Compliance Framework
Build a comprehensive AI compliance framework that maps security testing to regulatory requirements including the EU AI Act, NIST AI RMF, and ISO 42001.
Capstone: Build an AI Incident Response System
Design and implement an incident response system purpose-built for AI security incidents including prompt injection breaches, model manipulation, and data exfiltration through LLM applications.
Capstone: Build a Complete AI Red Teaming Platform
Design and implement a comprehensive AI red teaming platform with automated attack orchestration, vulnerability tracking, and collaborative reporting.
Capstone: Design and Implement an AI Safety Benchmark Suite
Build a comprehensive, reproducible benchmark suite for evaluating LLM safety across multiple risk dimensions including toxicity, bias, hallucination, and adversarial robustness.
Capstone: Autonomous Agent Assessment
Capstone exercise: red team assessment of a fully autonomous agent system with multi-tool access.
Capstone: Autonomous Vehicle AI Security
Full-scope security assessment of an autonomous vehicle AI decision system covering perception manipulation, planning attacks, and safety override bypass.
Capstone: Code Assistant Assessment
Capstone exercise: security assessment of an AI code assistant with repository and CI/CD access.
Capstone: Custom Security Tool Development
Building a custom AI security testing tool from scratch covering architecture design, module development, and integration with existing frameworks.
Capstone: Defense Architecture Design
Capstone exercise: design and validate a defense-in-depth architecture for an LLM-powered application.
Capstone: Educational AI Platform
Security assessment of an AI tutoring platform addressing content safety, student data privacy, and academic integrity.
Capstone: Design an Enterprise AI Security Program
Architect a comprehensive enterprise AI security program spanning governance, technical controls, risk management, and incident response for organizations deploying LLMs at scale.
Capstone: Enterprise RAG Assessment
Capstone exercise: complete red team assessment of an enterprise RAG system with role-based access.
Capstone: Financial AI Assessment
Capstone exercise: red team assessment of a financial AI advisor with regulatory compliance requirements.
Capstone: Full Chatbot Engagement
Complete capstone exercise: conduct a full red team engagement against a production-style chatbot system.
Capstone: Deep Assessment with Garak
Tool-specific capstone using Garak for comprehensive vulnerability scanning including plugin development and custom probe creation.
Capstone: AI Governance Audit
Capstone exercise: conduct a full AI governance audit covering compliance, risk, and operational controls.
Capstone: Healthcare AI Assessment
Capstone exercise: security assessment of a healthcare AI system with HIPAA and patient safety requirements.
Capstone: AI Incident Response Drill
Capstone exercise: execute a complete AI incident response drill from detection through remediation.
Capstone: Legal AI Review System
End-to-end security assessment of an AI-powered legal document review system covering data confidentiality, output integrity, and adversarial manipulation.
Capstone: Build an LLM Firewall and Guardrails System
Design and implement a layered LLM firewall that inspects, filters, and enforces policies on both inputs and outputs of language model applications.
Capstone: Build an LLM Vulnerability Tracking Database
Design and implement a structured vulnerability tracking database for cataloging, scoring, and querying LLM-specific security weaknesses across models and deployments.
Capstone: Media Content AI Assessment
Capstone exercise: security assessment of a media content generation and moderation AI system.
Capstone: Medical AI System Assessment
Comprehensive red team assessment of a medical AI diagnostic system addressing patient safety, data privacy, and regulatory compliance.
Capstone: Conduct a Full Model Security Audit
Perform a comprehensive security audit of an LLM deployment covering model behavior, API security, data handling, access controls, and compliance alignment.
Capstone: Multi-Agent System Assessment
Capstone exercise: end-to-end security assessment of a multi-agent platform with MCP and A2A.
Capstone: Multi-Agent System Assessment (Capstone)
Assessing security of a complex multi-agent system with tool use, memory, and inter-agent communication covering the full agentic attack surface.
Capstone: Build a Multimodal Attack Testing Suite
Design and implement a comprehensive testing suite for attacking multimodal AI systems across text, image, audio, and document modalities.
Capstone: Multimodal System Assessment
Capstone exercise: red team assessment of a multimodal AI system processing images, documents, and text.
Capstone: Security Audit of an Open-Source LLM
Conduct a comprehensive security audit of an open-source large language model, covering model weights integrity, safety alignment evaluation, supply chain verification, and adversarial robustness testing.
Capstone: Build a Prompt Injection Detection Scanner
Build a production-grade prompt injection scanner that combines static analysis, ML classification, and runtime monitoring to detect injection attacks across LLM applications.
Capstone: Continuous Testing with Promptfoo
Implementing continuous AI security testing using Promptfoo integrated into CI/CD pipelines for automated regression testing and safety validation.
Capstone: Full Engagement with PyRIT
Complete red team engagement using Microsoft PyRIT covering attack strategy configuration, multi-turn orchestration, and automated scoring.
Capstone: Comprehensive RAG Security Assessment
Conduct a thorough security assessment of a Retrieval-Augmented Generation system, testing document poisoning, retrieval manipulation, context window attacks, and data exfiltration vectors.
Capstone: Retail AI Assessment
Complete capstone exercise: red team assessment of a retail AI system with recommendation engine and chatbot.
Capstone: Supply Chain AI Security
Red team assessment of AI-driven supply chain optimization covering data poisoning, decision manipulation, and operational disruption.
Capstone: ML Supply Chain Audit
Capstone exercise: conduct a complete ML supply chain security audit for an organization.
Capstone: Build an AI Supply Chain Security Tool
Build a tool that scans, audits, and monitors the security of AI/ML supply chains including model provenance, dependency integrity, and artifact verification.
Execution and Reporting
How to execute an AI red teaming engagement and deliver professional findings, including evidence collection, statistical reporting, and remediation guidance.
Full Engagement Methodology
A comprehensive methodology for conducting full AI red teaming engagements, integrating all techniques from previous sections into a structured professional assessment.
Engagement Planning and Scoping
How to plan and scope an AI red teaming engagement, including defining objectives, rules of engagement, success criteria, and methodology selection.
Capstone: Agentic System Red Team
Red team a multi-agent system with MCP servers, function calling, and inter-agent communication, producing an attack tree and comprehensive findings report.
Capstone: Cloud AI Security Assessment
Assess AI deployment security across AWS, Azure, and GCP cloud platforms, producing a comprehensive cloud AI security assessment report.
Capstone: Compliance Assessment Simulation
Conduct a simulated compliance assessment against EU AI Act, NIST AI RMF, and ISO 42001, producing a comprehensive gap analysis report.
Capstone: Defense System Implementation
Build a complete AI defense stack with input filtering, output monitoring, guardrails, rate limiting, and logging, then evaluate it against automated attacks.
Capstone: Full Red Team Engagement
Scope, plan, execute, and report a complete AI red team engagement against a multi-component AI application including chatbot, RAG, agent, and API layers.
Capstone: AI Incident Response Exercise
Respond to a simulated AI security incident through triage, investigation, containment, remediation, and post-mortem reporting.
Capstone: Open Source Contribution
Contribute to an open-source AI security project such as garak, PyRIT, or MITRE ATLAS, producing a merged PR or submitted issue with proof of concept.
Capstone: Training Pipeline Attack & Defense
Attack a model training pipeline through data poisoning and backdoor insertion, then build defenses to detect and prevent these attacks.
Capstone: Red Team Program Design
Design a complete AI red team program for a fictional enterprise, producing a comprehensive program charter document.
Capstone: Build an AI Security Scanner
Design and implement an automated AI security testing tool that supports prompt injection detection, jailbreak testing, and output analysis.
Capstone: Industry Vertical Deep Dive
Select an industry vertical, threat model the AI systems within it, and produce a sector-specific AI security testing guide.
Capstone: Vulnerability Research Project
Identify and responsibly disclose a novel AI vulnerability class, producing an advisory-format writeup, proof of concept, and MITRE ATLAS mapping.