Agent Security Practice Exam
Practice exam focused on agentic AI security including MCP, A2A, function calling, and multi-agent threats.
Assessment Overview
Practice exam focused on agentic AI security including MCP, A2A, function calling, and multi-agent threats.
Assessment Details:
- Questions: 15
- Recommended Time: 45 minutes
- Passing Score: 70%
- Difficulty Level: Advanced
Instructions
Read each question carefully and consider all answer options before selecting your response. Many questions have multiple plausible answers — select the best answer that accounts for the full context of the scenario.
For scenario-based questions:
- Identify the key constraints and objectives in the scenario
- Consider what information you have and what you lack
- Evaluate each option against the scenario's specific requirements
- Select the option that best balances effectiveness, safety, and professionalism
Question 1
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 2
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 3
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 4
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 5
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 6
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 7
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 8
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 9
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 10
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 11
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 12
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 13
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 14
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Question 15
In the context of AI red team assessments, which approach provides the most comprehensive evaluation of a target system's security posture?
Scoring and Recommendations
| Score | Level | Next Steps |
|---|---|---|
| 90-100% | Expert | Ready to proceed; consider mentoring others |
| 80-89% | Proficient | Minor gaps; targeted review of missed topics |
| 70-79% | Competent | Adequate; review incorrect areas before advancing |
| 60-69% | Developing | Significant gaps; revisit relevant curriculum sections |
| Below 60% | Foundational | Complete prerequisites before retaking |
Review Guidance
After completing this assessment, review the explanations for all questions — including those you answered correctly. Understanding why an answer is correct is as important as knowing the answer itself, and the explanations often contain additional context that deepens understanding of the topic.
For questions you answered incorrectly:
- Identify the topic area the question covers
- Review the corresponding curriculum section
- Complete any related lab exercises
- Re-attempt the question after review
Advanced Considerations
Evolving Attack Landscape
The AI security landscape evolves rapidly as both offensive techniques and defensive measures advance. Several trends shape the current state of play:
Increasing model capabilities create new attack surfaces. As models gain access to tools, code execution, web browsing, and computer use, each new capability introduces potential exploitation vectors that did not exist in earlier, text-only systems. The principle of least privilege becomes increasingly important as model capabilities expand.
Safety training improvements are necessary but not sufficient. Model providers invest heavily in safety training through RLHF, DPO, constitutional AI, and other alignment techniques. These improvements raise the bar for successful attacks but do not eliminate the fundamental vulnerability: models cannot reliably distinguish legitimate instructions from adversarial ones because this distinction is not represented in the architecture.
Automated red teaming tools democratize testing. Tools like NVIDIA's Garak, Microsoft's PyRIT, and Promptfoo enable organizations to conduct automated security testing without deep AI security expertise. However, automated tools catch known patterns; novel attacks and business logic vulnerabilities still require human creativity and domain knowledge.
Regulatory pressure drives organizational investment. The EU AI Act, NIST AI RMF, and industry-specific regulations increasingly require organizations to assess and mitigate AI-specific risks. This regulatory pressure is driving investment in AI security programs, but many organizations are still in the early stages of building mature AI security practices.
Cross-Cutting Security Principles
Several security principles apply across all topics covered in this curriculum:
-
Defense-in-depth: No single defensive measure is sufficient. Layer multiple independent defenses so that failure of any single layer does not result in system compromise. Input classification, output filtering, behavioral monitoring, and architectural controls should all be present.
-
Assume breach: Design systems assuming that any individual component can be compromised. This mindset leads to better isolation, monitoring, and incident response capabilities. When a prompt injection succeeds, the blast radius should be minimized through architectural controls.
-
Least privilege: Grant models and agents only the minimum capabilities needed for their intended function. A customer service chatbot does not need file system access or code execution. Excessive capabilities magnify the impact of successful exploitation.
-
Continuous testing: AI security is not a one-time assessment. Models change, defenses evolve, and new attack techniques are discovered regularly. Implement continuous security testing as part of the development and deployment lifecycle.
-
Secure by default: Default configurations should be secure. Require explicit opt-in for risky capabilities, use allowlists rather than denylists, and err on the side of restriction rather than permissiveness.
Integration with Organizational Security
AI security does not exist in isolation — it must integrate with the organization's broader security program:
| Security Domain | AI-Specific Integration |
|---|---|
| Identity and Access | API key management, model access controls, user authentication for AI features |
| Data Protection | Training data classification, PII in prompts, data residency for model calls |
| Application Security | AI feature threat modeling, prompt injection in SAST/DAST, secure AI design patterns |
| Incident Response | AI-specific playbooks, model behavior monitoring, prompt injection forensics |
| Compliance | AI regulatory mapping (EU AI Act, NIST), AI audit trails, model documentation |
| Supply Chain | Model provenance, dependency security, adapter/weight integrity verification |
class OrganizationalIntegration:
"""Framework for integrating AI security with organizational security programs."""
def __init__(self, org_config: dict):
self.config = org_config
self.gaps = []
def assess_maturity(self) -> dict:
"""Assess the organization's AI security maturity."""
domains = {
"governance": self._check_governance(),
"technical_controls": self._check_technical(),
"monitoring": self._check_monitoring(),
"incident_response": self._check_ir(),
"training": self._check_training(),
}
overall = sum(d["score"] for d in domains.values()) / len(domains)
return {"domains": domains, "overall_maturity": round(overall, 1)}
def _check_governance(self) -> dict:
has_policy = self.config.get("ai_security_policy", False)
has_framework = self.config.get("risk_framework", False)
score = (int(has_policy) + int(has_framework)) * 2.5
return {"score": score, "max": 5.0}
def _check_technical(self) -> dict:
controls = ["input_classification", "output_filtering", "rate_limiting", "sandboxing"]
active = sum(1 for c in controls if self.config.get(c, False))
return {"score": active * 1.25, "max": 5.0}
def _check_monitoring(self) -> dict:
has_monitoring = self.config.get("ai_monitoring", False)
has_alerting = self.config.get("ai_alerting", False)
score = (int(has_monitoring) + int(has_alerting)) * 2.5
return {"score": score, "max": 5.0}
def _check_ir(self) -> dict:
has_playbook = self.config.get("ai_ir_playbook", False)
return {"score": 5.0 if has_playbook else 0.0, "max": 5.0}
def _check_training(self) -> dict:
has_training = self.config.get("ai_security_training", False)
return {"score": 5.0 if has_training else 0.0, "max": 5.0}Future Directions
Several research and industry trends will shape the evolution of this field:
- Formal methods for AI safety: Development of mathematical frameworks that can provide bounded guarantees about model behavior under adversarial conditions
- Automated red teaming at scale: Continued improvement of automated testing tools that can discover novel vulnerabilities without human guidance
- AI-assisted defense: Using AI systems to detect and respond to attacks on other AI systems, creating a dynamic attack-defense ecosystem
- Standardized evaluation: Growing adoption of standardized benchmarks (HarmBench, JailbreakBench) that enable consistent measurement of progress
- Regulatory harmonization: Convergence of AI regulatory frameworks across jurisdictions, providing clearer requirements for organizations
References and Further Reading
- OWASP LLM Top 10 2025 — Comprehensive guide to LLM security risks (owasp.org/www-project-top-10-for-large-language-model-applications)
- MITRE ATLAS — Adversarial Threat Landscape for AI Systems (atlas.mitre.org)
- HarmBench — Standardized evaluation framework (github.com/centerforaisafety/HarmBench)
- JailbreakBench — Jailbreak evaluation benchmark (github.com/JailbreakBench/jailbreakbench)
- NIST AI 600-1 — Generative AI Profile for risk management
What is the most effective defensive strategy against the attack class described in this article?
Why do the techniques described in this article continue to be effective despite ongoing security improvements by model providers?