代理 安全 Practice Exam
Practice exam focused on agentic AI security including MCP, A2A, function calling, and multi-agent threats.
評估 概覽
Practice exam focused on 代理式 AI 安全 including MCP, A2A, 函式呼叫, and multi-代理 threats.
評估 Details:
- Questions: 15
- Recommended Time: 45 minutes
- Passing Score: 70%
- Difficulty Level: Advanced
Instructions
Read each question carefully and 考慮 all answer options before selecting your response. Many questions have multiple plausible answers — select the best answer that accounts for the full context of the scenario.
For scenario-based questions:
- 識別 the key constraints and objectives in the scenario
- 考慮 what information you have and what you lack
- 評估 each option against the scenario's specific requirements
- Select the option that best balances effectiveness, 安全, and professionalism
Question 1
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 2
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 3
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 4
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 5
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 6
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 7
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 8
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 9
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 10
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 11
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 12
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 13
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 14
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Question 15
In the context of AI 紅隊 assessments, which approach provides the most comprehensive 評估 of a target system's 安全 posture?
Scoring and Recommendations
| Score | Level | Next Steps |
|---|---|---|
| 90-100% | Expert | Ready to proceed; 考慮 mentoring others |
| 80-89% | Proficient | Minor gaps; targeted review of missed topics |
| 70-79% | Competent | Adequate; review incorrect areas before advancing |
| 60-69% | Developing | Significant gaps; revisit relevant curriculum sections |
| Below 60% | Foundational | Complete prerequisites before retaking |
Review Guidance
After completing this 評估, review the explanations for all questions — including those you answered correctly. 理解 why an answer is correct is as important as knowing the answer itself, and the explanations often contain additional context that deepens 理解 of the topic.
For questions you answered incorrectly:
- 識別 the topic area the question covers
- Review the corresponding curriculum section
- Complete any related lab exercises
- Re-attempt the question after review
Advanced Considerations
Evolving 攻擊 Landscape
The AI 安全 landscape evolves rapidly as both offensive techniques and defensive measures advance. Several trends shape the current state of play:
Increasing model capabilities create new attack surfaces. As models gain access to tools, code execution, web browsing, and computer use, each new capability introduces potential 利用 vectors that did not exist in earlier, text-only systems. The principle of least privilege becomes increasingly important as model capabilities expand.
安全 訓練 improvements are necessary but not sufficient. Model providers invest heavily in 安全 訓練 through RLHF, DPO, constitutional AI, and other 對齊 techniques. These improvements raise the bar for successful attacks but do not eliminate the fundamental 漏洞: models cannot reliably distinguish legitimate instructions from 對抗性 ones 因為 this distinction is not represented in the architecture.
Automated 紅隊演練 tools democratize 測試. Tools like NVIDIA's Garak, Microsoft's PyRIT, and Promptfoo enable organizations to conduct automated 安全 測試 without deep AI 安全 expertise. 然而, automated tools catch known patterns; novel attacks and business logic 漏洞 still require human creativity and domain knowledge.
Regulatory pressure drives organizational investment. The EU AI Act, NIST AI RMF, and industry-specific regulations increasingly require organizations to 評估 and mitigate AI-specific risks. This regulatory pressure is driving investment in AI 安全 programs, but many organizations are still in the early stages of building mature AI 安全 practices.
Cross-Cutting 安全 Principles
Several 安全 principles apply across all topics covered 在本 curriculum:
-
防禦-in-depth: No single defensive measure is sufficient. Layer multiple independent 防禦 so that failure of any single layer does not result in system compromise. 輸入 classification, 輸出 filtering, behavioral 監控, and architectural controls should all be present.
-
Assume breach: Design systems assuming that any individual component can be compromised. This mindset leads to better isolation, 監控, and incident response capabilities. When a 提示詞注入 succeeds, the blast radius should be minimized through architectural controls.
-
Least privilege: Grant models and 代理 only the minimum capabilities needed for their intended function. A customer service chatbot does not need file system access or code execution. Excessive capabilities magnify the impact of successful 利用.
-
Continuous 測試: AI 安全 is not a one-time 評估. Models change, 防禦 evolve, and new attack techniques are discovered regularly. 實作 continuous 安全 測試 as part of the development and deployment lifecycle.
-
Secure by default: Default configurations should be secure. Require explicit opt-in for risky capabilities, use allowlists rather than denylists, and err on the side of restriction rather than permissiveness.
Integration with Organizational 安全
AI 安全 does not exist in isolation — it must integrate with the organization's broader 安全 program:
| 安全 Domain | AI-Specific Integration |
|---|---|
| Identity and Access | API key management, model access controls, user 認證 for AI features |
| Data Protection | 訓練資料 classification, PII in prompts, data residency for model calls |
| Application 安全 | AI feature threat modeling, 提示詞注入 in SAST/DAST, secure AI design patterns |
| Incident Response | AI-specific playbooks, model behavior 監控, 提示詞注入 forensics |
| Compliance | AI regulatory mapping (EU AI Act, NIST), AI audit trails, model documentation |
| Supply Chain | Model provenance, dependency 安全, adapter/weight integrity verification |
class OrganizationalIntegration:
"""Framework for integrating AI 安全 with organizational 安全 programs."""
def __init__(self, org_config: dict):
self.config = org_config
self.gaps = []
def assess_maturity(self) -> dict:
"""評估 the organization's AI 安全 maturity."""
domains = {
"governance": self._check_governance(),
"technical_controls": self._check_technical(),
"監控": self._check_monitoring(),
"incident_response": self._check_ir(),
"訓練": self._check_training(),
}
overall = sum(d["score"] for d in domains.values()) / len(domains)
return {"domains": domains, "overall_maturity": round(overall, 1)}
def _check_governance(self) -> dict:
has_policy = self.config.get("ai_security_policy", False)
has_framework = self.config.get("risk_framework", False)
score = (int(has_policy) + int(has_framework)) * 2.5
return {"score": score, "max": 5.0}
def _check_technical(self) -> dict:
controls = ["input_classification", "output_filtering", "rate_limiting", "sandboxing"]
active = sum(1 for c in controls if self.config.get(c, False))
return {"score": active * 1.25, "max": 5.0}
def _check_monitoring(self) -> dict:
has_monitoring = self.config.get("ai_monitoring", False)
has_alerting = self.config.get("ai_alerting", False)
score = (int(has_monitoring) + int(has_alerting)) * 2.5
return {"score": score, "max": 5.0}
def _check_ir(self) -> dict:
has_playbook = self.config.get("ai_ir_playbook", False)
return {"score": 5.0 if has_playbook else 0.0, "max": 5.0}
def _check_training(self) -> dict:
has_training = self.config.get("ai_security_training", False)
return {"score": 5.0 if has_training else 0.0, "max": 5.0}Future Directions
Several research and industry trends will shape the evolution of this field:
- Formal methods for AI 安全: Development of mathematical frameworks that can provide bounded guarantees about model behavior under 對抗性 conditions
- Automated 紅隊演練 at scale: Continued improvement of automated 測試 tools that can discover novel 漏洞 without human guidance
- AI-assisted 防禦: Using AI systems to detect and respond to attacks on other AI systems, creating a dynamic attack-防禦 ecosystem
- Standardized 評估: Growing adoption of standardized benchmarks (HarmBench, JailbreakBench) that enable consistent measurement of progress
- Regulatory harmonization: Convergence of AI regulatory frameworks across jurisdictions, providing clearer requirements for organizations
參考文獻 and Further Reading
- OWASP LLM Top 10 2025 — Comprehensive guide to LLM 安全 risks (owasp.org/www-project-top-10-for-large-language-model-applications)
- MITRE ATLAS — 對抗性 Threat Landscape for AI Systems (atlas.mitre.org)
- HarmBench — Standardized 評估 framework (github.com/centerforaisafety/HarmBench)
- JailbreakBench — 越獄 評估 benchmark (github.com/JailbreakBench/jailbreakbench)
- NIST AI 600-1 — Generative AI Profile for risk management
What is the most effective defensive strategy against the attack class described 在本 article?
Why do the techniques described 在本 article continue to be effective despite ongoing 安全 improvements by model providers?