# model
標記為「model」的 38 篇文章
GCP Model Garden Security
Security assessment of GCP Model Garden including model deployment, versioning, and access control.
Model Collapse in Recursive Training
Security implications of model collapse when models are trained on AI-generated data from previous generations.
Model Artifact Integrity Verification
Implementing integrity verification for model artifacts through checksums, signatures, and provenance tracking.
Model Distillation Security Lab
Extract model capabilities through distillation techniques using only black-box API access.
Reward Model Analysis Lab
Analyze and exploit reward model biases to craft responses that score high on safety while embedding harmful content.
Model Security Comparison Lab
Compare the security posture of different LLM models by running identical test suites across providers.
Model Whisperer Challenge
Extract the exact system prompt from 8 progressively harder LLM applications using creative extraction techniques.
Practical Model Fingerprinting Lab
Fingerprint the underlying model behind an API endpoint to identify its type, version, and configuration.
Multi-Model Pipeline Attack Lab
Attack a pipeline where multiple models process data sequentially, exploiting trust between pipeline stages.
Model Gateway Security Patterns
Security patterns for centralized model gateway deployments including authentication, authorization, and auditing.
Model Rollback Security
Security implications of model rollback procedures including exposure windows and state consistency.
Open Source Model Safety Comparison
Comparative safety analysis across open-source model families including Llama, Mistral, Qwen, and Phi.
Phi Models Security Analysis
Security analysis of Microsoft's Phi small language model family including safety vs capability tradeoffs.
Quantization Impact on Model Safety
How quantization affects safety alignment including GPTQ, AWQ, and GGUF format implications.
Qwen Models Security Analysis
Security analysis of Alibaba's Qwen model family including multilingual safety considerations.
Model Merging Security Analysis (Training Pipeline)
Security analysis of model merging techniques and propagation of vulnerabilities through merged models.
Model Supply Chain Attacks
Comprehensive analysis of model supply chain attack vectors from training data through deployment.
Vision Model Attack Walkthrough (Attack Walkthrough)
Step-by-step walkthrough of visual prompt injection, adversarial images, and OCR exploitation in vision-language models.
Model Behavior Monitoring Setup
Set up comprehensive model behavior monitoring to detect drift, anomalies, and potential compromise.
GCP 模型 Garden 安全
安全 assessment of GCP 模型 Garden including model deployment, versioning, and access control.
模型 Collapse in Recursive 訓練
安全 implications of model collapse when models are trained on AI-generated data from previous generations.
模型 Artifact Integrity Verification
Implementing integrity verification for model artifacts through checksums, signatures, and provenance tracking.
模型 Distillation 安全 實驗室
Extract model capabilities through distillation techniques using only black-box API access.
Reward 模型 Analysis 實驗室
Analyze and exploit reward model biases to craft responses that score high on safety while embedding harmful content.
模型 安全 Comparison 實驗室
Compare the security posture of different LLM models by running identical test suites across providers.
模型 Whisperer Challenge
Extract the exact system prompt from 8 progressively harder LLM applications using creative extraction techniques.
Practical 模型 Fingerprinting 實驗室
Fingerprint the underlying model behind an API endpoint to identify its type, version, and configuration.
Multi-模型 Pipeline 攻擊 實驗室
攻擊 a pipeline where multiple models process data sequentially, exploiting trust between pipeline stages.
模型 Gateway 安全 Patterns
安全 patterns for centralized model gateway deployments including authentication, authorization, and auditing.
模型 Rollback 安全
安全 implications of model rollback procedures including exposure windows and state consistency.
Open Source 模型 Safety Comparison
Comparative safety analysis across open-source model families including Llama, Mistral, Qwen, and Phi.
Phi 模型s 安全 Analysis
Security analysis of Microsoft's Phi small language model family including safety vs capability tradeoffs.
Quantization Impact on 模型 Safety
How quantization affects safety alignment including GPTQ, AWQ, and GGUF format implications.
Qwen 模型s 安全 Analysis
Security analysis of Alibaba's Qwen model family including multilingual safety considerations.
模型 Merging 安全 Analysis (訓練 Pipeline)
安全 analysis of model merging techniques and propagation of vulnerabilities through merged models.
模型 Supply Chain 攻擊s
Comprehensive analysis of model supply chain attack vectors from training data through deployment.
Vision 模型 攻擊 導覽 (攻擊 導覽)
Step-by-step walkthrough of visual prompt injection, adversarial images, and OCR exploitation in vision-language models.
模型 Behavior Monitoring Setup
Set up comprehensive model behavior monitoring to detect drift, anomalies, and potential compromise.