# distillation
13 articlestagged with “distillation”
Knowledge Distillation Attacks
Attacking knowledge distillation pipelines to transfer backdoors from teacher to student models or extract proprietary capabilities.
API-Based Model Extraction
Deep dive into extracting proprietary model capabilities through systematic API querying, active learning strategies, logprob exploitation, soft-label distillation, and evasion of query anomaly detection systems.
Model Extraction & IP Theft
Methodology for black-box model extraction, API-based distillation, side-channel extraction, watermark removal, and model fingerprinting bypass targeting deployed AI systems.
Model Distillation Security Implications
Security risks of knowledge distillation including capability transfer, safety property loss, and unauthorized model cloning.
Safety Loss During Model Distillation
Research on how safety alignment degrades during knowledge distillation from larger to smaller models.
Model Compression Security
Security implications of model pruning, quantization, and knowledge distillation on AI system robustness.
Model Distillation Security Lab
Extract model capabilities through distillation techniques using only black-box API access.
Inference Optimization Risks
Security implications of model optimization techniques — covering quantization safety degradation, pruning vulnerability introduction, distillation attacks, and speculative decoding risks.
Distillation Security Analysis
Security implications of knowledge distillation including backdoor transfer, capability extraction, and safety property degradation in student models.
Model Distillation Attacks
Stealing model capabilities via knowledge distillation: API-based distillation, bypassing access restrictions, task-specific capability theft, and defense against distillation-based model stealing.
Distillation-Based Model Extraction
Using knowledge distillation for model theft: student-teacher extraction attacks, API-based distillation, task-specific extraction, and defending against distillation-based model stealing.
Knowledge Distillation Safety Gap
Analysis of safety property loss during knowledge distillation from teacher to student models.
Knowledge Distillation Security
Security implications of knowledge distillation including capability extraction and safety alignment transfer.