# reasoning
27 articlestagged with “reasoning”
Manipulating Reasoning Chains
Techniques for influencing an AI agent's chain-of-thought reasoning to steer its planning, decision-making, and tool selection toward attacker-desired outcomes.
Reasoning Model Security Assessment
Assessment of chain-of-thought exploitation, reasoning trace manipulation, and thinking-token attacks.
Skill Verification: Reasoning Model Attacks
Verification of skills in reasoning trace manipulation, chain-of-thought exploitation, and thinking-token attacks.
Chain-of-Thought Exploitation Techniques
Deep analysis of how reasoning traces in CoT models can be manipulated to produce adversarial outputs while maintaining coherent reasoning.
Frontier Research
Cutting-edge AI security research covering reasoning model attacks, code generation security, computer use agents, AI-powered red teaming, robotics and embodied AI, and alignment faking.
Multi-Modal Reasoning Attacks
Attacking reasoning processes that span multiple modalities in vision-language and audio-language models.
Multimodal Reasoning Security
Security challenges specific to models that reason across text, image, audio, and video modalities simultaneously.
Chain-of-Thought Exploitation
Techniques for manipulating reasoning chains in CoT-enabled models: false premise injection, logic bombs, reasoning hijacking, and chain corruption attacks.
Reasoning Model Attacks
Overview of security risks in reasoning-enabled LLMs: how chain-of-thought models introduce new attack surfaces, exploit primitives, and defensive challenges.
Reasoning Model Jailbreaks
How reasoning capabilities create novel jailbreak surfaces: chain-of-thought exploitation, scratchpad attacks, and why higher reasoning effort increases attack success.
Steganographic Reasoning
Hidden communication channels within AI reasoning traces, where models encode information or coordinate behavior through patterns invisible to human overseers, including detection methods and implications for AI safety.
Thought Injection & Hidden CoT Manipulation
Techniques for injecting thoughts into hidden reasoning traces, exploiting summarized vs full CoT, and steering model reasoning without visible manipulation.
Unfaithful Chain-of-Thought Reasoning
Analysis of unfaithful chain-of-thought reasoning in language models, where the visible reasoning trace does not accurately reflect the model's actual computational process, including detection methods, implications for oversight, and exploitation techniques.
Reasoning Model Exploitation
Exploiting extended thinking and chain-of-thought reasoning in o1, Claude, and DeepSeek-R1 models.
Thinking Token Attacks on Reasoning Models
Exploiting extended thinking tokens and internal reasoning processes in models like Claude and o1 for jailbreaking.
Test-Time Compute Scaling Attacks (Frontier Research)
Attacking test-time compute scaling in reasoning models to exploit resource allocation and reasoning traces.
Injection in Reasoning Models
Research into injection attacks specific to reasoning-augmented models that exploit chain-of-thought processes and self-reflection mechanisms.
Lab: Reasoning Model Exploitation
Attack reasoning models like o1, o3, and DeepSeek-R1 by exploiting chain-of-thought manipulation, reasoning budget exhaustion, and thought-injection techniques.
Reasoning Chain Hijack: CoT Exploitation CTF
Hijack a reasoning model's chain-of-thought to redirect its conclusions while maintaining coherent reasoning.
Reasoning Model Chain-of-Thought Exploitation
Exploit extended thinking and chain-of-thought reasoning in models like o1 and DeepSeek-R1.
Reasoning Chain Hijacking
Hijack the model's chain-of-thought reasoning to redirect its conclusions toward attacker-controlled objectives.
DeepSeek-R1 Security Analysis
Security analysis of DeepSeek-R1's reasoning capabilities and MoE architecture vulnerabilities.
Reasoning Model Security Analysis
Security analysis of reasoning-augmented models (o1, DeepSeek-R1) focusing on chain-of-thought manipulation and reasoning-specific attack vectors.
Reasoning Model Exploitation Walkthrough
Exploit extended thinking in reasoning models to inject false premises and manipulate conclusion generation.
Thought Injection in Reasoning Models Walkthrough
Inject adversarial thoughts into the reasoning chain of thinking models to manipulate final outputs.
Chain-of-Thought Hijacking Walkthrough
Walkthrough of hijacking visible reasoning traces in CoT models to redirect conclusions and bypass safety checks.
Testing DeepSeek Models
Red team testing guide for DeepSeek models including R1 reasoning exploitation and MoE architecture.