# persona
標記為「persona」的 20 篇文章
Jailbreaking via Persona Engineering
Research on using sophisticated persona engineering to bypass safety training in frontier models.
Lab: Context Manipulation
Manipulate the conversational context to change model behavior, using techniques like context flooding, false history injection, and persona shifting.
Lab: Role-Play Attacks
Use persona-based approaches to bypass AI safety measures by assigning alternate identities, characters, and scenarios that override the model's trained refusal behaviors.
Lab: Role-Play & Persona Attacks
Hands-on lab exploring role-play and persona-based jailbreak techniques, where you craft scenarios that trick LLMs into adopting alternative identities to bypass safety guardrails.
Chatbot Persona and Capability Mapping
Map the configured persona, capabilities, and restrictions of an LLM chatbot through systematic probing.
Persona-Based Injection Attacks
Advanced persona and role-play based attacks that exploit instruction-following behavior.
Persona Establishment
Creating persistent alternate identities that survive across conversation turns, including character locking, identity anchoring, and progressive persona building.
Role-Play Attacks
Establishing alternate personas or fictional scenarios that cause models to bypass safety training, including DAN variants, character hijacking, and narrative framing.
Multi-Turn Persona Escalation Walkthrough
Gradually escalate through persona manipulation across multiple turns to achieve full safety bypass.
Virtual Persona Creation
Creating persistent alternate personas within LLM conversations to bypass safety training, establishing character identities that override the model's default behavioral constraints.
越獄ing via Persona Engineering
Research on using sophisticated persona engineering to bypass safety training in frontier models.
實驗室: Context Manipulation
Manipulate the conversational context to change model behavior, using techniques like context flooding, false history injection, and persona shifting.
實驗室: Role-Play 攻擊s
Use persona-based approaches to bypass AI safety measures by assigning alternate identities, characters, and scenarios that override the model's trained refusal behaviors.
實驗室: Role-Play & Persona 攻擊s
Hands-on lab exploring role-play and persona-based jailbreak techniques, where you craft scenarios that trick LLMs into adopting alternative identities to bypass safety guardrails.
Chatbot Persona and Capability Mapping
Map the configured persona, capabilities, and restrictions of an LLM chatbot through systematic probing.
Persona-Based Injection 攻擊s
進階 persona and role-play based attacks that exploit instruction-following behavior.
人格建立
建立跨對話輪次存活並抵擋回復預設行為之持久另一身份,包含角色鎖定、身份錨定與漸進式人格建構。
Role-Play 攻擊s
Establishing alternate personas or fictional scenarios that cause models to bypass safety training, including DAN variants, character hijacking, and narrative framing.
Multi-Turn Persona Escalation 導覽
Gradually escalate through persona manipulation across multiple turns to achieve full safety bypass.
Virtual Persona Creation
Creating persistent alternate personas within LLM conversations to bypass safety training, establishing character identities that override the model's default behavioral constraints.