Defending ChatGPT against jailbreak attack via self-reminders
Por um escritor misterioso
Descrição
Explainer: What does it mean to jailbreak ChatGPT
The ELI5 Guide to Prompt Injection: Techniques, Prevention Methods
Will AI ever be jailbreak proof? : r/ChatGPT
ChatGPT Jailbreak Prompt: Unlock its Full Potential
Attack Success Rate (ASR) of 54 Jailbreak prompts for ChatGPT with
Defending ChatGPT against jailbreak attack via self-reminders
OWASP Top 10 for Large Language Model Applications
How to Jailbreak ChatGPT?
Pause For Thought: The AI Pause Debate - by Scott Alexander
Cyber-criminals “Jailbreak” AI Chatbots For Malicious Ends
ChatGPT Jailbreak Prompts: Top 5 Points for Masterful Unlocking
GitHub - yjw1029/Self-Reminder: Code for our paper Defending