Hacker News
- How Johnny can persuade LLMs to jailbreak them https://chats-lab.github.io/persuasive_jailbreaker/ 4 comments
Linking pages
- DALL-E’s New Guardrails: Fast, Furious, and Far from Airtight https://garymarcus.substack.com/p/dall-es-new-guardrails-fast-furious 0 comments
- 7 methods to secure LLM apps from prompt injections and jailbreaks [Guest] https://artificialintelligencemadesimple.substack.com/p/mitigate-prompt-attacks 0 comments
- 7 methods to secure LLM apps from prompt injections and jailbreaks [Guest] http://mitigate-prompt-attacks.aitidbits.ai 0 comments
Related searches:
Search whole site: site:chats-lab.github.io
Search title: How Johnny Can Persuade LLMs to Jailbreak Them:Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs
See how to search.