Linking pages
Linked pages
- Jailbreak Chat https://www.jailbreakchat.com 528 comments
- Gandalf | Lakera - Prompt Injection https://gandalf.lakera.ai/ 359 comments
- Extracting Training Data from ChatGPT https://not-just-memorization.github.io/extracting-training-data-from-chatgpt.html 134 comments
- Kevin Liu on Twitter: "The entire prompt of Microsoft Bing Chat?! (Hi, Sydney.) https://t.co/ZNywWV9MNB" / Twitter https://twitter.com/kliu128/status/1623472922374574080 103 comments
- https://twitter.com/goodside/status/1746685366952735034 39 comments
- [2302.12173] Not what you've signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection https://arxiv.org/abs/2302.12173 26 comments
- Riley Goodside on Twitter: "Exploiting GPT-3 prompts with malicious inputs that order the model to ignore its previous directions. https://t.co/I0NVr9LOJq" / Twitter https://twitter.com/goodside/status/1569128808308957185 15 comments
- Universal and Transferable Attacks on Aligned Language Models https://llm-attacks.org/ 4 comments
- How Johnny Can Persuade LLMs to Jailbreak Them:Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs https://chats-lab.github.io/persuasive_jailbreaker/ 4 comments
- GitHub - NVIDIA/NeMo-Guardrails: NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems. https://github.com/NVIDIA/NeMo-Guardrails 1 comment
- GitHub - laiyer-ai/llm-guard: The Security Toolkit for LLM Interactions https://github.com/laiyer-ai/llm-guard 1 comment
- 12 techniques to reduce your LLM API bill and launch blazingly fast products https://www.aitidbits.ai/p/reduce-llm-latency-and-cost 1 comment
- https://twitter.com/ashbeauchamp/status/1748034519104450874 1 comment
- GitHub - whylabs/langkit: 🔍 LangKit: An open-source toolkit for monitoring Language Learning Models (LLMs). 📚 Extracts signals from prompts & responses, ensuring safety & security. 🛡️ Features include text quality, relevance metrics, & sentiment analysis. 📊 A comprehensive tool for LLM observability. 👀 https://github.com/whylabs/langkit 0 comments
- GitHub - protectai/rebuff: Rebuff.ai - Prompt Injection Detector https://github.com/protectai/rebuff 0 comments
- The future of Internet Search in the era of LLMs https://www.aitidbits.ai/p/future-of-internet-search 0 comments
- Harnessing research-backed prompting techniques for enhanced LLM performance https://www.aitidbits.ai/p/advanced-prompting 0 comments
- Public Domain Jailbreak : ChatGPT https://old.reddit.com/r/ChatGPT/comments/18wf1ie/public_domain_jailbreak/ 0 comments
- LVE Repository https://lve-project.org/index.html 0 comments
- [2311.16119] Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of LLMs through a Global Scale Prompt Hacking Competition https://arxiv.org/abs/2311.16119 0 comments
Related searches:
Search whole site: site:artificialintelligencemadesimple.substack.com
Search title: 7 methods to secure LLM apps from prompt injections and jailbreaks [Guest]
See how to search.