There are 16 repositories under prompt-injection topic.
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
Advanced Code and Text Manipulation Prompts for Various LLMs. Suitable for Siri, GPT-4o, Claude, Llama3, Gemini, and other high-performance open-source LLMs.
💼 another CV template for your job application, yet powered by Typst and more
Every practical and proposed defense against prompt injection.
# Prompt Engineering Hub ⭐️ If you find this helpful, give it a star to show your support! This repository is a one-stop resource for prompt engineering. Also available on: https://promptengineeringhub.dev/
Self-hardening firewall for large language models
Prompts of GPT-4V & DALL-E3 to full utilize the multi-modal ability. GPT4V Prompts, DALL-E3 Prompts.
This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
prompt attack-defense, prompt Injection, reverse engineering notes and examples | 提示词对抗、破解例子与笔记
A benchmark for prompt injection detection systems.
Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks
A prompt injection game to collect data for robust ML research
My inputs for the LLM Gandalf made by Lakera
Build production ready apps for GPT using Node.js & TypeScript
This project investigates the security of large language models by performing binary classification of a set of input prompts to discover malicious prompts. Several approaches have been analyzed using classical ML algorithms, a trained LLM model, and a fine-tuned LLM model.
Whispers in the Machine: Confidentiality in LLM-integrated Systems
Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platform provider.
This is The most comprehensive prompt hacking course available, which record our progress on a prompt engineering and prompt hacking course.
jailbreakme.xyz is an open-source decentralized app (dApp) where users are challenged to try and jailbreak pre-existing LLMs in order to find weaknesses and be rewarded. 🏆
Short list of indirect prompt injection attacks for OpenAI-based models.
The Security Toolkit for managing Generative AI(especially LLMs) and Supervised Learning processes(Learning and Inference).
LLM | Security | Operations in one github repo with good links and pictures.
My solutions for Lakera's Gandalf
Guard your LangChain applications against prompt injection with Lakera ChainGuard.
🤯 AI Security EXPOSED! Live Demos Showing Hidden Risks of 🤖 Agentic AI Flows: 💉Prompt Injection, ☣️ Data Poisoning. Watch the recorded session:
MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.
Manual Prompt Injection / Red Teaming Tool
A prompt defence is a multi-layer defence that can be used to protect your applications against prompt injection attacks.