There are 17 repositories under prompt-injection topic.
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
Advanced Code and Text Manipulation Prompts for Various LLMs. Suitable for Siri, GPT-4o, Claude, Llama3, Gemini, and other high-performance open-source LLMs.
💼 another CV template for your job application, yet powered by Typst and more
# Prompt Engineering Hub ⭐️ If you find this helpful, give it a star to show your support! This repository is a one-stop resource for prompt engineering. Also available on: https://promptengineeringhub.dev/
Every practical and proposed defense against prompt injection.
Self-hardening firewall for large language models
Prompts of GPT-4V & DALL-E3 to full utilize the multi-modal ability. GPT4V Prompts, DALL-E3 Prompts.
This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
prompt attack-defense, prompt Injection, reverse engineering notes and examples | 提示词对抗、破解例子与笔记
gpt_server是一个用于生产级部署LLMs或Embedding的开源框架。
A benchmark for prompt injection detection systems.
Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks
A prompt injection game to collect data for robust ML research
My inputs for the LLM Gandalf made by Lakera
Build production ready apps for GPT using Node.js & TypeScript
This project investigates the security of large language models by performing binary classification of a set of input prompts to discover malicious prompts. Several approaches have been analyzed using classical ML algorithms, a trained LLM model, and a fine-tuned LLM model.
This is The most comprehensive prompt hacking course available, which record our progress on a prompt engineering and prompt hacking course.
jailbreakme.xyz is an open-source decentralized app (dApp) where users are challenged to try and jailbreak pre-existing LLMs in order to find weaknesses and be rewarded. 🏆
Whispers in the Machine: Confidentiality in LLM-integrated Systems
Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platform provider.
LLM | Security | Operations in one github repo with good links and pictures.
Short list of indirect prompt injection attacks for OpenAI-based models.
My solutions for Lakera's Gandalf
The Security Toolkit for managing Generative AI(especially LLMs) and Supervised Learning processes(Learning and Inference).
Guard your LangChain applications against prompt injection with Lakera ChainGuard.
🤯 AI Security EXPOSED! Live Demos Showing Hidden Risks of 🤖 Agentic AI Flows: 💉Prompt Injection, ☣️ Data Poisoning. Watch the recorded session:
Manual Prompt Injection / Red Teaming Tool
MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.