Easy to use LLM Prompt Injection Detection / Detector Python Package.
Application which investigates defensive measures against prompt injection attacks on LLMs, focusing on external tool exposure.
Short list of indirect prompt injection attacks for OpenAI-based models.
Fine-tuning base models to create robust task-specific models for better performance.
This repository contains the official code for the paper on prompt injection and parameterization.
Official implementation of StruQ, which defends against prompt injection attacks using structured queries.
The official implementation of InjecGuard, a tool for benchmarking and mitigating over-defense in prompt injection guardrail models.
A writeup for the Gandalf prompt injection game.
This project investigates the security of large language models by classifying prompts to discover malicious injections.
Repo for the research paper "SecAlign: Defending Against Prompt Injection with Preference Optimization"
Uses the ChatGPT model to filter out potentially dangerous user-supplied questions.
Project Mantis is a tool designed to counter LLM-driven cyberattacks using prompt injection techniques.