The official implementation of InjecGuard, a tool for benchmarking and mitigating over-defense in prompt injection guardrail models.
Repo for the research paper "SecAlign: Defending Against Prompt Injection with Preference Optimization"
The official implementation of a pre-print paper on prompt injection attacks against large language models.
A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.
This repository provides a benchmark for prompt Injection attacks and defenses.
LLM Prompt Injection Detector designed to protect AI applications from prompt injection attacks.
A resource for understanding adversarial prompting in LLMs and techniques to mitigate risks.