Category
Explore by categories

AI Research PapersModel Robustness EnhancementPrompt Injection Defense
InjecGuard
Details
The official implementation of InjecGuard, a tool for benchmarking and mitigating over-defense in prompt injection guardrail models.

AI Research PapersModel Robustness EnhancementPrompt Injection Defense
SecAlign
Details
Repo for the research paper "SecAlign: Defending Against Prompt Injection with Preference Optimization"

AI ModelsAI Security MonitoringModel Robustness Enhancement
Universal-Prompt-Injection
Details
The official implementation of a pre-print paper on prompt injection attacks against large language models.

AI ModelsSecurity ResearchModel Robustness Enhancement
BIPIA
Details
A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.

Input Validation & FilteringModel Robustness EnhancementPrompt Injection Defense
Open-Prompt-Injection
Details
This repository provides a benchmark for prompt Injection attacks and defenses.

AI Security MonitoringModel Robustness EnhancementPrompt Injection Defense
Rebuff
Details
LLM Prompt Injection Detector designed to protect AI applications from prompt injection attacks.

AI Ethics ResourcesModel Robustness EnhancementPrompt Injection Defense
Prompt Engineering Guide
Details
A resource for understanding adversarial prompting in LLMs and techniques to mitigate risks.
