Category
Explore by categories

AI ModelsAI Security MonitoringJailbreak Prevention
GPTect
Details
A system prompt to prevent prompt leakage and adversarial attacks in GPTs.

AI ModelsAI Application PlatformsJailbreak Prevention
jackhhao/jailbreak-classification
Details
Dataset for classifying prompts as jailbreak or benign to enhance LLM safety.

Model Backdoor DefenseAI Security MonitoringJailbreak Prevention
Awesome-Jailbreak-on-LLMs
Details
A collection of state-of-the-art jailbreak methods for LLMs, including papers, codes, datasets, and analyses.

AI Ethics ResourcesAI Research PapersJailbreak Prevention
jailbreak_llms
Details
A dataset of 15,140 ChatGPT prompts, including 1,405 jailbreak prompts, collected from various platforms for research purposes.
