A system prompt to prevent prompt leakage and adversarial attacks in GPTs.
Dataset for classifying prompts as jailbreak or benign to enhance LLM safety.
A collection of state-of-the-art jailbreak methods for LLMs, including papers, codes, datasets, and analyses.
A dataset of 15,140 ChatGPT prompts, including 1,405 jailbreak prompts, collected from various platforms for research purposes.