Newsletter
Join the Community
Subscribe to our newsletter for the latest news and updates
Dataset for classifying prompts as jailbreak or benign to enhance LLM safety.
The jackhhao/jailbreak-classification dataset is designed to classify prompts as either jailbreak or benign. This dataset is crucial for enhancing the safety of large language models (LLMs) by helping to detect and prevent harmful jailbreak prompts that users might employ when interacting with these models.