Tag
Explore by tags

AI ModelsAI Ethics ResourcesSecurity Research
Safety-Prompts
Details
Chinese safety prompts for evaluating and improving the safety of LLMs.

AI ModelsAI Ethics ResourcesAI Research Papers
CValues
Details
A research project assessing and aligning the values of Chinese large language models focusing on safety and responsibility.

AI Ethics ResourcesAI Research Papers
SafetyBench
Details
Official GitHub repository for SafetyBench, a benchmark to evaluate the safety of large language models (LLMs).

AI Research PapersModel Robustness EnhancementPrompt Injection Defense
InjecGuard
Details
The official implementation of InjecGuard, a tool for benchmarking and mitigating over-defense in prompt injection guardrail models.