Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
灵取证是一款专业的安卓设备数据取证工具,专为执法部门和安全调查人员设计。
Python SDK for Agent AI Observability, Monitoring and Evaluation Framework.
Repository accompanying a paper on Red-Teaming for Large Language Models (LLMs).
Breaker AI is a CLI tool that detects prompt injection risks and vulnerabilities in AI prompts.
Red AI Range (RAR) is a security platform for AI red teaming and vulnerability assessment using Docker.
A CLI that provides a generic automation layer for assessing the security of ML models.
AgentFence is an open-source platform for automatically testing AI agent security, identifying vulnerabilities like prompt injection and secret leakage.
The official implementation of InjecGuard, a tool for benchmarking and mitigating over-defense in prompt injection guardrail models.
A benchmark for evaluating prompt injection detection systems.
This repository provides a benchmark for prompt Injection attacks and defenses.
Open-source tool by AIShield for AI model insights and vulnerability scans, securing the AI supply chain.