
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.

灵取证是一款专业的安卓设备数据取证工具,专为执法部门和安全调查人员设计。

Python SDK for Agent AI Observability, Monitoring and Evaluation Framework.

Repository accompanying a paper on Red-Teaming for Large Language Models (LLMs).

Breaker AI is a CLI tool that detects prompt injection risks and vulnerabilities in AI prompts.

Red AI Range (RAR) is a security platform for AI red teaming and vulnerability assessment using Docker.

A CLI that provides a generic automation layer for assessing the security of ML models.

AgentFence is an open-source platform for automatically testing AI agent security, identifying vulnerabilities like prompt injection and secret leakage.

The official implementation of InjecGuard, a tool for benchmarking and mitigating over-defense in prompt injection guardrail models.

A benchmark for evaluating prompt injection detection systems.

This repository provides a benchmark for prompt Injection attacks and defenses.

Open-source tool by AIShield for AI model insights and vulnerability scans, securing the AI supply chain.