The Best Your Ultimate AI Security Toolkit
Curated AI security tools & LLM safety resources for cybersecurity professionals
Curated AI security tools & LLM safety resources for cybersecurity professionals

AdaptixC2 is an extensible post-exploitation framework for penetration testers, supporting multi-platform operations.

An experimental high-performance DNS query bruteforce tool built with AF_XDP for bulk DNS lookups.

Python SDK for AI agent monitoring, LLM cost tracking, benchmarking, and more.

Python SDK for Agent AI Observability, Monitoring and Evaluation Framework.

A curated list of awesome LLM agents frameworks.

A comprehensive library for implementing LLMs with a unified training pipeline and model evaluation.

A guidebook sharing insights and knowledge about evaluating Large Language Models (LLMs).

Latitude is the open-source prompt engineering platform to build, evaluate, and refine your prompts with AI.

A comprehensive collection of papers focused on evaluating large language models (LLMs).

Automatically evaluate your LLMs in Google Colab with LLM AutoEval.

Official GitHub repository for SafetyBench, a benchmark to evaluate the safety of large language models (LLMs).

Self-evaluating interview for AI coders.