A PyTorch adversarial library for attack and defense methods on images and graphs.
Advbox is a toolbox for generating adversarial examples to test the robustness of neural networks across various frameworks.
A Python toolbox for adversarial robustness research, implemented in PyTorch.
TextAttack is a Python framework for adversarial attacks, data augmentation, and model training in NLP.
A Python library designed to enhance machine learning security against adversarial threats.
A Python toolbox to create adversarial examples that fool neural networks in PyTorch, TensorFlow, and JAX.
An adversarial example library for constructing attacks, building defenses, and benchmarking both.
AgentFence is an open-source platform for automatically testing AI agent security, identifying vulnerabilities like prompt injection and secret leakage.
A novel approach to hacking AI assistants using Unicode Tags to bypass security measures in large language models.
This research proposes defense strategies against prompt injection in large language models to improve their robustness and security against unwanted outputs.
GitHub repository for techniques to prevent prompt injection in AI chatbots using LLMs.