Curated reading list for adversarial perspective and robustness in deep reinforcement learning.
A CLI that provides a generic automation layer for assessing the security of ML models.
A PyTorch adversarial library for attack and defense methods on images and graphs.
Advbox is a toolbox for generating adversarial examples to test the robustness of neural networks across various frameworks.
A Python toolbox for adversarial robustness research, implemented in PyTorch.
A Python library designed to enhance machine learning security against adversarial threats.
A Python toolbox to create adversarial examples that fool neural networks in PyTorch, TensorFlow, and JAX.
An adversarial example library for constructing attacks, building defenses, and benchmarking both.
This research proposes defense strategies against prompt injection in large language models to improve their robustness and security against unwanted outputs.
GitHub repository for techniques to prevent prompt injection in AI chatbots using LLMs.