AI Security Tools & Resources
Curated collection of 65 tools and frameworks for AI/ML security across 10 categories
Red Teaming
Tools for offensive security testing of AI/ML systems
Garak
LLM vulnerability scanner and red teaming framework from NVIDIA Research
PyRIT
Microsoft Python Risk Identification Tool for generative AI red teaming
PromptFoo
LLM evaluation and red teaming framework with security testing plugins
AI Goat
Vulnerable LLM application for security training and testing
Prompt Injection Playground
Interactive environment for testing prompt injection techniques
Counterfit
Microsoft tool for adversarial ML attack simulation
Caldera Adversary Emulation
MITRE adversary emulation platform with ML attack plugins
DeepTeam
Open-source LLM red teaming framework with guardrail testing capabilities
AutoRTAI
Agent-based automated red teaming at scale for AI systems
Crucible
AI/ML vulnerability practice platform by Dreadnode
Vulnerability Scanners
Automated scanning tools for AI security vulnerabilities
Nuclei
Fast vulnerability scanner with AI/LLM security templates
Semgrep
Static analysis tool with rules for AI/ML security vulnerabilities
Trivy
Container and IaC scanner for AI model deployments
ModelScan
Security scanner for ML model files detecting malicious code
AIShield
Automated vulnerability assessment for ML models
Mindgard DAST-AI
Dynamic application security testing specialized for AI systems
Woodpecker
AI, Kubernetes, and API vulnerability testing platform
LLM Defense
Defensive tools for protecting LLM applications
LLM Guard
Security toolkit for LLM applications with input/output validation
NeMo Guardrails
NVIDIA toolkit for building safe and controllable LLM applications
Rebuff
Self-hardening prompt injection detection service
Guardrails AI
Framework for adding validation and structured outputs to LLMs
LangKit
WhyLabs toolkit for LLM monitoring and safety
Lakera Guard
Runtime protection layer for generative AI applications
Adversarial Testing
Frameworks for adversarial attack simulation
Adversarial Robustness Toolbox
IBM library for adversarial ML attacks and defenses
CleverHans
Library for adversarial example generation and robustness testing
Foolbox
Python toolbox for adversarial attacks with multiple backends
SecML
Machine learning security library for attack and defense
ARTKIT
Framework for multi-turn attacker-target simulations in LLM systems
Evaluation & Benchmarking
Tools for evaluating AI safety and security
HarmBench
Benchmark for evaluating automated red teaming and LLM safety
TrustLLM
Comprehensive benchmark for LLM trustworthiness
DecodingTrust
Comprehensive trustworthiness evaluation for GPT models
Inspect AI
UK AI Safety Institute evaluation framework
LM Evaluation Harness
Framework for evaluating language models including safety
JailbreakBench
Standardized benchmark for evaluating LLM jailbreak defenses
HELM Safety
Stanford safety benchmark suite for holistic LLM evaluation
OS-HARM
Safety benchmark for computer-use AI agents
Fuzzing
Automated fuzzing tools for LLM testing
GPTFuzzer
Automated fuzzing framework for jailbreak testing
PromptBench
Benchmark for evaluating LLM robustness to adversarial prompts
Privacy & Data Protection
Tools for PII detection and privacy-preserving ML
Presidio
Microsoft data protection and PII detection for AI systems
Opacus
Differential privacy library for PyTorch training
TensorFlow Privacy
Privacy-preserving machine learning library
Model Security
Tools for securing ML model files and deployments
AI Verify
Singapore AI governance testing framework and toolkit
MLflow
ML lifecycle platform with security and governance features
Weights & Biases
ML experiment tracking with security and collaboration features
Supply Chain Security
Tools for ML supply chain security
Sigstore
Signing and verification for ML model supply chain
SLSA Framework
Supply chain security framework for ML artifacts
In-toto
Software supply chain security framework
Syft
SBOM generation for ML containers and dependencies
Monitoring & Observability
Tools for monitoring AI system security
Langfuse
Open source LLM observability and security monitoring
Arize AI
ML observability platform with security monitoring
Evidently AI
Open source ML monitoring and testing platform
Additional Resources
OWASP Top 10 for LLM Applications
Comprehensive guide to the most critical security risks in LLM applications
Visit ResourceNIST AI Risk Management Framework
Comprehensive framework for managing AI risks
Visit Resource