AI Security Tools & Resources

Curated collection of 65 tools and frameworks for AI/ML security across 10 categories

65Tools
56Open Source
10Categories
Category:
65 tools

Red Teaming

Tools for offensive security testing of AI/ML systems

10 tools

Garak

2.8k

LLM vulnerability scanner and red teaming framework from NVIDIA Research

Maintained by NVIDIA Research
LLM TestingRed TeamAutomationNVIDIA

PyRIT

1.5k

Microsoft Python Risk Identification Tool for generative AI red teaming

Maintained by Microsoft
MicrosoftRed TeamPythonAutomation

PromptFoo

3.2k

LLM evaluation and red teaming framework with security testing plugins

EvaluationRed TeamCI/CDTesting

AI Goat

450

Vulnerable LLM application for security training and testing

TrainingCTFVulnerable AppLearning

Prompt Injection Playground

320

Interactive environment for testing prompt injection techniques

Prompt InjectionPlaygroundLearning

Counterfit

680

Microsoft tool for adversarial ML attack simulation

Maintained by Microsoft
MicrosoftAttack SimulationCLI

Caldera Adversary Emulation

5.2k

MITRE adversary emulation platform with ML attack plugins

Maintained by MITRE
MITREEmulationRed Team

DeepTeam

1.2k

Open-source LLM red teaming framework with guardrail testing capabilities

Maintained by Confident AI
Red TeamGuardrailsOpen SourceLLM

AutoRTAI

Agent-based automated red teaming at scale for AI systems

Maintained by HiddenLayer
HiddenLayerAgent-basedScalableEnterprise

Crucible

AI/ML vulnerability practice platform by Dreadnode

Maintained by Dreadnode
DreadnodePracticeCTFLearning

Vulnerability Scanners

Automated scanning tools for AI security vulnerabilities

8 tools

Nuclei

17.2k

Fast vulnerability scanner with AI/LLM security templates

Maintained by ProjectDiscovery
TemplatesAutomationCI/CD

Semgrep

9.8k

Static analysis tool with rules for AI/ML security vulnerabilities

SASTCode AnalysisCustom Rules

Trivy

21.3k

Container and IaC scanner for AI model deployments

Maintained by Aqua Security
ContainerKubernetesSupply Chain

ModelScan

850

Security scanner for ML model files detecting malicious code

Maintained by Protect AI
Model ScanningMalware DetectionCI/CD

AIShield

Automated vulnerability assessment for ML models

Maintained by Bosch
EnterpriseModel SecurityAutomation

Mindgard DAST-AI

Dynamic application security testing specialized for AI systems

Maintained by Mindgard
DASTDynamic TestingEnterpriseAI Security

Woodpecker

AI, Kubernetes, and API vulnerability testing platform

Maintained by Operant AI
Operant AIK8sAPIInfrastructure

BurpGPT

1.1k

LLM-integrated extension for Burp Suite web security testing

Burp SuiteWeb SecurityLLM IntegrationPentesting

LLM Defense

Defensive tools for protecting LLM applications

7 tools

LLM Guard

1.2k

Security toolkit for LLM applications with input/output validation

Maintained by Protect AI
Input ValidationOutput FilteringPII Detection

NeMo Guardrails

3.8k

NVIDIA toolkit for building safe and controllable LLM applications

Maintained by NVIDIA
Safety RailsNVIDIAEnterpriseColang

Rebuff

680

Self-hardening prompt injection detection service

Maintained by Protect AI
Prompt InjectionDetectionAPI

Vigil

420

LLM prompt injection scanner and detection system

DetectionScannerMonitoring

Guardrails AI

3.5k

Framework for adding validation and structured outputs to LLMs

ValidationStructured OutputEnterprise

LangKit

380

WhyLabs toolkit for LLM monitoring and safety

Maintained by WhyLabs
MonitoringSafetyMetrics

Lakera Guard

Runtime protection layer for generative AI applications

Maintained by Lakera
LakeraRuntime ProtectionGuardrailsEnterprise

Adversarial Testing

Frameworks for adversarial attack simulation

7 tools

Adversarial Robustness Toolbox

4.5k

IBM library for adversarial ML attacks and defenses

Maintained by IBM
IBMDefenseAttackComprehensive

TextAttack

2.8k

Framework for adversarial attacks on NLP models

NLPAdversarialResearch

CleverHans

6.1k

Library for adversarial example generation and robustness testing

TensorFlowPyTorchJAX

Foolbox

2.6k

Python toolbox for adversarial attacks with multiple backends

Multi-frameworkResearchBenchmarking

SecML

350

Machine learning security library for attack and defense

ResearchAttacksDefenses

ARTKIT

420

Framework for multi-turn attacker-target simulations in LLM systems

Multi-turnSimulationAttack PatternsOpen Source

BrokenHill

520

GCG-based jailbreak attack generator for LLM security testing

Maintained by Bishop Fox
Bishop FoxGCGJailbreakAdversarial

Evaluation & Benchmarking

Tools for evaluating AI safety and security

9 tools

HarmBench

420

Benchmark for evaluating automated red teaming and LLM safety

Maintained by Center for AI Safety
BenchmarkSafetyRed Team

TrustLLM

380

Comprehensive benchmark for LLM trustworthiness

TrustworthinessBenchmarkSafety

DecodingTrust

320

Comprehensive trustworthiness evaluation for GPT models

GPTTrustworthinessComprehensive

Inspect AI

680

UK AI Safety Institute evaluation framework

Maintained by UK AI Safety Institute
GovernmentSafetyEvaluation

LM Evaluation Harness

5.2k

Framework for evaluating language models including safety

Maintained by EleutherAI
EvaluationBenchmarkingComprehensive

JailbreakBench

650

Standardized benchmark for evaluating LLM jailbreak defenses

NeurIPS 2024BenchmarkJailbreakEvaluation

HELM Safety

1.9k

Stanford safety benchmark suite for holistic LLM evaluation

Maintained by Stanford CRFM
StanfordBenchmarkSafetyHolistic

OS-HARM

180

Safety benchmark for computer-use AI agents

Agent SafetyComputer UseBenchmarkResearch

OpenAI Evals

14.5k

Framework for evaluating LLM behavior and capabilities

Maintained by OpenAI
OpenAIEvaluationBenchmarkingOpen Source

Fuzzing

Automated fuzzing tools for LLM testing

4 tools

GPTFuzzer

580

Automated fuzzing framework for jailbreak testing

FuzzingJailbreakAutomation

FuzzLLM

280

Fuzzing framework for LLM-based applications

FuzzingLLMSecurity Testing

PromptBench

1.8k

Benchmark for evaluating LLM robustness to adversarial prompts

Maintained by Microsoft
MicrosoftBenchmarkRobustness

FuzzyAI

850

Automated fuzzing framework using genetic algorithms for AI model testing

Maintained by CyberArk
CyberArkFuzzingGenetic AlgorithmsAutomated

Privacy & Data Protection

Tools for PII detection and privacy-preserving ML

5 tools

Presidio

3.2k

Microsoft data protection and PII detection for AI systems

Maintained by Microsoft
PIIMicrosoftAnonymization

Opacus

1.6k

Differential privacy library for PyTorch training

Maintained by Meta
Differential PrivacyPyTorchMeta

TensorFlow Privacy

1.9k

Privacy-preserving machine learning library

Maintained by Google
TensorFlowGoogleDP-SGD

Private AI

Enterprise PII detection and redaction API

APIEnterpriseCompliance

PII Catcher

420

Automated PII detection for databases and data pipelines

DatabaseScanningAutomation

Model Security

Tools for securing ML model files and deployments

4 tools

AI Verify

320

Singapore AI governance testing framework and toolkit

Maintained by AI Verify Foundation
GovernanceTestingCompliance

MLflow

17.8k

ML lifecycle platform with security and governance features

MLOpsTrackingDeployment

Weights & Biases

8.4k

ML experiment tracking with security and collaboration features

MonitoringVersioningCollaboration

DVC

12.8k

Version control for ML projects with security features

Version ControlDataReproducibility

Supply Chain Security

Tools for ML supply chain security

5 tools

Sigstore

2.1k

Signing and verification for ML model supply chain

Maintained by OpenSSF
SigningSLSAOpenSSF

SLSA Framework

1.5k

Supply chain security framework for ML artifacts

Maintained by Google
FrameworkGoogleStandards

In-toto

850

Software supply chain security framework

Maintained by CNCF
AttestationCNCFProvenance

Syft

5.7k

SBOM generation for ML containers and dependencies

Maintained by Anchore
SBOMContainerDependencies

SafeTensors

2.4k

Safe serialization format for ML models preventing code execution

Maintained by Hugging Face
SerializationSecurityHugging Face

Monitoring & Observability

Tools for monitoring AI system security

6 tools

Langfuse

4.8k

Open source LLM observability and security monitoring

MonitoringTracingAnalytics

Arize AI

ML observability platform with security monitoring

EnterpriseObservabilityML Monitoring

Evidently AI

4.6k

Open source ML monitoring and testing platform

MonitoringTestingOpen Source

Fiddler AI

ML model performance and safety monitoring

EnterpriseExplainabilityMonitoring

Deepchecks

3.2k

Testing and monitoring for ML models and data

TestingMonitoringValidation

Galah

380

LLM-powered web honeypot for threat intelligence gathering

HoneypotThreat IntelLLM-poweredDetection

Additional Resources

Documentation

OWASP Top 10 for LLM Applications

Comprehensive guide to the most critical security risks in LLM applications

Visit Resource
Framework

MITRE ATLAS

Adversarial threat landscape for AI systems

Visit Resource
Standard

NIST AI Risk Management Framework

Comprehensive framework for managing AI risks

Visit Resource
Database

AI Incident Database

Repository of AI failures and incidents

Visit Resource
Platform

Hugging Face Model Security

Security scanning for Hugging Face models

Visit Resource
Framework

Google Secure AI Framework

Google approach to securing AI systems

Visit Resource