ai-security-tooling

Use this skill when adding or organizing:

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "ai-security-tooling" with this command: npx skills add gmh5225/awesome-ai-security/gmh5225-awesome-ai-security-ai-security-tooling

AI Security Tooling

Scope

Use this skill when adding or organizing:

  • LLM security tools (guardrails, detectors)

  • Adversarial ML libraries

  • AI vulnerability scanners

  • Model safety tools

  • Security benchmarks and frameworks

Tool Categories

LLM Security Tools

  • Guardrails: NeMo Guardrails, LLM Guard, Rebuff

  • Detectors: Vigil-LLM, Nova Framework, Garak

  • Scanners: ModelScan, AI Security Analyzer

Adversarial ML Libraries

  • Attack libraries: ART, CleverHans, Foolbox, TextAttack

  • Defense libraries: SecML

  • Fuzzing: OSS-Fuzz-Gen, Brainstorm

AI Red Teaming

  • Microsoft: Counterfit, PyRIT

  • Meta: PurpleLlama

  • NVIDIA: Garak, NeMo Guardrails

Benchmarks

  • Robustness: RobustBench

  • Jailbreak: JailbreakBench

  • Safety: Stanford AIR-Bench

  • Hallucination: Vectara Leaderboard

Standards & Frameworks

  • MITRE ATLAS: AI threat matrix

  • NIST AI RMF: Risk management framework

  • OWASP: LLM Top 10, GenAI Security Project

Categorization Rules

  • LLM guardrails/detectors → AI Security & Attacks → Model Security

  • Prompt injection tools → AI Security & Attacks → Prompt Injection

  • Adversarial ML libraries → AI Security & Attacks → Adversarial Attacks or AI Security Libraries

  • AI RE/debugging tools → AI Security Tools & Frameworks → AI Reverse Engineering

  • AI vulnerability scanners → AI Security Tools & Frameworks → AI Vulnerability Detection

  • Benchmarks → Benchmarks & Standards

  • MCP security tools → AI Pentesting & Red Teaming → AI Security MCP Tools

Quality Bar

  • Prefer canonical repos

  • Avoid forks unless they add meaningful features

  • Add short descriptions

  • Never duplicate an existing URL

  • Tool must be AI/ML-focused

Key Vendor Tools

Vendor Tools

Microsoft Counterfit, PyRIT

Meta PurpleLlama (Llama Guard, Prompt Guard, Code Shield)

NVIDIA Garak, NeMo Guardrails

IBM Adversarial Robustness Toolbox (ART)

Google OSS-Fuzz-Gen

ProtectAI Rebuff, LLM Guard, ModelScan

Notes

Keep additions:

  • AI/ML security focused

  • Non-duplicated URLs

  • Minimal structural changes

Data Source

For detailed and up-to-date resources, fetch the complete list from:

https://raw.githubusercontent.com/gmh5225/awesome-ai-security/refs/heads/main/README.md

Use this URL to get the latest curated links when you need specific tools, papers, or resources not covered in this skill.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Security

llm-attacks-security

No summary provided by upstream source.

Repository SourceNeeds Review
Security

ai-powered-pentesting

No summary provided by upstream source.

Repository SourceNeeds Review
Security

adversarial-machine-learning

No summary provided by upstream source.

Repository SourceNeeds Review
Security

reverse-engineering-tools

No summary provided by upstream source.

Repository SourceNeeds Review