logo
0
0
WeChat Login
Rafael<rafael@brinhosa.com.br>
Merge pull request #1 from mldangelo/update-promptfoo

Awesome AI Security banner

Awesome AI Security

🔐🤖 A curated list of AI/LLM security tools, frameworks, guides, papers, and training — focused on open-source and community resources.

Awesome CC0 License GitHub stars PRs Welcome Last Commit


About

The awesome-ai-security repository is a community-driven collection of AI Security, LLM Security, and Prompt Injection tools and resources. The focus is on open-source tools and resources that benefit the community.

This repository covers:

  • Large Language Model (LLM) security testing and vulnerability assessment
  • Prompt injection attacks and defenses
  • AI red teaming and adversarial testing
  • Jailbreak detection and prevention
  • Model poisoning and extraction attacks
  • Hallucination detection and prevention
  • AI application security best practices
  • MLSecOps and LLMOps security
  • Model Context Protocol (MCP) security
  • AI supply chain security

📌 Please read the Contributions section before opening a pull request.


🧪 AI Security Testing Tools

NameAuthorDescription
garakNVIDIALLM vulnerability scanner – tests 120+ categories (hallucination, data leakage, prompt injection, misinformation, toxicity, jailbreaks).
PyRITMicrosoftPython Risk Identification Tool for GenAI; adversarial testing automation with multi-turn orchestration.
promptmaputkusenAutomated prompt injection scanner with white-box testing.
aiapwnkarimhabushAutomatic prompt injection testing with tailored payload generation.
FuzzyAICyberArkLLM fuzzing framework to identify jailbreaks and vulns.
LLMFuzzermnnsFuzzing framework for LLM API integrations.
promptfoopromptfooAdaptive red teaming for LLM agents with multi-turn attacks (PAIR, tree-of-attacks, crescendo) that probe tool use, RAG, and agentic workflows. Used by 250K+ developers; featured in OpenAI and Anthropic developer education.
LLM WardenjackhhaoSimple jailbreak detection (Hugging Face model).
VigildeadbitsModular scanners (vectors, YARA, transformers) via lib & REST API.
picklescanmmaitre314Detects malicious code in Python pickle model files.
ModelScanProtect AIMulti-format ML model file scanner (pickle, SavedModel, etc.).
Open-Prompt-InjectionYupei Liu et al.Toolkit/benchmark for prompt injection attacks/defenses.
ARTKITBCG-XOpen-source framework for automated LLM red-teaming with multi-turn attacker-target interactions.
GiskardGiskard AIAdvanced automated red-teaming platform with 50+ specialized probes and adaptive attack engine.
MindgardMindgardDAST-AI platform for automated red teaming across the AI lifecycle with artifact scanning.
CodeGateStacklokSecurity proxy for LLMs and IDEs that filters input/output to prevent API key leakage and insecure code.
AIJackKoukyosyumeiOpen-source simulator for modeling security and privacy threats targeting ML systems.
Strixusestrix"AI hacker" agents for CLI & CI/CD with automated security testing.

🧨 Prompt Injection Resources

NameAuthorDescription
PayloadsAllTheThings – Prompt InjectionswisskyrepoPrompt injection payloads and bypasses.
PIPE – Prompt Injection PrimerjthackAttack scenarios and payloads for engineers.
Basic-ML-prompt-injectionsZieraxEducational payloads.
OWASP LLM Prompt Injection PreventionOWASPPrevention cheat sheet and best practices.
NeuralTrust AI GuideNeuralTrustComprehensive guide to implementing prompt injection detection with real-time alerting.

🕵️ Jailbreak Detection & Red Teaming

NameAuthorDescription
Adversarial Robustness Toolbox (ART)IBM / LF AIDefenses against evasion, poisoning, extraction, inference attacks with 39 attack modules and 29 defense modules.
HEARTIBMHardened ART extension for T&E workflows.
RebuffProtect AISelf-hardening prompt injection detector (multi-layer).
PurpleLlamaMetaLlama Guard, CyberSecEval, and more.
HarmBenchCenter for AI SafetyStandardized evaluation framework for automated red teaming with 18 methods comparison.
Splx AISplx AICommercial platform for multi-modal AI red teaming with CI/CD integration.
Lasso MCP GatewayLasso SecurityOpen-source MCP Gateway for Model Context Protocol security testing.

🧩 Deliberately Vulnerable AI Applications

NameAuthorDescription
AI GoatdhammonLocal LLM CTF challenges; no fees/sign-ups.
GandalfLakeraPrompt injection game with difficulty levels (world's largest red team experiment).
LLM Security CTFTrustAI-laboratoryFree web-based vulnerable LLM CTFs.
DamnVulnerableLLMProjectharishsg993010DV LLM app for training/education.

🎓 Training, Labs & CTF Challenges

OrganizationNameDescription
SANSSEC545: GenAI & LLM AppSecHands-on GenAI security.
SANSSEC495: Building & Securing RAGRAG security training.
SANSSEC411: AI Security PrinciplesFundamentals with Docker labs.
SANSSANS AI Cybersecurity Summit 2025Hands-on workshops and live demos for AI/ML integration in cybersecurity (Denver, March 31 – April 7, 2025).
AppSecEngineerAI Combat & ConstructAttack/defend AI apps.
Practical DevSecOpsCAISP60 days of labs; MITRE ATLAS defenses.
HackAPromptHackAPrompt 1.0Prompt hacking competition.
HackAPromptHackAPrompt 2.0Large-scale red-teaming hackathon.
AI VillageDEF CON AI CTFAnnual LLM security CTF.

📚 Books & Publications

Author(s)PublisherNameDescription
VariousSpringerLarge Language Models in Cybersecurity: Threats, Exposure and MitigationOpen-access guide (2024).
VariousSpringerGenerative AI Security: Theories and PracticesGenAI impacts across security (2024).
VariousSpringerAI-Driven Cybersecurity and Threat IntelligenceAI x security (2024).
Steve WilsonO'ReillyDeveloper's Playbook for LLM SecurityPractical LLM AppSec (2024).

📎 Cheatsheets & Guides

NameAuthorDescription
OWASP LLM Prompt Injection PreventionOWASPPrevention best practices.
LangChain Security PolicyLangChain"Four Perimeters" and app hardening.
CISA AI Security Best PracticesCISAAI system security guidance.
NVIDIA AI Red Team Practical AdviceNVIDIAKey findings from AIRT assessments on securing AI-powered applications.
Salesforce Prompt Injection Detection GuideSalesforceBuilding trusted AI systems against prompt injection threats.

🧭 Frameworks & Standards

NameOrgDescription
OWASP Top 10 for LLM Apps (2025)OWASPLLM01–LLM10 risks including new entries: System Prompt Leakage (LLM07), Vector and Embedding Weaknesses (LLM08), Misinformation (LLM09), Unbounded Consumption (LLM10).
NIST AI RMF + GenAI ProfileNISTGovern, Map, Measure, Manage.
MITRE ATLASMITREAI adversary TTPs (modeled after ATT&CK).
CISA AI GuidelinesCISAJoint guidance for AI/ML systems.
OWASP Top 10 2025OWASPUpdated to include A03: Software Supply Chain Failures and A10: Mishandling of Exceptional Conditions.

🛡️ Defense & Guardrails

NameAuthorDescription
NeMo GuardrailsNVIDIAProgrammable input/output/dialog/retrieval/execution controls.
LLM GuardProtect AIRuntime scanning, PII redaction, content filtering.
LocalModKOKOSdeSelf-hosted content moderation with prompt injection, toxicity, PII, and NSFW detection. 100% offline.
Guardrails AIGuardrails AIValidation rules & structured outputs using RAIL.
Lakera GuardLakeraReal-time prompt injection/jailbreak detection with near-real-time alerts.
Prompt ArmorPrompt ArmorReal-time detection and filtering of malicious prompts.
HiddenLayer AIM SecurityHiddenLayerAI application monitoring, real-time threat detection, and zero-trust access controls.
CalypsoAI ModeratorCalypsoAICommercial tool with audit trails, malicious code detection, and data loss protection.

🎓 Certifications & Courses

OrgNameDescription
ISACAAAISM™AI Security Management (CISM/CISSP req.).
ISC2Building AI Strategy CertificateStrategy, governance, risk.
Practical DevSecOpsCAISPHands-on certification with labs.
SecuritiAI Security & GovernanceGovernance, privacy, compliance.

🎤 Conferences & Events

NameDateLocationDescription
DEF CONAug 2025Las VegasAI Village & GenAI red team challenges.
Black Hat USAAug 2025Las VegasAI Security Summit & trainings.
RSA ConferenceApr-May 2025San FranciscoAI security tracks, expo.
AI Risk SummitAug 19-20, 2025Ritz-Carlton, Half Moon Bay, CASecurity executives, AI researchers, and policymakers discuss adversarial AI, deepfakes, and regulatory challenges.
GCSCC AI Cybersecurity Conference 20252025Oxford, UKSecuring the Cyber Future: Cyber Resilience in the Age of AI and Geopolitical Uncertainty.
AI Security & Privacy Conference 20252025TBD400+ CISOs and C-Level Executives with expert-led discussions and case studies.
Cyber-AI 2025 ConferenceSep 1-4, 2025Varna, BulgariaFour-day conference on cutting-edge advancements in cybersecurity and AI.
AI VillageOngoingVirtual/VariousCommunity, meetups, and CTFs.

📄 Research Papers & Datasets

NameTopicDescription
Ignore This Title and HackAPromptPrompt InjectionEMNLP'23; taxonomy of prompt hacking.
SelfCheckGPTHallucinationSelf-consistency for hallucination detection.
Survey on Model Extraction Attacks (2025)Model SecuritySurvey of extraction attacks/defenses.
SECURE BenchmarkCybersecurityMulti-dataset security evaluation suite.
TruthfulQASafetyTruthfulness under misconceptions.
ToxiGenSafetyToxicity dataset & benchmarks.
In-The-Wild Jailbreak Prompts DatasetJailbreak15,140 prompts with 1,405 jailbreak prompts from Reddit, Discord, websites (2022-2023).
JailbreakBenchJailbreakOpen-source robustness benchmark with 200 distinct behaviors and jailbreak artifacts.
JailBreakV-28KJailbreak28,000 jailbreak test cases for MLLMs (20K text-based, 8K image-based).
Forbidden Question SetSafetyCurated dataset of forbidden questions across high-risk categories.
LLM Jailbreak + Safety DataJailbreak~10K fine-tuning examples and ~3K adversarial prompts for chatbot safety.

📈 Observability & Monitoring

NameAuthorDescription
LangSmithLangChainTracing + evals for LLM apps.
Weights & BiasesWeights & BiasesExperiment tracking & prompt management for LLMs.
LangfuseLangfuseOpen-source tracing & cost monitoring.
PhoenixArize AIOpen-source eval/monitoring.
HeliconeHeliconeProxy-based logging & analytics.
Dynatrace Davis AIDynatraceAI-driven root cause analysis with multidimensional baselining and predictive analytics.

🗡️ Penetration Testing Tools

NameAuthorDescription
PentestGPTGreyDGLGPT-powered pentesting assistant.
AI-penetration-testingMr-InfectCurated offensive/defensive AI pentest techniques.
PentAGIvxcontrolAutonomous agent system for pentesting.
AI-OPSantoninoLorenzoAssistant for exploit dev & research.
HackSynthaielte-researchPlanner + summarizer pentest agent.
HexStrike AI MCP0x4m4150+ tools + AI agents automation.
Strixusestrix"AI hacker" agents; CLI & CI/CD.
BurpGPTBurp SuiteBurp Suite extension integrating LLMs for enhanced vulnerability scanning and traffic analysis.

🔐 MCP Security

Model Context Protocol (MCP) security resources

NameAuthorDescription
Lasso MCP GatewayLasso SecurityFirst security-centric open-source solution for Model Context Protocol.

⭐ Awesome Lists

NameAuthorDescription
awesome-llm-securitycorca-aiLLM Security resources.
awesome-gpt-securitycckuailongSecurity tools & cases for GPT apps.
awesome-llm-cybersecurity-toolsTenableLLM tools for cybersecurity.
Awesome-LLMSecOpswearetyomsmnvLLM SecOps lifecycle & threats.
awesome-llm-supply-chain-securityShenaoWSupply chain security resources.
awesome-MLSecOpsRiccardoBiosasMLSecOps tools & best practices.
awesome-hallucination-detectionEdinburghNLPHallucination detection papers.
oss-llm-securitykaplanliorCurated list of open-source LLM security tools including EasyJailbreak, fast-llm-security, and more.

📰 Podcasts & Newsletters

NameHost/AuthorDescription
AI Security PodcastAshish Rajan & Caleb SimaVendor-neutral AI security conversations.
The AI Fix PodcastGraham Cluley & Mark StockleyDeepfakes, policy, and security.
Smashing SecurityGraham Cluley & Carole TheriaultWeekly infosec pod with AI topics.
Resilient Cyber NewsletterChris HughesAI, supply chain, cloud, AppSec.

📺 YouTube Channels

Cybersecurity and AI security YouTube channels

NameFocusDescription
PowerDMARCEmail SecurityEmail authentication, DMARC, spoofing, phishing, and fraud tactics.
John HammondGeneral Cybersecurity1.28M subscribers; CTF challenges, hacking tutorials, and real-time problem-solving.
The Cyber MentorEthical HackingPractical ethical hacking, penetration testing, and step-by-step tutorials.
NetworkChuckNetworking & SecurityExploring cybersecurity, networking, and technology concepts.
Hak5Hacking ToolsCybersecurity tools, privacy, tech gadgets, and entertaining tutorials.
MalwareTechMalware AnalysisDeep-dive malware analysis, cybersecurity research, and threat intelligence.
David BombalNetwork SecurityEthical hacking, network security, and certifications.
LiveOverflowBinary ExploitationLow-level security, reverse engineering, and CTF writeups.
CyberRisk TVAI SecurityBlack Hat 2025 coverage with focus on AI security, agentic AI, and trust.

🧰 Other Resources

NameAuthorDescription
Rez0's AI Security BlogJoseph ThackerAI hacking fundamentals & techniques.
Simon Willison's BlogSimon WillisonPrompt injection & agent security.
Lakera AI BlogLakera TeamGenAI security thought leadership.
Anthropic Transparency HubAnthropicSystem cards & red team reports.
OpenAI Red Teaming NetworkOpenAIRed teaming docs & invites.
MLSecOps CommunityCommunityBest practices & community.
OWASP GenAI Security ProjectOWASPGlobal community-driven initiative for GenAI security guidance and resources.
OWASP AI Security Solutions LandscapeOWASPLandmark guide outlining key risks and critical controls for securing LLMs and GenAI applications.
Lasso Security BlogLasso SecurityResources on LLM & AI cybersecurity, MCP security, and red teaming.

👤 Thought Leaders

NamePlatformNotability
Simon WillisonTwitter/X / BlogPrompt injection & agent security.
Joseph Thacker (rez0)Twitter/X / BlogProlific AI vuln research & guides.
Lakera TeamTwitter/XGandalf & Lakera Guard creators.
NVIDIA AI Red TeamTwitter/XTeam behind garak and practical security guidance.
Microsoft AI Red TeamTwitter/XPyRIT & public red teaming lessons.
Steve WilsonLinkedInOWASP Top 10 for LLM Applications Project Lead.
Ads DawsonLinkedInTechnical Lead & Vulnerability Entries Lead for OWASP Top 10 LLMs.

Contributions

  1. Purpose: Collect AI/LLM security & prompt-injection resources. Prefer open-source/community content.
  2. Out of Scope: Ads, closed-source/proprietary, trials/freemium, or items needing private details.
  3. Relevance: Must directly relate to AI/LLM security, jailbreaks, red teaming, model/app security.
  4. No Duplicates: Avoid redundant entries.
  5. Thought Leaders: Prefer figures tied to content/tools listed here.
  6. Accuracy: Authors can open issues/PRs to update their entries.
  7. Books: Paid books allowed for educational value.

How to contribute


License

CC0

To the extent possible under law, the contributors have waived all copyright and related or neighboring rights to this work.


⬆️ Back to top