logo
0
0
WeChat Login
wearetyomsmnv<129667548+wearetyomsmnv@users.noreply.github.com>
Update README.md

LLMSecOps

🚀 Awesome LLMSecOps

Awesome GitHub stars GitHub forks GitHub last commit

🔐 A curated list of awesome resources for LLMSecOps (Large Language Model Security Operations) 🧠

by @wearetyomsmnv and people

Architecture | Vulnerabilities | Tools | Defense | Threat Modeling | Jailbreaks | RAG Security | PoC's | Study Resources | Books | Blogs | Datasets for Testing | OPS Security | Frameworks | Best Practices | Research | Tutorials | Companies | Community Resources

LLM safety is a huge body of knowledge that is important and relevant to society today. The purpose of this Awesome list is to provide the community with the necessary knowledge on how to build an LLM development process - safe, as well as what threats may be encountered along the way. Everyone is welcome to contribute.

IMPORTANT

This repository, unlike many existing repositories, emphasizes the practical implementation of security and does not provide a lot of references to arxiv in the description.


Architecture risks

Overview of fundamental architectural risks and challenges in LLM systems.

RiskDescription
Recursive PollutionLLMs can produce incorrect output with high confidence. If such output is used in training data, it can cause future LLMs to be trained on polluted data, creating a feedback loop problem.
Data DebtLLMs rely on massive datasets, often too large to thoroughly vet. This lack of transparency and control over data quality presents a significant risk.
Black Box OpacityMany critical components of LLMs are hidden in a "black box" controlled by foundation model providers, making it difficult for users to manage and mitigate risks effectively.
Prompt ManipulationManipulating the input prompts can lead to unstable and unpredictable LLM behavior. This risk is similar to adversarial inputs in other ML systems.
Poison in the DataTraining data can be contaminated intentionally or unintentionally, leading to compromised model integrity. This is especially problematic given the size and scope of data used in LLMs.
Reproducibility EconomicsThe high cost of training LLMs limits reproducibility and independent verification, leading to a reliance on commercial entities and potentially unreviewed models.
Model TrustworthinessThe inherent stochastic nature of LLMs and their lack of true understanding can make their output unreliable. This raises questions about whether they should be trusted in critical applications.
Encoding IntegrityData is often processed and re-represented in ways that can introduce bias and other issues. This is particularly challenging with LLMs due to their unsupervised learning nature.

From Berryville Institute of Machine Learning (BIML) paper

Vulnerabilities description

by Giskard

Common vulnerabilities and security issues found in LLM applications.

VulnerabilityDescription
Hallucination and MisinformationThese vulnerabilities often manifest themselves in the generation of fabricated content or the spread of false information, which can have far-reaching consequences such as disseminating misleading content or malicious narratives.
Harmful Content GenerationThis vulnerability involves the creation of harmful or malicious content, including violence, hate speech, or misinformation with malicious intent, posing a threat to individuals or communities.
Prompt InjectionUsers manipulating input prompts to bypass content filters or override model instructions can lead to the generation of inappropriate or biased content, circumventing intended safeguards.
RobustnessThe lack of robustness in model outputs makes them sensitive to small perturbations, resulting in inconsistent or unpredictable responses that may cause confusion or undesired behavior.
Output FormattingWhen model outputs do not align with specified format requirements, responses can be poorly structured or misformatted, failing to comply with the desired output format.
Information DisclosureThis vulnerability occurs when the model inadvertently reveals sensitive or private data about individuals, organizations, or entities, posing significant privacy risks and ethical concerns.
Stereotypes and DiscriminationIf model's outputs are perpetuating biases, stereotypes, or discriminatory content, it leads to harmful societal consequences, undermining efforts to promote fairness, diversity, and inclusion.

LLMSecOps Life Cycle

Group 2

🛠 Tools for scanning

Security scanning and vulnerability assessment tools for LLM applications.

ToolDescriptionStars
🔧 GarakLLM vulnerability scannerGitHub stars
🔧 ps-fuzz 2Make your GenAI Apps Safe & Secure 🚀 Test & harden your system promptGitHub stars
🗺️ LLMmapTool for mapping LLM vulnerabilitiesGitHub stars
🛡️ Agentic SecuritySecurity toolkit for AI agentsGitHub stars
🔒 LLM ConfidentialityTool for ensuring confidentiality in LLMsGitHub stars
🔒 PyRITThe Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.GitHub stars
🔧 promptfooLLM red teaming and evaluation framework. Test for jailbreaks, prompt injection, and other vulnerabilities with adversarial attacks (PAIR, tree-of-attacks, crescendo). CI/CD integration.GitHub stars
🔧 LLaMatorFramework for testing vulnerabilities of large language models with support for Russian languageGitHub stars
🔧 SpikeeComprehensive testing framework for LLM applications. Tests prompt injection, jailbreaks, and other vulnerabilities. Supports custom targets, attacks, judges, and guardrail evaluationGitHub stars
🛡️ LocalModSelf-hosted content moderation API with prompt injection detection, toxicity filtering, PII detection, and NSFW filtering. Runs 100% offline.GitHub stars

🛡️Defense

Defensive mechanisms, guardrails, and security controls for protecting LLM applications.

Security by Design

CategoryMethod / TechnologyPrinciple of Operation (Mechanism)Examples of Use / Developers
1. Fundamental AlignmentRLHF (Reinforcement Learning from Human Feedback)Training a model with reinforcement learning based on a reward model, which is trained on human evaluations. It optimizes for "usefulness" and "safety."OpenAI (GPT-4), Yandex (YandexGPT)
DPO (Direct Preference Optimization)Direct optimization of response probabilities based on preference pairs, bypassing the creation of a separate reward model. It is described as more stable and effective.Meta (Llama 3), Mistral, open models
Constitutional AI / RLAIFUsing the model itself to criticize and correct its responses according to a set of rules ("Constitution"). AI replaces human labeling (RLAIF).Anthropic (Claude 3)
2. Internal Control (Interpretability)Representation Engineering (RepE)Detection and suppression of neuron activation vectors responsible for undesirable concepts (e.g., falsehood, lust for power) in real-time.Center for AI Safety (CAIS)
Circuit BreakersRedirection ("short-circuiting") of internal representations of malicious queries into orthogonal space, causing failure or nonsense.GraySwan AI, researchers
Machine UnlearningAlgorithmic "erasure" of dangerous knowledge or protected data from model weights (e.g., via Gradient Ascent) so that the model physically "forgets" them.Research groups, Microsoft
3. External Filters (Guardrails)Llama GuardA specialized LLM-classifier that checks incoming prompts and outgoing responses for compliance with a risk taxonomy (MLCommons).Meta
NeMo GuardrailsA programmable dialogue management system. It uses the Colang language for strict topic adherence and attack blocking.NVIDIA
Prompt Guard / ShieldsLightweight models (based on BERT/DeBERTA) for detecting jailbreaks and prompt injections before they reach the LLM.Meta, Azure AI
SmoothLLMA randomized smoothing method: creating copies of a prompt with symbolic perturbations to disrupt the structure of adversarial attacks (e.g., GCG suffixes).Researchers (SmoothLLM authors)
Google Safety FiltersMulti-level content filtering with customizable sensitivity thresholds and semantic vector analysis.Google (Gemini API)
4. System InstructionsSystem Prompts / TagsUsing special tokens (e.g., </start_header_id>) to separate system and user instructions.OpenAI, Meta, Anthropic
Instruction HierarchyPrioritizing system instructions over user instructions to protect against prompt injection, especially when the model learns to ignore "forget past instructions" commands.OpenAI (GPT-4o Mini)
5. Testing (Red Teaming)Automated Attacks (GCG, AutoDAN)Using algorithms and other LLMs to generate hundreds of thousands of adversarial prompts to find vulnerabilities.Research groups
ToolDescriptionStars
🛡️ PurpleLlamaSet of tools to assess and improve LLM security.GitHub stars
🛡️ RebuffAPI with built-in rules for identifying prompt injection and detecting data leakage through canary words. (ProtectAI is now part of Palo Alto Networks)GitHub stars
🔒 LLM GuardSelf-hostable tool with multiple prompt and output scanners for various security issues.GitHub stars
🚧 NeMo GuardrailsTool that protects against jailbreak and hallucinations with customizable rulesets.GitHub stars
👁️ VigilOffers dockerized and local setup options, using proprietary HuggingFace datasets for security detection.GitHub stars
🧰 LangKitProvides functions for jailbreak detection, prompt injection, and sensitive information detection.GitHub stars
🛠️ GuardRails AIFocuses on functionality, detects presence of secrets in responses.GitHub stars
🦸 Hyperion AlphaDetects prompt injections and jailbreaks.N/A
🛡️ LLM-GuardTool for securing LLM interactions. (ProtectAI is now part of Palo Alto Networks)GitHub stars
🚨 WhistleblowerTool for detecting and preventing LLM vulnerabilities.GitHub stars
🔍 PlexiglassSecurity tool for LLM applications.GitHub stars
🔍 Prompt Injection defensesRules for protected LLMGitHub stars
🔍 LLM Data ProtectorTools for protected LLM in chatbotsN/A
🔍 Gen AI & LLM Security for developers: Prompt attack mitigations on GeminiSecurity tool for LLM applications.GitHub stars
🔍 TrustGateGenerative Application Firewall that detects and blocks attacks against GenAI Applications.GitHub stars
🛡️ TenuoCapability tokens for AI agents with task-scoped TTLs, offline verification and proof-of-possession binding.GitHub stars
🛡️ AIDEFENDPractical knowledge base for AI security defenses. Based on MAESTRO framework, MITRE D3FEND, ATLAS, ATT&CK, Google Secure AI Framework, and OWASP Top 10 LLM 2025/ML Security 2023.N/A

Threat Modeling

Frameworks and methodologies for identifying and modeling threats in LLM systems.

ToolDescription
Secure LLM Deployment: Navigating and Mitigating Safety RisksResearch paper on LLM security [sorry, but is really cool]
ThreatModelsRepository for LLM threat models
Threat Modeling LLMsAI Village resource on threat modeling for LLMs
Sberbank AI Cybersecurity Threat ModelSberbank's threat model for AI cybersecurity
Pangea Attack TaxonomyComprehensive taxonomy of AI/LLM attacks and vulnerabilities

image image

Monitoring

Tools and platforms for monitoring LLM applications, detecting anomalies, and tracking security events.

ToolDescription
LangfuseOpen Source LLM Engineering Platform with security capabilities.
HiveTraceLLM monitoring and security platform for GenAI applications. Detects prompt injection, jailbreaks, malicious HTML/Markdown elements, and PII. Provides real-time anomaly detection and security alerts.

Watermarking

Tools and techniques for watermarking LLM-generated content to detect AI-generated text.

ToolDescription
MarkLLMAn Open-Source Toolkit for LLM Watermarking.

Jailbreaks

Resources, databases, and benchmarks for understanding and testing jailbreak techniques against LLMs.

ResourceDescriptionStars
JailbreakBenchWebsite dedicated to evaluating and analyzing jailbreak methods for language modelsN/A
L1B3RT45GitHub repository containing information and tools related to AI jailbreakingGitHub stars
llm-hacking-databaseThis repository contains various attacks against Large Language ModelsGitHub stars
HaizeLabs jailbreak DatabaseThis database contains jailbreaks for multimodal language modelsN/A
Lakera PINT BenchmarkA comprehensive benchmark for prompt injection detection systems. Evaluates detection systems across multiple categories (prompt injection, jailbreak, hard negatives, chat, documents) and supports evaluation in 20+ languages. Open-source benchmark with Jupyter notebook for custom evaluations.GitHub stars
EasyJailbreakAn easy-to-use Python framework to generate adversarial jailbreak promptsGitHub stars

LLM Interpretability

Resources for understanding and interpreting LLM behavior, decision-making, and internal mechanisms.

ResourceDescription
Интерпретируемость LLMDmitry Kolodezev's web page, which provides useful resources with LLM interpretation techniques

PINT Benchmark scores (by lakera)

Prompt Injection Test (PINT) benchmark scores comparing different prompt injection detection systems.

NamePINT ScoreTest Date
Lakera Guard95.2200%2025-05-02
Azure AI Prompt Shield for Documents89.1241%2025-05-02
protectai/deberta-v3-base-prompt-injection-v279.1366%2025-05-02
Llama Prompt Guard 2 (86M)78.7578%2025-05-05
Google Model Armor70.0664%2025-08-27
Aporia Guardrails66.4373%2025-05-02
Llama Prompt Guard61.8168%2025-05-02

Note: ProtectAI is now part of Palo Alto Networks

Hallucinations Leaderboard

Top 25 Hallucination Rates

Note: For the complete and most up-to-date interactive leaderboard, visit the Hugging Face leaderboard or the GitHub repository.

From this repo (last updated: December 18, 2025)

image

This is a Safety Benchmark from Stanford University


RAG Security

Security considerations, attacks, and defenses for Retrieval-Augmented Generation (RAG) systems.

ResourceDescription
Security Risks in RAGArticle on security risks in Retrieval-Augmented Generation (RAG)
How RAG Poisoning Made LLaMA3 RacistBlog post about RAG poisoning and its effects on LLaMA3
Adversarial AI - RAG Attacks and MitigationsGitHub repository on RAG attacks, mitigations, and defense strategies
PoisonedRAGGitHub repository about poisoned RAG systems
ConfusedPilot: Compromising Enterprise Information Integrity and Confidentiality with Copilot for Microsoft 365Article about RAG vulnerabilities
Awesome Jailbreak on LLMs - RAG AttacksCollection of RAG-based LLM attack techniques

image

Agentic security

Security tools, benchmarks, and research focused on autonomous AI agents and their vulnerabilities.

ToolDescriptionStars
invariantA trace analysis tool for AI agents.GitHub stars
AgentBenchA Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)GitHub stars
Agent Hijacking, the true impact of prompt injectionGuide for attack langchain agentsArticle
Breaking Agents: Compromising Autonomous LLM Agents Through Malfunction AmplificationResearch about typical agent vulnerabilitiesArticle
Model Context Protocol (MCP) at First Glance: Studying the Security and Maintainability of MCP ServersFirst large-scale empirical study of MCP servers security and maintainabilityArticle
Awesome MCP SecurityCurated list of MCP security resourcesGitHub stars
Awesome LLM Agent SecurityComprehensive collection of LLM agent security resources, attacks, vulnerabilitiesGitHub stars
MCP Security AnalysisResearch paper on MCP security vulnerabilities and analysisArticle
TenuoCapability-based authorization framework for AI agents. Task-scoped warrants with cryptographic attenuation, PoP binding, offline verification. LangChain/LangGraph/MCP integrations.GitHub stars

Agentic Browser Security

Security research and analysis of AI-powered browser agents and their unique attack vectors.

ResourceDescriptionSource
From Inbox to Wipeout: Perplexity Comet's AI Browser Quietly Erasing Google DriveResearch on zero-click Google Drive wiper attack via Perplexity Comet. Shows how polite, well-structured emails can trigger destructive actions in agentic browsers.Straiker STAR Labs
Agentic Browser Security AnalysisResearch paper on security vulnerabilities in agentic browsersArticle
Browser AI Agents: The New Weakest LinkAnalysis of security risks in browser-based AI agentsSqrx Labs
Comet Prompt Injection VulnerabilityBrave's analysis of prompt injection vulnerabilities in Perplexity Comet browserBrave

PoC

Proof of Concept implementations demonstrating various LLM attacks, vulnerabilities, and security research.

ToolDescriptionStars
Visual Adversarial ExamplesJailbreaking Large Language Models with Visual Adversarial ExamplesGitHub stars
Weak-to-Strong GeneralizationWeak-to-Strong Generalization: Eliciting Strong Capabilities With Weak SupervisionGitHub stars
Image HijacksRepository for image-based hijacks of large language modelsGitHub stars
CipherChatSecure communication tool for large language modelsGitHub stars
LLMs Finetuning SafetySafety measures for fine-tuning large language modelsGitHub stars
Virtual Prompt InjectionTool for virtual prompt injection in language modelsGitHub stars
FigStepJailbreaking Large Vision-language Models via Typographic Visual PromptsGitHub stars
stealing-part-lm-supplementarySome code for "Stealing Part of a Production Language Model"GitHub stars
Hallucination-AttackAttack to induce LLMs within hallucinationsGitHub stars
llm-hallucination-surveyReading list of hallucination in LLMs. Check out our new survey paper: "Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language Models"GitHub stars
LMSanitatorLMSanitator: Defending Large Language Models Against Stealthy Prompt Injection AttacksGitHub stars
ImperioImperio: Robust Prompt Engineering for Anchoring Large Language ModelsGitHub stars
Backdoor Attacks on Fine-tuned LLaMABackdoor Attacks on Fine-tuned LLaMA ModelsGitHub stars
CBAConsciousness-Based Authentication for LLM SecurityGitHub stars
MuScleLoRAA Framework for Multi-scenario Backdoor Fine-tuning of LLMsGitHub stars
BadActsBadActs: Backdoor Attacks against Large Language Models via Activation SteeringGitHub stars
TrojTextTrojan Attacks on Text ClassifiersGitHub stars
AnyDoorCreate Arbitrary Backdoor Instances in Language ModelsGitHub stars
PromptWareA Jailbroken GenAI Model Can Cause Real Harm: GenAI-powered Applications are Vulnerable to PromptWaresGitHub stars
BrokenHillAutomated attack tool that generates crafted prompts to bypass restrictions in LLMs using greedy coordinate gradient (GCG) attackGitHub stars
OWASP Agentic AIOWASP Top 10 for Agentic AI (AI Agent Security) - Pre-release versionGitHub stars

Study resource

Educational platforms, CTF challenges, courses, and training resources for learning LLM security.

ToolDescription
GandalfInteractive LLM security challenge game
Prompt AirlinesPlatform for learning and practicing prompt engineering
PortSwigger LLM AttacksEducational resource on WEB LLM security vulnerabilities and attacks
Invariant Labs CTF 2024CTF. You should hack LLM agentic
Invariant Labs CTF Summer 24Hugging Face Space with CTF challenges
CrucibleLLM security training platform
Poll Vault CTFCTF challenge with ML/LLM components
MyLLMDocLLM security training platform
AI CTF PHDFest2 2025AI CTF competition from PHDFest2 2025
AI in SecurityRussian platform for AI security training
DeepLearning.AI Red Teaming CourseShort course on red teaming LLM applications
Learn Prompting: Offensive MeasuresGuide on offensive prompt engineering techniques
Application Security LLM TestingFree LLM security testing
Salt Security Blog: ChatGPT Extensions VulnerabilitiesArticle on security flaws in ChatGPT browser extensions
safeguarding-llmsTMLS 2024 Workshop: A Practitioner's Guide To Safeguarding Your LLM Applications
Damn Vulnerable LLM AgentIntentionally vulnerable LLM agent for security testing and education
GPT Agents ArenaPlatform for testing and evaluating LLM agents in various scenarios
AI BattleInteractive game focusing on AI security challenges
AI/LLM Exploitation ChallengesChallenges to test your knowledge of AI, ML, and LLMs
TryHackMe AI/ML Security ThreatsWalkthrough and writeup for TryHackMe AI/ML Security Threats room

image

📊 Community research articles

Research articles, security advisories, and technical papers from the security community.

TitleAuthorsYear
📄 Bypassing Meta's LLaMA Classifier: A Simple JailbreakRobust Intelligence2024
📄 Vulnerabilities in LangChain Gen AIUnit422024
📄 Detecting Prompt Injection: BERT-based ClassifierWithSecure Labs2024
📄 Practical LLM Security: Takeaways From a Year in the TrenchesNVIDIA2024
📄 Security ProbLLMs in xAI's GrokEmbrace The Red2024
📄 Persistent Pre-Training Poisoning of LLMsSpyLab AI2024
📄 Navigating the Risks: A Survey of Security, Privacy, and Ethics Threats in LLM-Based AgentsMultiple Authors2024
📄 Practical AI Agent SecurityMeta2025
📄 Security Advisory: Anthropic's Slack MCP Server Vulnerable to Data ExfiltrationEmbrace The Red2025

🎓 Tutorials

Step-by-step guides and tutorials for understanding and implementing LLM security practices.

ResourceDescription
📚 HADESS - Web LLM AttacksUnderstanding how to carry out web attacks using LLM
📚 Red Teaming with LLMsPractical methods for attacking AI systems
📚 Lakera LLM SecurityOverview of attacks on LLM

📚 Books

Comprehensive books covering LLM security, adversarial AI, and secure AI development practices.

📖 Title🖋️ Author(s)🔍 Description
The Developer's Playbook for Large Language Model SecuritySteve Wilson🛡️ Comprehensive guide for developers on securing LLMs
Generative AI Security: Theories and Practices (Future of Business and Finance)Ken Huang, Yang Wang, Ben Goertzel, Yale Li, Sean Wright, Jyoti Ponnapalli🔬 In-depth exploration of security theories, laws, terms and practices in Generative AI
Adversarial AI Attacks, Mitigations, and Defense Strategies: A cybersecurity professional's guide to AI attacks, threat modeling, and securing AI with MLSecOpsJohn SotiropoulosPractical examples of code for your best mlsecops pipeline

BLOGS

Security blogs, Twitter feeds, and Telegram channels focused on AI/LLM security.

Websites & Twitter

ResourceDescription
Embrace The RedBlog on AI security, red teaming, and LLM vulnerabilities
HiddenLayerAI security company blog
CyberArkBlog on AI agents, identity risks, and security
StraikerAI security research and agentic browser security
FiretailLLM security, prompt injection, and AI vulnerabilities
Palo Alto NetworksUnit 42 research on AI security and agentic AI attacks
Trail of BitsSecurity research including AI/ML pickle file security
NCSCUK National Cyber Security Centre blog on AI safeguards
KnosticAI Security Posture Management (AISPM)
0dinSecure LLM and RAG deployment practices
@llm_secTwitter feed on LLM security
@LLM_Top10Twitter feed on OWASP LLM Top 10
@aivillage_dcAI Village Twitter
@elder_pliniusTwitter feed on AI security

Telegram Channels

ChannelLanguageDescription
PWN AIRUPractical AI Security and MLSecOps: LLM security, agents, guardrails, real-world threats
Борис_ь с mlRUMachine Learning + Information Security: ML, data science and cyber/AI security
Евгений Кокуйкин — RaftRUBuilding Raft AI and GPT-based applications: trust & safety, reliability and security
LLM SecurityRUFocused on LLM security: jailbreaks, prompt injection, adversarial attacks, benchmarks
AISecHubENGlobal AI security hub: curated research, articles, reports and tools
AI Security LabRULaboratory by Raft x ITMO University: breaking and defending AI systems
ML&Sec FeedRU/ENAggregated feed for ML & security: news, tools, research links
AISec [x_feed]RU/ENDigest of AI security content from X, blogs and papers
AI SecOpsRUAI Security Operations: monitoring, incident response, SIEM/SOC integrations
OK MLRUML/DS/AI channel with focus on repositories, tools and vulnerabilities
AI AttacksENStream of AI attack examples and threat intelligence
AGI SecurityENArtificial General Intelligence Security discussions

DATA

Datasets for testing LLM security, prompt injection examples, and safety evaluation data.

ResourceDescription
Safety and privacy with Large Language ModelsGitHub repository on LLM safety and privacy
Jailbreak LLMsData for jailbreaking Large Language Models
ChatGPT System PromptRepository containing ChatGPT system prompts
Do Not AnswerProject related to LLM response control
ToxiGenMicrosoft dataset
SafetyPromptsA Living Catalogue of Open Datasets for LLM Safety
llm-security-prompt-injectionThis project investigates the security of large language models by performing binary classification of a set of input prompts to discover malicious prompts. Several approaches have been analyzed using classical ML algorithms, a trained LLM model, and a fine-tuned LLM model.
Prompt Injections DatasetDataset containing prompt injection examples for testing and research

OPS

Operational security considerations: supply chain risks, infrastructure vulnerabilities, and production deployment security.

Group 4

ResourceDescription
https://sysdig.com/blog/llmjacking-stolen-cloud-credentials-used-in-new-ai-attack/LLMJacking: Stolen Cloud Credentials Used in New AI Attack
https://huggingface.co/docs/hub/securityHugging Face Hub Security Documentation
https://github.com/ShenaoW/awesome-llm-supply-chain-securityLLM Supply chain security resources
https://developer.nvidia.com/blog/secure-llm-tokenizers-to-maintain-application-integrity/Secure LLM Tokenizers to Maintain Application Integrity
https://sightline.protectai.com/Sightline by ProtectAI (ProtectAI is now part of Palo Alto Networks)

Check vulnerabilities on:
• Nemo by Nvidia
• Deep Lake
• Fine-Tuner AI
• Snorkel AI
• Zen ML
• Lamini AI
• Comet
• Titan ML
• Deepset AI
• Valohai

For finding LLMops tools vulnerabilities
https://wearetyomsmnv.github.io/ml_supply_chain_map/My supply_chain map
ShadowMQ: How Code Reuse Spread Critical Vulnerabilities Across the AI EcosystemResearch on critical RCE vulnerabilities in AI inference servers (Meta Llama Stack, NVIDIA TensorRT-LLM, vLLM, SGLang, Modular) caused by unsafe ZeroMQ and pickle deserialization

🏗 Frameworks

Comprehensive security frameworks, standards, and governance models for LLM and AI security.


OWASP Top 10 for LLM Applications 2025 (v2.0)

Updated list including System Prompt Leakage, Vector and Embedding Weaknesses

OWASP Top 10 for Agentic Applications (2026 Edition)

First industry standard for autonomous AI agent risks (released Dec 2025)

OWASP AI Testing Guide v1

Open standard for testing AI system trustworthiness (Nov 2025)

GenAI Security Solutions Reference Guide

Vendor-neutral guide for GenAI security architecture (Q2-Q3 2025)

LLM AI Cybersecurity & Governance Checklist

Security and governance checklist

LLMSecOps Cybersecurity Solution Landscape

Solution landscape overview

All OWASP GenAI Resources: genai.owasp.org/resources/

LLMSECOPS, by OWASP

Group 12

Additional Security Frameworks

FrameworkOrganizationDescription
MCP Security GovernanceCloud Security AllianceGovernance framework for the Model Context Protocol ecosystem. Developing policies, standards, and assessment tools for secure MCP server deployment.
Databricks AI Security Framework (DASF) 2.0DatabricksActionable framework for managing AI security. Includes 62 security risks across three stages and 64 controls applicable to any data and AI platform.
Google Secure AI Framework (SAIF) 2.0GoogleSecure AI Framework focused on agents. Practitioner-focused framework for building powerful agents users can trust.
Snowflake AI Security FrameworkSnowflakeComprehensive framework for securing AI deployments on Snowflake platform.

AI Security Solutions Radar

2025 AI Security Solutions Radar

Source: 2025 AI Security Solutions Radar by RiskInsight-Wavestone


🌐 Community

Community resources, platforms, and collaborative spaces for LLM security practitioners.

PlatformDetails
OWASP SLACKChannels:
• #project-top10-for-llm
• #ml-risk-top5
• #project-ai-community
• #project-mlsec-top10
• #team-llm_ai-secgov
• #team-llm-redteam
• #team-llm-v2-brainstorm
Awesome LLM SecurityGitHub repository
Awesome AI Security TelegramCurated list of Telegram channels and chats on AI Security, AI/MLSecOps, LLM Security
LVE_ProjectOfficial website
Lakera AI Security resource hubGoogle Sheets document
llm-testing-findingsTemplates with recommendations, CWE and other
Arcanum Prompt Injection TaxonomyStructured taxonomy of prompt injection attacks categorizing attack intents, techniques, and evasions. Resource for security researchers, AI developers, and red teamers.

Benchmarks

Security benchmarks, evaluation frameworks, and standardized tests for assessing LLM security capabilities.

ResourceDescriptionStars
Backbone Breaker Benchmark (b3)Human-grounded benchmark for testing AI agent security. Built by Lakera with UK AI Security Institute using 194,000+ human attack attempts from Gandalf: Agent Breaker. Tests backbone LLM resilience across 10 threat snapshots.Article
Backbone Breaker Benchmark PaperResearch paper on the Backbone Breaker Benchmark methodology and findingsArticle
CyberSoCEvalMeta's benchmark for evaluating LLM capabilities in malware analysis and threat intelligence reasoningMeta Research
Agent Security Bench (ASB)Benchmark for agent securityGitHub stars
AI Safety BenchmarkComprehensive benchmark for AI safety evaluationN/A
AI Safety Benchmark PaperResearch paper on AI safety benchmarking methodologiesArticle
Evaluating Prompt Injection DatasetsAnalysis and evaluation framework for prompt injection datasetsHiddenLayer
LLM Security Guidance BenchmarksBenchmarking lightweight, open-source LLMs for security guidance effectiveness using SECURE datasetGitHub stars
SECUREBenchmark for evaluating LLMs in cybersecurity scenarios, focusing on Industrial Control SystemsGitHub stars
NIST AI TEVVAI Test, Evaluation, Validation and Verification framework by NISTN/A
Taming the Beast: Inside the Llama 3 Red Teaming ProcessDEF CON 32 presentation on Llama 3 red teaming2024

About

LLM | Security | Operations in one github repo with good links and pictures.

Language
HTML100%