owasp-llm-top10 by mastepanoski/claude-skills
npx skills add https://github.com/mastepanoski/claude-skills --skill owasp-llm-top10此技能使 AI 代理能够使用 OWASP GenAI 安全项目发布的 OWASP LLM 应用十大安全风险(2025),对大型语言模型(LLM)和生成式 AI 应用进行全面的安全评估。
OWASP LLM 应用十大安全风险识别了集成大型语言模型的系统中最关键的安全风险,涵盖了从提示注入到无限制资源消耗等各种漏洞。这是 LLM 应用安全领域的权威行业标准。
使用此技能来识别安全漏洞、评估风险敞口、确定修复优先级,并为 AI 驱动的应用程序建立安全的开发实践。
可结合 "NIST AI RMF" 进行全面的风险管理,或结合 "ISO 42001 AI Governance" 以满足治理合规性要求。
在以下情况下调用此技能:
执行此审计时,请收集:
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
严重性 : 严重
描述 : 攻击者通过精心构造的输入(直接或间接)操纵 LLM 操作,以绕过预期功能、访问未经授权的数据或触发意外操作。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 严重
描述 : LLM 无意中通过其输出暴露机密数据,包括 PII、专有算法、凭证、知识产权或内部系统信息。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 高
描述 : 受损的第三方组件(模型、数据集、库、插件)引入了安全风险,包括恶意软件、后门或有偏见的行为。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 高
描述 : 攻击者操纵训练或微调数据,以引入漏洞、后门或偏见,从而损害模型的安全性和可靠性。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 高
描述 : 应用程序盲目执行或渲染 LLM 输出而不进行验证,从而可能导致代码注入、XSS、SQL 注入、SSRF 和其他攻击。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 高
描述 : AI 代理拥有过多的权限和自主能力,可能通过受损的提示、幻觉或恶意操纵造成重大危害。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 中
描述 : 用于指导 AI 行为的系统指令暴露给用户或攻击者,从而泄露内部逻辑、安全控制或敏感配置。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 中
描述 : 向量数据库和基于嵌入的检索系统(RAG)中的漏洞允许对存储的数据进行投毒、注入或未经授权的访问。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 中
描述 : LLM 生成看似合理但虚假的信息(幻觉/虚构),用户可能信任并据此采取行动,从而造成危害。
攻击向量:
影响:
评估清单:
缓解策略:
严重性 : 中
描述 : 不受控制的 LLM 使用导致拒绝服务、系统崩溃或通过资源耗尽产生过高的运营成本。
攻击向量:
影响:
评估清单:
缓解策略:
针对 10 个漏洞中的每一个进行评估:
对发现的每个漏洞,使用以下标准进行评分:
可能性 : 被利用的可能性有多大?
影响 : 潜在的损害是什么?
编制全面的安全评估报告。
生成全面的 OWASP LLM 安全审计报告:
# OWASP LLM 十大安全审计报告
**应用程序**: [名称]
**LLM 提供商/模型**: [提供商 - 模型]
**日期**: [日期]
**评估者**: [AI 代理或人类]
**OWASP LLM 十大版本**: 2025
---
## 执行摘要
### 整体安全态势: [严重 / 高风险 / 中等风险 / 低风险 / 安全]
**应用程序类型**: [聊天机器人 / 代理 / RAG 系统 / 内容生成器 / 代码助手 / 其他]
**数据敏感性**: [公开 / 内部 / 机密 / 受限]
**用户群**: [内部 / B2B / B2C / 公开]
### 关键发现
| # | 漏洞 | 严重性 | 状态 |
|---|---|---|---|
| LLM01 | 提示注入 | 严重 | [易受攻击 / 已缓解 / 不适用] |
| LLM02 | 敏感信息泄露 | 严重 | [易受攻击 / 已缓解 / 不适用] |
| LLM03 | 供应链 | 高 | [易受攻击 / 已缓解 / 不适用] |
| LLM04 | 数据/模型投毒 | 高 | [易受攻击 / 已缓解 / 不适用] |
| LLM05 | 不当的输出处理 | 高 | [易受攻击 / 已缓解 / 不适用] |
| LLM06 | 过度代理 | 高 | [易受攻击 / 已缓解 / 不适用] |
| LLM07 | 系统提示泄露 | 中 | [易受攻击 / 已缓解 / 不适用] |
| LLM08 | 向量/嵌入弱点 | 中 | [易受攻击 / 已缓解 / 不适用] |
| LLM09 | 虚假信息 | 中 | [易受攻击 / 已缓解 / 不适用] |
| LLM10 | 无限制消耗 | 中 | [易受攻击 / 已缓解 / 不适用] |
### 前三大关键问题
1. [问题] - [影响描述]
2. [问题] - [影响描述]
3. [问题] - [影响描述]
---
## 详细发现
### LLM01: 提示注入
**状态**: [易受攻击 / 部分缓解 / 已缓解]
**严重性**: [严重 / 高 / 中 / 低]
**可能性**: [高 / 中 / 低]
**发现:**
1. [附带证据的发现]
2. [附带证据的发现]
**攻击场景:**
[描述如何利用此漏洞]
**建议:**
1. [具体的修复步骤]
2. [具体的修复步骤]
**工作量**: [低 / 中 / 高]
---
[继续 LLM02 到 LLM10...]
---
## 架构安全审查
### 数据流分析
[带有信任边界标记的数据流图或描述]
### 攻击面摘要
| 攻击面 | 风险级别 | 控制措施 |
|---|---|---|
| 用户输入 | [级别] | [控制措施] |
| API 端点 | [级别] | [控制措施] |
| 向量存储 | [级别] | [控制措施] |
| 插件/工具 | [级别] | [控制措施] |
| 输出渲染 | [级别] | [控制措施] |
---
## 修复路线图
### 阶段 1: 关键 (0-7 天)
1. [ ] [带负责人的行动项]
2. [ ] [带负责人的行动项]
### 阶段 2: 高优先级 (7-30 天)
1. [ ] [带负责人的行动项]
### 阶段 3: 中等优先级 (30-90 天)
1. [ ] [带负责人的行动项]
### 阶段 4: 加固 (持续进行)
1. [ ] [持续改进实践]
---
## 安全控制矩阵
| 控制措施 | 已实施 | 有效 | 建议 |
|---|---|---|---|
| 输入验证 | [是/否/部分] | [是/否] | [建议] |
| 输出清理 | [是/否/部分] | [是/否] | [建议] |
| 速率限制 | [是/否/部分] | [是/否] | [建议] |
| 身份验证 | [是/否/部分] | [是/否] | [建议] |
| 授权 | [是/否/部分] | [是/否] | [建议] |
| 日志记录/监控 | [是/否/部分] | [是/否] | [建议] |
| 内容过滤 | [是/否/部分] | [是/否] | [建议] |
| 人类参与 | [是/否/部分] | [是/否] | [建议] |
---
## 后续步骤
1. [ ] 确定关键发现的优先级并分配任务
2. [ ] 实施快速见效的措施(输入验证、速率限制)
3. [ ] 为高风险区域安排渗透测试
4. [ ] 建立持续监控
5. [ ] 修复后计划后续审计
---
## 资源
- [OWASP LLM 应用十大安全风险 2025](https://genai.owasp.org/resource/owasp-top-10-for-llm-applications-2025/)
- [OWASP GenAI 安全项目](https://genai.owasp.org/)
- [OWASP LLM AI 安全与治理清单](https://owasp.org/www-project-top-10-for-large-language-model-applications/)
- [OWASP GitHub 存储库](https://github.com/OWASP/www-project-top-10-for-large-language-model-applications)
---
**审计版本**: 1.0
**日期**: [日期]
| 优先级 | 漏洞 | 理由 |
|---|---|---|
| P0 | LLM01 (提示注入), LLM02 (数据泄露) | 直接利用,影响大 |
| P1 | LLM05 (输出处理), LLM06 (过度代理) | 可能导致系统被入侵 |
| P2 | LLM03 (供应链), LLM04 (投毒) | 较难利用但影响严重 |
| P3 | LLM07 (提示泄露), LLM08 (向量弱点) | 为后续攻击提供便利 |
| P4 | LLM09 (虚假信息), LLM10 (无限制消耗) | 操作风险 |
1.0 - 初始版本 (OWASP LLM 应用十大安全风险 2025)
请记住 : LLM 安全是一个不断发展的领域。新的攻击向量会定期出现。此审计提供基线评估;持续监控和定期重新评估对于维持安全态势至关重要。
每周安装数
87
仓库
GitHub 星标数
13
首次出现
2026年2月5日
安全审计
安装于
codex84
gemini-cli83
github-copilot83
opencode83
cursor80
kimi-cli79
This skill enables AI agents to perform a comprehensive security assessment of Large Language Model (LLM) and Generative AI applications using the OWASP Top 10 for LLM Applications 2025 , published by the OWASP GenAI Security Project.
The OWASP Top 10 for LLM Applications identifies the most critical security risks in systems that integrate large language models, covering vulnerabilities from prompt injection to unbounded resource consumption. This is the authoritative industry standard for LLM application security.
Use this skill to identify security vulnerabilities, assess risk exposure, prioritize remediation, and establish secure development practices for AI-powered applications.
Combine with "NIST AI RMF" for comprehensive risk management or "ISO 42001 AI Governance" for governance compliance.
Invoke this skill when:
When executing this audit, gather:
Severity : Critical
Description : Attackers manipulate LLM operations through crafted inputs, either directly or indirectly, to bypass intended functionality, access unauthorized data, or trigger unintended actions.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : Critical
Description : LLMs inadvertently expose confidential data including PII, proprietary algorithms, credentials, intellectual property, or internal system information through their outputs.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : High
Description : Compromised third-party components (models, datasets, libraries, plugins) introduce security risks including malware, backdoors, or biased behavior.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : High
Description : Attackers manipulate training or fine-tuning data to introduce vulnerabilities, backdoors, or biases that compromise model security and reliability.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : High
Description : Applications blindly execute or render LLM outputs without validation, enabling code injection, XSS, SQL injection, SSRF, and other attacks.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : High
Description : AI agents possess excessive permissions and autonomous capabilities, enabling significant harm through compromised prompts, hallucinations, or malicious manipulation.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : Medium
Description : System instructions intended to guide AI behavior are exposed to users or attackers, revealing internal logic, security controls, or sensitive configurations.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : Medium
Description : Vulnerabilities in vector databases and embedding-based retrieval systems (RAG) allow poisoning, injection, or unauthorized access to stored data.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : Medium
Description : LLMs generate plausible but false information (hallucinations/confabulations) that users may trust and act upon, causing harm.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity : Medium
Description : Uncontrolled LLM usage causes denial-of-service, system crashes, or excessive operational costs through resource exhaustion.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
System inventory:
Threat modeling:
For each of the 10 vulnerabilities, assess:
For each vulnerability found, score using:
Likelihood : How likely is exploitation?
Impact : What is the potential damage?
Compile comprehensive security assessment.
Generate a comprehensive OWASP LLM security audit report:
# OWASP LLM Top 10 Security Audit Report
**Application**: [Name]
**LLM Provider/Model**: [Provider - Model]
**Date**: [Date]
**Evaluator**: [AI Agent or Human]
**OWASP LLM Top 10 Version**: 2025
---
## Executive Summary
### Overall Security Posture: [Critical / High Risk / Medium Risk / Low Risk / Secure]
**Application Type**: [Chatbot / Agent / RAG System / Content Generator / Code Assistant / Other]
**Data Sensitivity**: [Public / Internal / Confidential / Restricted]
**User Base**: [Internal / B2B / B2C / Public]
### Critical Findings
| # | Vulnerability | Severity | Status |
|---|---|---|---|
| LLM01 | Prompt Injection | Critical | [Vulnerable / Mitigated / N/A] |
| LLM02 | Sensitive Info Disclosure | Critical | [Vulnerable / Mitigated / N/A] |
| LLM03 | Supply Chain | High | [Vulnerable / Mitigated / N/A] |
| LLM04 | Data/Model Poisoning | High | [Vulnerable / Mitigated / N/A] |
| LLM05 | Improper Output Handling | High | [Vulnerable / Mitigated / N/A] |
| LLM06 | Excessive Agency | High | [Vulnerable / Mitigated / N/A] |
| LLM07 | System Prompt Leakage | Medium | [Vulnerable / Mitigated / N/A] |
| LLM08 | Vector/Embedding Weaknesses | Medium | [Vulnerable / Mitigated / N/A] |
| LLM09 | Misinformation | Medium | [Vulnerable / Mitigated / N/A] |
| LLM10 | Unbounded Consumption | Medium | [Vulnerable / Mitigated / N/A] |
### Top 3 Critical Issues
1. [Issue] - [Impact description]
2. [Issue] - [Impact description]
3. [Issue] - [Impact description]
---
## Detailed Findings
### LLM01: Prompt Injection
**Status**: [Vulnerable / Partially Mitigated / Mitigated]
**Severity**: [Critical / High / Medium / Low]
**Likelihood**: [High / Medium / Low]
**Findings:**
1. [Finding with evidence]
2. [Finding with evidence]
**Attack Scenario:**
[Description of how this could be exploited]
**Recommendations:**
1. [Specific remediation step]
2. [Specific remediation step]
**Effort**: [Low / Medium / High]
---
[Continue for LLM02 through LLM10...]
---
## Architecture Security Review
### Data Flow Analysis
[Diagram or description of data flows with trust boundaries marked]
### Attack Surface Summary
| Surface | Risk Level | Controls |
|---|---|---|
| User Input | [Level] | [Controls] |
| API Endpoints | [Level] | [Controls] |
| Vector Store | [Level] | [Controls] |
| Plugins/Tools | [Level] | [Controls] |
| Output Rendering | [Level] | [Controls] |
---
## Remediation Roadmap
### Phase 1: Critical (0-7 days)
1. [ ] [Action item with owner]
2. [ ] [Action item with owner]
### Phase 2: High Priority (7-30 days)
1. [ ] [Action item with owner]
### Phase 3: Medium Priority (30-90 days)
1. [ ] [Action item with owner]
### Phase 4: Hardening (Ongoing)
1. [ ] [Continuous improvement practices]
---
## Security Controls Matrix
| Control | Implemented | Effective | Recommendation |
|---|---|---|---|
| Input validation | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Output sanitization | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Rate limiting | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Authentication | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Authorization | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Logging/Monitoring | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Content filtering | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Human-in-the-loop | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
---
## Next Steps
1. [ ] Prioritize and assign critical findings
2. [ ] Implement quick wins (input validation, rate limiting)
3. [ ] Schedule penetration testing for high-risk areas
4. [ ] Establish continuous monitoring
5. [ ] Plan follow-up audit after remediation
---
## Resources
- [OWASP Top 10 for LLM Applications 2025](https://genai.owasp.org/resource/owasp-top-10-for-llm-applications-2025/)
- [OWASP GenAI Security Project](https://genai.owasp.org/)
- [OWASP LLM AI Security & Governance Checklist](https://owasp.org/www-project-top-10-for-large-language-model-applications/)
- [OWASP GitHub Repository](https://github.com/OWASP/www-project-top-10-for-large-language-model-applications)
---
**Audit Version**: 1.0
**Date**: [Date]
| Priority | Vulnerabilities | Rationale |
|---|---|---|
| P0 | LLM01 (Prompt Injection), LLM02 (Data Disclosure) | Direct exploitation, high impact |
| P1 | LLM05 (Output Handling), LLM06 (Excessive Agency) | System compromise potential |
| P2 | LLM03 (Supply Chain), LLM04 (Poisoning) | Harder to exploit but severe impact |
| P3 | LLM07 (Prompt Leakage), LLM08 (Vector Weaknesses) | Enables further attacks |
| P4 | LLM09 (Misinformation), LLM10 (Unbounded Consumption) | Operational risk |
1.0 - Initial release (OWASP Top 10 for LLM Applications 2025)
Remember : LLM security is an evolving field. New attack vectors emerge regularly. This audit provides a baseline assessment; continuous monitoring and periodic re-assessment are essential for maintaining security posture.
Weekly Installs
87
Repository
GitHub Stars
13
First Seen
Feb 5, 2026
Security Audits
Gen Agent Trust HubPassSocketPassSnykPass
Installed on
codex84
gemini-cli83
github-copilot83
opencode83
cursor80
kimi-cli79
超能力技能使用指南:AI助手技能调用优先级与工作流程详解
52,100 周安装
Ruby on Rails 应用开发指南:构建功能全面的Rails应用,包含模型、控制器、身份验证与最佳实践
298 周安装
Reddit 只读浏览技能 - 安全获取帖子、评论与搜索,助力智能体开发
297 周安装
原生广告投放指南:Taboola/Outbrain平台优化策略、创意测试与预算建议
313 周安装
自主智能体架构指南:从ReAct模式到可靠部署的最佳实践与风险规避
296 周安装
GitHub API 集成与自动化操作指南 - 使用 Membrane 管理仓库、议题和拉取请求
72 周安装
AI电影剧本创作器 - 专业剧本生成与角色一致性工具,助力AI视频制作
319 周安装