重要前提
安装AI Skills的关键前提是:必须科学上网,且开启TUN模式,这一点至关重要,直接决定安装能否顺利完成,在此郑重提醒三遍:科学上网,科学上网,科学上网。查看完整安装教程 →
npx skills add https://github.com/jwynia/agent-skills --skill fact-check对生成内容中的声明进行系统性验证。旨在捕捉幻觉、虚构和未经证实的断言。
根本问题: 大语言模型通过预测接下来应该出现的内容来生成听起来合理的内容。同样的机制也会产生幻觉——那些感觉真实但实际并非如此的自信陈述。处于生成模式的大语言模型无法可靠地捕捉自身的幻觉,因为:
解决方案: 验证必须是一个独立的认知过程,具备以下特点:
症状: 内容生成并交付,未进行任何事实核查。风险: 幻觉未被发现而通过。干预措施: 在交付前运行验证过程。提取声明,根据来源逐一检查。
症状: 在生成过程中,要求同一过程“检查你的事实”。风险: 虚假信心——由产生错误的同一过程确认错误。干预措施: 先完成生成,然后运行独立的验证过程,并明确要求提供来源。
症状: 根据“我所知道的”检查声明,未使用外部来源。风险: 幻觉被虚构的知识所验证。干预措施: 要求为每个已验证的声明提供明确的来源引用。如果没有可用来源,则标记为未经验证。
仅检查部分声明;其余声明被假定为正确。 未检查的声明可能包含错误。 系统地提取所有可验证的声明。逐一检查,或明确标记未检查的项目。
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
症状: 所有声明已提取,每个声明均已根据来源进行检查,并分配了置信度级别。指标: 存在来源引用,未经验证的声明已标记,置信度明确。
从内容中提取每一个可验证的陈述。
需要提取的声明类型:
需要跳过的内容:
根据可验证性对每个声明进行分类:
| 类别 | 描述 | 验证策略 |
|---|---|---|
| 可验证-硬事实 | 数字、日期、名称、引用 | 必须与来源完全匹配 |
| 可验证-软事实 | 一般事实、过程、机制 | 来源应基本支持 |
| 归属 | "X 说过..."、"根据..." | 验证来源存在且说过类似内容 |
| 推论 | 从证据得出的结论 | 验证前提,评估推理 |
| 观点陈述为事实 | 主观主张陈述为客观事实 | 标记以进行重新措辞或限定 |
对每个声明尝试进行验证:
## 声明验证日志
### 声明 1:"[确切的声明文本]"
- **类别:** [可验证-硬事实/软事实/归属/推论]
- **已检查的来源:** [具体来源]
- **发现:** [已确认/部分支持/未找到/被反驳]
- **置信度:** [高/中/低]
- **备注:** [差异、需要的限定条件]
### 声明 2:...
验证结果:
| 结果 | 含义 | 操作 |
|---|---|---|
| 已确认 | 来源明确支持声明 | 保留,引用来源 |
| 部分支持 | 来源支持部分而非全部 | 限定或缩小声明范围 |
| 未找到 | 未找到任何来源 | 标记为未经验证,考虑移除 |
| 被反驳 | 来源说法相反 | 移除或更正 |
| 已过时 | 来源已过时;当前状态可能不同 | 更新或添加时效性说明 |
为内容分配整体置信度:
| 级别 | 标准 |
|---|---|
| 高 | 所有关键声明已验证;未发现矛盾 |
| 中 | 大多数声明已验证;部分未验证但合理 |
| 低 | 重要声明未经验证;需要一些更正 |
| 不可靠 | 发现多处矛盾;需要重大修订 |
需要注意的常见幻觉类型:
模式: 听起来正确但不存在的具体细节。示例: 虚假的论文引用、不存在的统计数据、捏造的引语。检测方法: 根据原始来源验证具体声明。
模式: 合理的推断被陈述为既定事实。示例: "研究表明..."(无具体研究)、"专家同意..."(无引用)。检测方法: 要求为任何声称有外部支持的声明提供具体来源。
模式: 混合不同时期的信息。示例: 将旧统计数据呈现为当前数据、将已不存在的组织描述为活跃状态。检测方法: 检查来源日期,验证当前状态。
模式: 正确信息归属于错误的来源。示例: 引语归属于错误的人、发现归属于错误的研究。检测方法: 专门验证归属,而不仅仅是内容。
模式: 将多个来源的细节合并成一个虚构的来源。示例: 捏造一项研究,其中结合了来自不同论文的真实发现。检测方法: 验证具体来源存在且包含所有被归因的声明。
模式: 为模糊的知识添加虚假的精确度。示例: "大约 47.3%",而实际来源仅支持"大约一半"。检测方法: 检查来源是否确实提供了该级别的精确度。
在发布经过事实核查的内容之前:
| 研究阶段 | 事实核查角色 |
|---|---|
| 研究期间 | 验证来源本身的声明 |
| 综合之后 | 验证综合内容是否准确代表了来源 |
| 交付之前 | 最终检查以捕捉输出中的幻觉 |
交接模式:
| 情境 | 验证级别 |
|---|---|
| 已发布内容 | 需要完全验证 |
| 决策支持 | 关键声明必须经过验证 |
| 教育内容 | 期望高准确性 |
| 随意对话 | 可接受轻度验证 |
| 创意小说 | 不适用(不同标准) |
| 模式 | 问题 | 修正方法 |
|---|---|---|
| "我很自信" | 自信 ≠ 准确 | 要求提供来源引用 |
| "据我所知" | 记忆不可靠 | 检查外部来源 |
| "一般来说" | 模糊性掩盖了不确定性 | 具体说明或标记为未经验证 |
| "研究表明" | 哪项研究? | 引用具体来源 |
| 边生成边验证 | 同一过程无法捕捉自身错误 | 必须分开进行 |
| 检查一个,假设其余 | 部分验证 | 检查所有或标记未检查项 |
交付经过事实核查的内容时:
## [内容标题]
[包含声明的内容正文]
---
### 验证状态
**整体置信度:** [高/中/低]
**已验证的声明:**
- [声明 1] — 来源:[引用]
- [声明 2] — 来源:[引用]
**未经验证的声明:**
- [声明 3] — 未找到来源;请视为不确定
**已进行的更正:**
- [原始声明] → [更正后的声明] (来源:[引用])
**注意事项:**
- [任何限制或限定条件]
此技能将主要输出写入文件,以便工作在不同会话间持久保存。
在进行任何其他工作之前:
context/output-config.mdexplorations/fact-check/ 或适合此项目的合理位置context/output-config.md 中.fact-check-output.md 中对于此技能,持久化保存:
| 写入文件 | 保留在对话中 |
|---|---|
| 验证状态报告 | 关于来源的讨论 |
| 逐个声明的结果 | 澄清性问题 |
| 置信度评估 | 验证过程 |
| 更正和注意事项 | 实时反馈 |
模式:{内容名称}-factcheck-{日期}.md 示例:research-synthesis-factcheck-2025-01-15.md
此技能通过生成后验证扩展了研究集群。与研究(收集信息)不同,它作为输出质量控制运行。
相关:skills/research/SKILL.md(生成前)、references/doppelganger/(真相层级)
每周安装数
186
代码仓库
GitHub 星标数
37
首次出现
Jan 20, 2026
安全审计
安装于
opencode158
gemini-cli151
codex150
github-copilot140
cursor136
amp129
Systematic verification of claims in generated content. Designed to catch hallucinations, confabulations, and unsupported assertions.
The Fundamental Problem: LLMs generate plausible-sounding content by predicting what should come next. This same mechanism produces hallucinations—confident statements that feel true but aren't. An LLM in generation mode cannot reliably catch its own hallucinations because:
The Solution: Verification must be a separate cognitive pass with:
Symptoms: Content generated and delivered without any fact-checking. Risk: Hallucinations pass through undetected. Intervention: Run verification pass before delivery. Extract claims, check each against sources.
Symptoms: Same pass asked to "check your facts" while generating. Risk: False confidence—errors confirmed by same process that created them. Intervention: Complete generation first, then run separate verification pass with explicit source requirements.
Symptoms: Claims checked against "what I know" without external sources. Risk: Hallucinations verified by hallucinated knowledge. Intervention: Require explicit source citation for each verified claim. If no source available, mark as unverified.
Symptoms: Only some claims checked; others assumed correct. Risk: Unchecked claims may contain errors. Intervention: Systematic extraction of ALL verifiable claims. Check each, or explicitly mark unchecked items.
Symptoms: All claims extracted, each checked against sources, confidence levels assigned. Indicators: Source citations present, unverified claims marked, confidence explicit.
Extract every verifiable statement from the content.
Claim types to extract:
What to skip:
Categorize each claim by verifiability:
| Category | Description | Verification Strategy |
|---|---|---|
| Verifiable-Hard | Numbers, dates, names, quotes | Must match source exactly |
| Verifiable-Soft | General facts, processes, mechanisms | Source should substantially support |
| Attribution | "X said...", "According to..." | Verify source exists and said something similar |
| Inference | Conclusions drawn from evidence | Verify premises, assess reasoning |
| Opinion-as-Fact | Subjective claim stated as objective | Flag for rewording or qualification |
For each claim, attempt verification:
## Claim Verification Log
### Claim 1: "[exact claim text]"
- **Category:** [Verifiable-Hard/Soft/Attribution/Inference]
- **Source checked:** [specific source]
- **Finding:** [Confirmed/Partially supported/Not found/Contradicted]
- **Confidence:** [High/Medium/Low]
- **Notes:** [discrepancies, qualifications needed]
### Claim 2: ...
Verification outcomes:
| Outcome | Meaning | Action |
|---|---|---|
| Confirmed | Source explicitly supports claim | Keep, cite source |
| Partially supported | Source supports part, not all | Qualify or narrow claim |
| Not found | No source located | Mark unverified, consider removing |
| Contradicted | Source says opposite | Remove or correct |
| Outdated | Source is dated; current state may differ | Update or add recency caveat |
Assign overall confidence to the content:
| Level | Criteria |
|---|---|
| High | All key claims verified; no contradictions found |
| Medium | Most claims verified; some unverified but plausible |
| Low | Significant claims unverified; some corrections needed |
| Unreliable | Multiple contradictions found; major revision needed |
Common hallucination types to watch for:
Pattern: Specific details that sound right but don't exist. Examples: Fake paper citations, non-existent statistics, invented quotes. Detection: Verify specific claims against primary sources.
Pattern: Reasonable inference stated as established fact. Examples: "Studies show..." (no specific study), "Experts agree..." (no citation). Detection: Require specific source for any claim of external support.
Pattern: Mixing information from different time periods. Examples: Old statistics presented as current, defunct organizations described as active. Detection: Check dates on sources, verify current status.
Pattern: Correct information attributed to wrong source. Examples: Quote assigned to wrong person, finding attributed to wrong study. Detection: Verify attribution specifically, not just content.
Pattern: Combining details from multiple sources into one fictional source. Examples: Invented study that combines real findings from separate papers. Detection: Verify the specific source exists and contains all attributed claims.
Pattern: Adding false precision to vague knowledge. Examples: "Approximately 47.3%" when only "about half" is supported. Detection: Check if source actually provides that level of precision.
Before releasing fact-checked content:
| Research Phase | Fact-Check Role |
|---|---|
| During research | Verify claims in sources themselves |
| After synthesis | Verify that synthesis accurately represents sources |
| Before delivery | Final pass to catch hallucinations in output |
Handoff pattern:
| Context | Verification Level |
|---|---|
| Published content | Full verification required |
| Decision support | Key claims must be verified |
| Educational content | High accuracy expected |
| Casual conversation | Light verification acceptable |
| Creative fiction | N/A (different standards) |
| Pattern | Problem | Fix |
|---|---|---|
| "I'm confident" | Confidence ≠ accuracy | Require source citation |
| "To the best of my knowledge" | Memory is unreliable | Check external source |
| "Generally speaking" | Vagueness hides uncertainty | Be specific or mark unverified |
| "Research shows" | Which research? | Cite specific source |
| Verify-while-generating | Same pass can't catch own errors | Separate passes mandatory |
| Check one, assume rest | Partial verification | Check all or mark unchecked |
When delivering fact-checked content:
## [Content Title]
[Content body with claims]
---
### Verification Status
**Overall Confidence:** [High/Medium/Low]
**Verified Claims:**
- [Claim 1] — Source: [citation]
- [Claim 2] — Source: [citation]
**Unverified Claims:**
- [Claim 3] — No source found; treat as uncertain
**Corrections Made:**
- [Original claim] → [Corrected claim] (Source: [citation])
**Caveats:**
- [Any limitations or qualifications]
This skill writes primary output to files so work persists across sessions.
Before doing any other work:
context/output-config.md in the projectexplorations/fact-check/ or a sensible location for this projectcontext/output-config.md if context network exists.fact-check-output.md at project root otherwiseFor this skill, persist:
| Goes to File | Stays in Conversation |
|---|---|
| Verification status report | Discussion of sources |
| Claim-by-claim results | Clarifying questions |
| Confidence assessment | Verification process |
| Corrections and caveats | Real-time feedback |
Pattern: {content-name}-factcheck-{date}.md Example: research-synthesis-factcheck-2025-01-15.md
This skill extends the research cluster with post-generation verification. Distinct from research (which gathers information) and operates as quality control on output.
Related: skills/research/SKILL.md (pre-generation), references/doppelganger/ (truth hierarchies)
Weekly Installs
186
Repository
GitHub Stars
37
First Seen
Jan 20, 2026
Security Audits
Gen Agent Trust HubPassSocketPassSnykPass
Installed on
opencode158
gemini-cli151
codex150
github-copilot140
cursor136
amp129
超能力技能使用指南:AI助手技能调用优先级与工作流程详解
55,300 周安装