pol-probe-advisor by deanpeters/product-manager-skills
npx skills add https://github.com/deanpeters/product-manager-skills --skill pol-probe-advisor指导产品经理根据他们的假设、风险和可用资源,选择合适的概念验证(PoL)探针类型(共5种)。当您需要消除特定风险或测试一个狭窄的假设,但不确定使用哪种验证方法时,请使用此技能。这个互动技能确保您将最便宜的原型与最严酷的真相相匹配——而不是您最擅长构建的原型。
这不是用来决定是否应该验证的工具(您应该验证)。它是一个用于选择如何最有效验证的决策框架。
常见的失败模式: 产品经理根据工具熟悉度("我会用 Figma,所以我会设计一个原型")而不是学习目标来选择验证方法。结果:验证了错误的东西,错过了实际的风险。
解决方案: 从假设出发,逆向工作。提问:"我要消除的具体风险是什么?通往严酷真相的最廉价路径是什么?"
| 类型 | 核心问题 | 最适合 | 时间线 |
|---|---|---|---|
| 可行性检查 | "我们能构建这个吗?" | 技术未知、API 依赖、数据完整性 | 1-2 天 |
| 任务导向测试 | "用户能否无摩擦地完成这项工作?" | 关键 UI 时刻、字段标签、决策点 | 2-5 天 |
| 叙事原型 | "这个工作流程能获得利益相关者的认同吗?" |
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
| 讲故事、解释复杂流程、达成一致 |
| 1-3 天 |
| 合成数据模拟 | "我们能否在不冒生产风险的情况下建模?" | 边缘情况、未知的未知、统计建模 | 2-4 天 |
| 氛围编码 PoL 探针 | "这个解决方案能经受住真实用户的接触吗?" | 具有真实交互的工作流程/UX 验证 | 2-3 天 |
黄金法则: "使用能揭示最严酷真相的最便宜原型。"
✅ 在以下情况使用此技能:
❌ 不要在以下情况使用此技能:
problem-statement.md 或 problem-framing-canvas.md)使用 workshop-facilitation 作为此技能的默认交互协议。
它定义了:
其他(请说明))此文件定义了特定领域的评估内容。如果存在冲突,请遵循此文件的领域逻辑。
此互动技能使用自适应提问,根据您的上下文推荐正确的 PoL 探针类型。
代理询问:
让我们找出哪种 PoL 探针类型适合您的验证需求。首先,我需要一些上下文信息:
1. 您正在测试什么假设?(用一句话描述,或使用"如果[我们做 X]给[用户画像],那么[结果]"的格式)
2. 您试图消除什么具体风险? 例如:
3. 您的时间线如何?
4. 您有哪些可用资源? 例如:
代理综合用户输入并询问:
根据您的假设和风险,您真正想回答以下哪个核心问题?
提供 5 个选项(与探针类型对应):
用户响应: [选择一个数字,或者如果都不合适则描述]
根据用户选择,代理推荐匹配的探针类型:
→ 推荐探针:可行性检查
这是什么: 一个为期 1-2 天的快速探索和删除测试,旨在暴露技术风险。不是为了给人留下深刻印象——而是为了快速揭示阻碍因素。
方法:
时间线: 1-2 天
工具:
成功标准示例:
处置计划: 记录发现后删除所有探索代码。
下一步: 您希望我生成一个记录此可行性检查的 pol-probe 工件吗?
→ 推荐探针:任务导向测试
这是什么: 使用专门的测试工具验证关键时刻——字段标签、决策点、导航、流失区域。专注于可观察的任务完成度,而不是意见。
方法:
时间线: 2-5 天
工具:
成功标准示例:
处置计划: 归档会话录音,记录学习成果,删除测试原型。
下一步: 您希望我生成一个记录此任务导向测试的 pol-probe 工件吗?
→ 推荐探针:叙事原型
这是什么: 讲述故事,而不是测试界面。使用视频演练或幻灯片故事板来解释工作流程并衡量兴趣。这是"讲述而非测试"——您验证的是叙事,而不是 UI。
方法:
storyboard.md 组件技能)时间线: 1-3 天
工具:
成功标准示例:
处置计划: 归档视频,记录反馈,删除支持文件。
下一步: 您希望我生成一个记录此叙事原型的 pol-probe 工件吗?
→ 推荐探针:合成数据模拟
这是什么: 使用模拟用户、合成数据或提示逻辑测试来探索边缘情况和未知的未知,而无需触及生产环境。想想"风洞测试,比事后分析便宜"。
方法:
时间线: 2-4 天
工具:
成功标准示例:
处置计划: 删除合成数据,归档发现结果,记录边缘情况。
下一步: 您希望我生成一个记录此合成数据模拟的 pol-probe 工件吗?
→ 推荐探针:氛围编码 PoL 探针
这是什么: 一个弗兰肯软件堆栈(ChatGPT Canvas + Replit + Airtable),创造足够的假象让用户与半功能工作流程交互。不是生产级的——只是足够在 48 小时内捕捉 UX/工作流程信号。
⚠️ 警告: 这是风险最高的探针类型。它看起来足够真实,容易将势头误认为成熟度。仅在需要真实用户接触且其他方法无法满足时使用。
方法:
时间线: 2-3 天
堆栈示例:
成功标准示例:
处置计划: 用户会话后删除所有代码,归档 Loom 录音,记录学习成果。
下一步: 您希望我生成一个记录此氛围编码探针的 pol-probe 工件吗?
代理提供:
我建议对您的假设使用 [选定的探针类型]。您希望我:
pol-probe 工件(使用 pol-probe.md 模板),包含您的假设、成功标准、时间线和处置计划?用户选择选项,代理相应执行。
如果用户不确定或假设太宽泛:
代理询问:
您的假设似乎很宽泛,或者同时测试了多个东西。PoL 探针在狭窄且可证伪时效果最好。让我问几个澄清性问题:
澄清性问题:
然后,代理根据优化后的假设重新推荐探针类型。
用户上下文:
代理推理:
推荐: 使用 UsabilityHub 进行任务导向测试
用户上下文:
代理推理:
推荐: 使用提示工程快速探索进行可行性检查
结果: 错误率为 18%。决定不构建该功能。节省了 6 周的开发时间。
用户上下文:
代理推理:
推荐: 使用 Loom 演练进行叙事原型
为什么不是氛围编码探针: 您不需要用户与安全设置交互——您需要他们理解这个概念。叙事更便宜、更快。
用户上下文:
代理响应:
⚠️ 这是原型剧场,不是 PoL 探针。
您测试的是内部政治,而不是用户行为或技术可行性。PoL 探针消除的是产品风险,而不是组织风险。
推荐: 完全跳过原型。相反:
如果高管需要演示,请在用真实用户验证假设之后再构建它。
失败模式: "我会用 Figma,所以我会设计一个 UI 原型"(即使设计不是风险所在)。
后果: 验证了错误的东西;错过了实际的风险。
解决方法: 先回答核心问题,然后选择方法。如果您需要进行可行性检查但只懂设计工具,请与工程师合作 1 天。
失败模式: "我们就构建它,看看会发生什么。"
后果: 在发现测试了错误假设之前,已经进行了 2 周的开发。
解决方法: 提问:"什么是能揭示最严酷真相的最便宜原型?"通常它不是代码。
失败模式: 氛围编码探针"看起来真实",所以团队将其视为生产代码。
后果: 范围蔓延、技术债务、抵制处置。
解决方法: 在构建之前设定处置日期。氛围编码探针是有意设计的弗兰肯软件——庆祝其粗糙,学习后删除。
失败模式: "让我们在一个探针中测试工作流程、定价和 UI。"
后果: 结果模糊不清——您不知道是哪个变量导致了失败。
解决方法: 一个探针,一个假设。如果您有 3 个假设,就运行 3 个探针。
失败模式: "我们看到了就知道了。"
后果: 没有严酷的真相——只有意见和虚荣指标。
解决方法: 在构建之前写下成功标准。定义"通过"、"失败"和"学习"的阈值。
每周安装数
212
代码库
GitHub 星标数
1.5K
首次出现
2026年2月12日
安全审计
安装于
codex187
opencode185
gemini-cli181
github-copilot180
cursor178
kimi-cli177
Guide product managers through selecting the right Proof of Life (PoL) probe type (of 5 flavors) based on their hypothesis, risk, and available resources. Use this when you need to eliminate a specific risk or test a narrow hypothesis, but aren't sure which validation method to use. This interactive skill ensures you match the cheapest prototype to the harshest truth—not the prototype you're most comfortable building.
This is not a tool for deciding if you should validate (you should). It's a decision framework for choosing how to validate most effectively.
Common failure mode: PMs choose validation methods based on tooling comfort ("I know Figma, so I'll design a prototype") rather than learning goal. Result: validate the wrong thing, miss the actual risk.
Solution: Work backwards from the hypothesis. Ask: "What specific risk am I eliminating? What's the cheapest path to harsh truth?"
| Type | Core Question | Best For | Timeline |
|---|---|---|---|
| Feasibility Check | "Can we build this?" | Technical unknowns, API dependencies, data integrity | 1-2 days |
| Task-Focused Test | "Can users complete this job without friction?" | Critical UI moments, field labels, decision points | 2-5 days |
| Narrative Prototype | "Does this workflow earn stakeholder buy-in?" | Storytelling, explaining complex flows, alignment | 1-3 days |
| Synthetic Data Simulation | "Can we model this without production risk?" | Edge cases, unknown-unknowns, statistical modeling | 2-4 days |
| Vibe-Coded PoL Probe | "Will this solution survive real user contact?" | Workflow/UX validation with real interactions | 2-3 days |
Golden Rule: "Use the cheapest prototype that tells the harshest truth."
✅ Use this when:
❌ Don't use this when:
problem-statement.md or problem-framing-canvas.md first)Use workshop-facilitation as the default interaction protocol for this skill.
It defines:
Other (specify) when useful)This file defines the domain-specific assessment content. If there is a conflict, follow this file's domain logic.
This interactive skill uses adaptive questioning to recommend the right PoL probe type based on your context.
Agent asks:
Let's figure out which PoL probe type is right for your validation needs. First, I need some context:
1. What hypothesis are you testing? (Describe in one sentence, or use "If [we do X] for [persona], then [outcome]" format)
2. What specific risk are you trying to eliminate? Examples:
3. What's your timeline?
4. What resources do you have available? Examples:
Agent synthesizes user input and asks:
Based on your hypothesis and risk, which of these core questions are you really trying to answer?
Offer 5 options (aligned to probe types):
User response: [Select one number, or describe if none fit]
Based on user selection, agent recommends the matching probe type:
→ Recommended Probe: Feasibility Check
What it is: A 1-2 day spike-and-delete test to surface technical risk. Not meant to impress anyone—meant to reveal blockers fast.
Methods:
Timeline: 1-2 days
Tools:
Success Criteria Example:
Disposal Plan: Delete all spike code after documenting findings.
Next Step: Would you like me to generate a pol-probe artifact documenting this feasibility check?
→ Recommended Probe: Task-Focused Test
What it is: Validate critical moments—field labels, decision points, navigation, drop-off zones—using specialized testing tools. Focus on observable task completion , not opinions.
Methods:
Timeline: 2-5 days
Tools:
Success Criteria Example:
Disposal Plan: Archive session recordings, document learnings, delete test prototype.
Next Step: Would you like me to generate a pol-probe artifact documenting this task-focused test?
→ Recommended Probe: Narrative Prototype
What it is: Tell the story, don't test the interface. Use video walkthroughs or slideware storyboards to explain workflows and measure interest. This is "tell vs. test"—you're validating the narrative, not the UI.
Methods:
storyboard.md component skill)Timeline: 1-3 days
Tools:
Success Criteria Example:
Disposal Plan: Archive video, document feedback, delete supporting files.
Next Step: Would you like me to generate a pol-probe artifact documenting this narrative prototype?
→ Recommended Probe: Synthetic Data Simulation
What it is: Use simulated users, synthetic data, or prompt logic testing to explore edge cases and unknown-unknowns without touching production. Think "wind tunnel testing, cheaper than postmortem."
Methods:
Timeline: 2-4 days
Tools:
Success Criteria Example:
Disposal Plan: Delete synthetic data, archive findings, document edge cases.
Next Step: Would you like me to generate a pol-probe artifact documenting this synthetic data simulation?
→ Recommended Probe: Vibe-Coded PoL Probe
What it is: A Frankensoft stack (ChatGPT Canvas + Replit + Airtable) that creates just enough illusion for users to interact with a semi-functional workflow. Not production-grade—just enough to catch UX/workflow signals in 48 hours.
⚠️ Warning: This is the riskiest probe type. It looks real enough to confuse momentum with maturity. Use only when you need real user contact and other methods won't suffice.
Methods:
Timeline: 2-3 days
Stack Example:
Success Criteria Example:
Disposal Plan: Delete all code after user sessions, archive Loom recordings, document learnings.
Next Step: Would you like me to generate a pol-probe artifact documenting this vibe-coded probe?
Agent offers:
I recommend using [selected probe type] for your hypothesis. Would you like me to:
pol-probe artifact (using the pol-probe.md template) with your hypothesis, success criteria, timeline, and disposal plan?User selects option, and agent executes accordingly.
If user is uncertain or hypothesis is too broad:
Agent asks:
Your hypothesis seems broad or tests multiple things at once. PoL probes work best when they're narrow and falsifiable. Let me ask a few clarifying questions:
Clarifying Questions:
Agent then re-recommends probe type based on refined hypothesis.
User Context:
Agent Reasoning:
Recommendation: Task-Focused Test using UsabilityHub
User Context:
Agent Reasoning:
Recommendation: Feasibility Check using prompt engineering spike
Outcome: Error rate was 18%. Decided NOT to build feature. Saved 6 weeks of development.
User Context:
Agent Reasoning:
Recommendation: Narrative Prototype using Loom walkthrough
Why NOT Vibe-Coded Probe: You don't need users to interact with security settings—you need them to understand the concept. Narrative is cheaper and faster.
User Context:
Agent Response:
⚠️ This is prototype theater, not a PoL probe.
You're testing internal politics, not user behavior or technical feasibility. PoL probes eliminate product risk, not organizational risk.
Recommendation: Skip the prototype entirely. Instead:
If executives need a demo, build it after you've validated the hypothesis with real users.
Failure Mode: "I know Figma, so I'll design a UI prototype" (even if design isn't the risk).
Consequence: Validate the wrong thing; miss the actual risk.
Fix: Answer the core question first , then pick the method. If you need a Feasibility Check but only know design tools, pair with an engineer for 1 day.
Failure Mode: "Let's just build it and see what happens."
Consequence: 2 weeks of development before learning you tested the wrong hypothesis.
Fix: Ask: "What's the cheapest prototype that tells the harshest truth?" Usually it's NOT code.
Failure Mode: Vibe-Coded probe "looks real," so team treats it like production code.
Consequence: Scope creep, technical debt, resistance to disposal.
Fix: Set disposal date before building. Vibe-Coded probes are Frankensoft by design —celebrate the jank, delete after learning.
Failure Mode: "Let's test the workflow, the pricing, and the UI in one probe."
Consequence: Ambiguous results—you won't know which variable caused failure.
Fix: One probe, one hypothesis. If you have 3 hypotheses, run 3 probes.
Failure Mode: "We'll know it when we see it."
Consequence: No harsh truth—just opinions and vanity metrics.
Fix: Write success criteria before building. Define "pass," "fail," and "learn" thresholds.
Weekly Installs
212
Repository
GitHub Stars
1.5K
First Seen
Feb 12, 2026
Security Audits
Gen Agent Trust HubPassSocketFailSnykWarn
Installed on
codex187
opencode185
gemini-cli181
github-copilot180
cursor178
kimi-cli177
Vue 3 调试指南:解决响应式、计算属性与监听器常见错误
9,900 周安装