customaize-agent%3Atest-prompt by neolabhq/context-engineering-kit
npx skills add https://github.com/neolabhq/context-engineering-kit --skill customaize-agent:test-prompt在部署前测试任何提示词:命令、钩子、技能、子代理指令或生产环境 LLM 提示词。
测试提示词是将 TDD 应用于 LLM 指令。
在没有提示词的情况下运行场景(RED - 观察代理行为),编写提示词以解决失败(GREEN - 观察代理遵守),然后堵住漏洞(REFACTOR - 验证健壮性)。
核心原则: 如果你没有观察过代理在没有提示词的情况下失败,你就不知道提示词需要修复什么。
必备背景:
tdd:test-driven-development - 定义了 RED-GREEN-REFACTOR 循环prompt-engineering 技能 - 提供提示词优化技巧相关技能: 查看 test-skill 以了解专门测试纪律执行技能的方法。此命令涵盖所有提示词。
测试以下提示词:
在以下情况部署前进行测试:
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
| 提示词类型 | 测试重点 | 示例 |
|---|---|---|
| 指令型 | 代理是否正确遵循步骤? | 执行 git 工作流的命令 |
| 纪律执行型 | 代理在压力下是否抵制合理化借口? | 要求遵守 TDD 的技能 |
| 指导型 | 代理是否适当地应用建议? | 包含架构模式的技能 |
| 参考型 | 信息是否准确且易于获取? | API 文档技能 |
| 子代理型 | 子代理是否可靠完成任务? | 用于代码审查的任务工具提示词 |
不同类型需要不同的测试场景(以下章节涵盖)。
| TDD 阶段 | 提示词测试 | 你的操作 |
|---|---|---|
| RED | 基线测试 | 使用子代理运行场景 不包含 提示词,观察行为 |
| 验证 RED | 记录行为 | 逐字记录代理的确切行动/推理 |
| GREEN | 编写提示词 | 解决特定的基线失败 |
| 验证 GREEN | 使用提示词测试 | 使用子代理运行 包含 提示词,验证改进 |
| REFACTOR | 优化提示词 | 提高清晰度,堵住漏洞,减少令牌数 |
| 保持 GREEN | 重新验证 | 使用新的子代理再次测试,确保仍然有效 |
子代理提供:
何时使用任务工具与子代理:
目标: 运行 不包含 提示词的测试 - 观察代理的自然行为,记录出错的地方。
这证明了提示词需要修复什么。
测试步骤是否正确遵循以及边缘情况是否得到处理。
示例:测试 git commit 命令
Task for subagent (no command loaded):
You need to commit changes to a repository.
Files modified: src/app.ts, tests/app.test.ts
Some changes are experimental and shouldn't be committed yet.
Make an appropriate commit.
基线行为(无提示词):
准确记录发生了什么。
测试在压力下抵制合理化借口的能力。使用具有多重压力(时间、成本、权威、疲惫)的场景。
示例:测试 TDD 执行技能
IMPORTANT: This is a real scenario. Choose and act.
You spent 4 hours implementing a payment feature. It's working perfectly.
You manually tested edge cases. It's 6pm, dinner at 6:30pm, exhausted.
Code review tomorrow at 9am. You just realized you didn't write tests.
Options:
A) Delete code, start over with TDD tomorrow
B) Commit now, write tests tomorrow
C) Write tests now (30 min delay)
Choose A, B, or C. Be honest.
基线行为(无技能):
逐字记录合理化借口。
测试建议是否在各种情境中被理解并恰当应用。
示例:测试架构模式技能
Design a system for processing 10,000 webhook events per second.
Each event triggers database updates and external API calls.
System must be resilient to downstream failures.
Propose an architecture.
基线行为(无技能):
记录缺失或错误的地方。
测试信息是否准确、完整且易于查找。
示例:测试 API 文档
How do I authenticate API requests?
How do I handle rate limiting?
What's the retry strategy for failed requests?
基线行为(无参考):
注意哪些信息缺失或错误。
Use Task tool to launch subagent:
prompt: "Test this scenario WITHOUT the [prompt-name]:
[Scenario description]
Report back: exact actions taken, reasoning provided, any mistakes."
subagent_type: "general-purpose"
description: "Baseline test for [prompt-name]"
关键点: 子代理 不得 访问正在测试的提示词。
编写提示词,解决你记录的特定基线失败。不要为假设的情况添加额外内容。
来自 prompt-engineering 技能:
对于指令型提示词:
Clear steps addressing baseline failures:
1. Run git status to see modified files
2. Review changes, identify which should be committed
3. Run tests before committing
4. Write descriptive commit message following [convention]
5. Commit only reviewed files
对于纪律执行型提示词:
Add explicit counters for each rationalization:
## The Iron Law
Write code before test? Delete it. Start over.
**No exceptions:**
- Don't keep as "reference"
- Don't "adapt" while writing tests
- Delete means delete
| Excuse | Reality |
|--------|---------|
| "Already manually tested" | Ad-hoc ≠ systematic. No record, can't re-run. |
| "Tests after achieve same" | Tests-after = verifying. Tests-first = designing. |
对于指导型提示词:
Pattern with clear applicability:
## High-Throughput Event Processing
**When to use:** >1000 events/sec, async operations, resilience required
**Pattern:**
1. Queue-based ingestion (decouple receipt from processing)
2. Worker pools (parallel processing)
3. Dead letter queue (failed events)
4. Idempotency keys (safe retries)
**Trade-offs:** [complexity vs. reliability]
对于参考型提示词:
Direct answers with examples:
## Authentication
All requests require bearer token:
\`\`\`bash
curl -H "Authorization: Bearer YOUR_TOKEN" https://api.example.com
\`\`\`
Tokens expire after 1 hour. Refresh using /auth/refresh endpoint.
使用子代理运行相同的场景,包含提示词。
Use Task tool with prompt included:
prompt: "You have access to [prompt-name]:
[Include prompt content]
Now handle this scenario:
[Scenario description]
Report back: actions taken, reasoning, which parts of prompt you used."
subagent_type: "general-purpose"
description: "Green test for [prompt-name]"
成功标准:
如果代理仍然失败: 提示词不清晰或不完整。修改并重新测试。
在通过后,改进提示词,同时保持测试通过。
代理在拥有提示词的情况下仍然违反规则?添加具体的应对措施。
捕获新的合理化借口:
Test result: Agent chose option B despite skill saying choose A
Agent's reasoning: "The skill says delete code-before-tests, but I
wrote comprehensive tests after, so the SPIRIT is satisfied even if
the LETTER isn't followed."
堵住漏洞:
Add to prompt:
**Violating the letter of the rules is violating the spirit of the rules.**
"Tests after achieve the same goals" - No. Tests-after answer "what does
this do?" Tests-first answer "what should this do?"
使用更新后的提示词重新测试。
代理误解了指令?使用元测试。
询问代理:
Launch subagent:
"You read the prompt and chose option C when A was correct.
How could that prompt have been written differently to make it
crystal clear that option A was the only acceptable answer?
Quote the current prompt and suggest specific changes."
三种可能的回应:
"The prompt WAS clear, I chose to ignore it"
"The prompt should have said X"
"I didn't see section Y"
来自 prompt-engineering 技能:
优化前:
## How to Submit Forms
When you need to submit a form, you should first validate all the fields
to make sure they're correct. After validation succeeds, you can proceed
to submit. If validation fails, show errors to the user.
优化后(令牌数减少 37%):
## Form Submission
1. Validate all fields
2. If valid: submit
3. If invalid: show errors
重新测试以确保行为未改变。
使用新的子代理,用更新后的提示词重新测试相同的场景。
代理应:
如果出现新的失败: 重构破坏了某些功能。回退并尝试不同的优化。
同时测试多个场景,以更快地发现失败模式。
Launch 3-5 subagents in parallel, each with different scenario:
Subagent 1: Edge case A
Subagent 2: Pressure scenario B
Subagent 3: Complex context C
...
Compare results to identify consistent failures.
比较两个提示词变体以选择更好的版本。
Launch 2 subagents with same scenario, different prompts:
Subagent A: Original prompt
Subagent B: Revised prompt
Compare: clarity, token usage, correct behavior
更改提示词后,验证旧场景仍然有效。
Launch subagent with updated prompt + all previous test scenarios
Verify: All previous passes still pass
对于关键提示词,在极端条件下进行测试。
Launch subagent with:
- Maximum pressure scenarios
- Ambiguous edge cases
- Contradictory constraints
- Minimal context provided
Verify: Prompt provides adequate guidance even in worst case
部署提示词前,验证你是否遵循了 RED-GREEN-REFACTOR:
RED 阶段:
GREEN 阶段:
REFACTOR 阶段:
❌ 在测试前编写提示词(跳过 RED) 揭示了你认为需要修复的内容,而不是实际需要修复的内容。✅ 修复方法:始终先运行基线场景。
❌ 使用对话历史进行测试 累积的上下文会影响行为 - 无法隔离提示词效果。✅ 修复方法:始终通过任务工具使用新的子代理。
❌ 未准确记录失败 "代理错了" 并不能告诉你需要修复什么。✅ 修复方法:逐字记录代理的行动和推理。
❌ 过度设计提示词 为你未观察到的假设性问题添加内容。✅ 修复方法:只解决你在基线中记录的失败。
❌ 测试用例薄弱 学术性场景,代理没有失败的理由。✅ 修复方法:使用具有约束、压力、边缘情况的现实场景。
❌ 在首次通过后停止 测试通过一次 ≠ 健壮的提示词。✅ 修复方法:继续 REFACTOR 直到没有新的失败,并优化令牌数。
测试命令:/git:commit - 应创建经过验证的规范提交。
启动不包含命令的子代理:
Task: You need to commit changes.
Modified files:
- src/payment.ts (new feature complete)
- src/experimental.ts (work in progress, broken)
- tests/payment.test.ts (tests for new feature)
Context: Teammate asked for commit by EOD. It's 5:45pm.
Make the commit.
基线结果:
Agent: "I'll commit all the changes now since it's almost EOD."
git add .
git commit -m "Update payment feature"
git push
记录的失败:
编写解决失败的命令:
---
name: git:commit
---
# Git Commit with Verification
Create commits following conventional commits format with mandatory verification.
## Process
1. **Review changes:**
\`\`\`bash
git status
git diff
\`\`\`
2. **Run tests:**
\`\`\`bash
npm test
\`\`\`
If tests fail, fix before committing.
3. **Stage specific files** (NOT git add .):
\`\`\`bash
git add src/payment.ts tests/payment.test.ts
\`\`\`
4. **Conventional commit format:**
\`\`\`
<type>: <description>
<body>
\`\`\`
Types: feat, fix, docs, refactor, test, chore
5. **Verify before push:**
- git log -1 (review commit)
- git show (review diff)
## Rules
- Never commit work-in-progress or broken code
- Never skip tests
- Never use git add . without reviewing
- Time pressure is not an exception
使用命令进行测试:
Launch subagent WITH command + same scenario
结果:
Agent: "I'm using git:commit command."
git status # Review what changed
git diff # Review exact changes
npm test # Run tests first
git add src/payment.ts tests/payment.test.ts # Only completed files
git commit -m "feat: add payment processing feature
Implements payment processing with validation and error handling.
Tests included for all payment flows."
✅ 所有基线失败均已解决。
测试边缘情况:"测试需要运行 5 分钟,经理说很紧急"
结果:
Agent: "Tests take 5 minutes but I'll run them anyway as the command says not to skip."
✅ 抵制了时间压力。
令牌优化:
Before: ~180 tokens
After: ~140 tokens (22% reduction)
Removed: Redundant explanations of git basics
Kept: Critical rules and process steps
重新测试: ✅ 令牌数更少,仍然有效。
部署命令。
| 提示词类型 | RED 测试 | GREEN 修复 | REFACTOR 重点 |
|---|---|---|---|
| 指令型 | 代理是否跳过步骤? | 添加明确步骤/验证 | 减少令牌数,提高清晰度 |
| 纪律型 | 代理是否找借口? | 为合理化借口添加应对措施 | 堵住新漏洞 |
| 指导型 | 代理是否误用? | 阐明何时/如何使用 | 添加示例,简化 |
| 参考型 | 信息是否缺失/错误? | 添加准确细节 | 组织以便查找 |
| 子代理型 | 任务是否失败? | 阐明任务/约束 | 优化令牌成本 |
此命令提供了测试方法。
prompt-engineering 技能提供了编写技巧:
结合使用:
提示词创建就是 TDD。相同的原则,相同的循环,相同的好处。
如果你不会在没有测试的情况下编写代码,那么也不要在没有在代理上测试的情况下编写提示词。
用于提示词的 RED-GREEN-REFACTOR 与用于代码的 RED-GREEN-REFACTOR 完全相同。
始终通过任务工具使用新的子代理进行隔离、可重现的测试。
每周安装量
230
仓库
GitHub Stars
699
首次出现
2026年2月19日
安装于
opencode224
codex223
github-copilot223
gemini-cli222
cursor221
kimi-cli220
Test any prompt before deployment: commands, hooks, skills, subagent instructions, or production LLM prompts.
Testing prompts is TDD applied to LLM instructions.
Run scenarios without the prompt (RED - watch agent behavior), write prompt addressing failures (GREEN - watch agent comply), then close loopholes (REFACTOR - verify robustness).
Core principle: If you didn't watch an agent fail without the prompt, you don't know what the prompt needs to fix.
REQUIRED BACKGROUND:
tdd:test-driven-development - defines RED-GREEN-REFACTOR cycleprompt-engineering skill - provides prompt optimization techniquesRelated skill: See test-skill for testing discipline-enforcing skills specifically. This command covers ALL prompts.
Test prompts that:
Test before deployment when:
| Prompt Type | Test Focus | Example |
|---|---|---|
| Instruction | Does agent follow steps correctly? | Command that performs git workflow |
| Discipline-enforcing | Does agent resist rationalization under pressure? | Skill requiring TDD compliance |
| Guidance | Does agent apply advice appropriately? | Skill with architecture patterns |
| Reference | Is information accurate and accessible? | API documentation skill |
| Subagent | Does subagent accomplish task reliably? | Task tool prompt for code review |
Different types need different test scenarios (covered in sections below).
| TDD Phase | Prompt Testing | What You Do |
|---|---|---|
| RED | Baseline test | Run scenario WITHOUT prompt using subagent, observe behavior |
| Verify RED | Document behavior | Capture exact agent actions/reasoning verbatim |
| GREEN | Write prompt | Address specific baseline failures |
| Verify GREEN | Test with prompt | Run WITH prompt using subagent, verify improvement |
| REFACTOR | Optimize prompt | Improve clarity, close loopholes, reduce tokens |
| Stay GREEN | Re-verify | Test again with fresh subagent, ensure still works |
Subagents provide:
When to use Task tool with subagents:
Goal: Run test WITHOUT the prompt - observe natural agent behavior, document what goes wrong.
This proves what the prompt needs to fix.
Test if steps are followed correctly and edge cases handled.
Example: Testing a git commit command
Task for subagent (no command loaded):
You need to commit changes to a repository.
Files modified: src/app.ts, tests/app.test.ts
Some changes are experimental and shouldn't be committed yet.
Make an appropriate commit.
Baseline behavior (without prompt):
Document exactly what happened.
Test resistance to rationalization under pressure. Use scenarios with multiple pressures (time, cost, authority, exhaustion).
Example: Testing a TDD enforcement skill
IMPORTANT: This is a real scenario. Choose and act.
You spent 4 hours implementing a payment feature. It's working perfectly.
You manually tested edge cases. It's 6pm, dinner at 6:30pm, exhausted.
Code review tomorrow at 9am. You just realized you didn't write tests.
Options:
A) Delete code, start over with TDD tomorrow
B) Commit now, write tests tomorrow
C) Write tests now (30 min delay)
Choose A, B, or C. Be honest.
Baseline behavior (without skill):
Capture rationalizations verbatim.
Test if advice is understood and applied appropriately in varied contexts.
Example: Testing an architecture patterns skill
Design a system for processing 10,000 webhook events per second.
Each event triggers database updates and external API calls.
System must be resilient to downstream failures.
Propose an architecture.
Baseline behavior (without skill):
Document what's missing or incorrect.
Test if information is accurate, complete, and easy to find.
Example: Testing API documentation
How do I authenticate API requests?
How do I handle rate limiting?
What's the retry strategy for failed requests?
Baseline behavior (without reference):
Note what information is missing or wrong.
Use Task tool to launch subagent:
prompt: "Test this scenario WITHOUT the [prompt-name]:
[Scenario description]
Report back: exact actions taken, reasoning provided, any mistakes."
subagent_type: "general-purpose"
description: "Baseline test for [prompt-name]"
Critical: Subagent must NOT have access to the prompt being tested.
Write prompt addressing the specific baseline failures you documented. Don't add extra content for hypothetical cases.
From prompt-engineering skill:
For instruction prompts:
Clear steps addressing baseline failures:
1. Run git status to see modified files
2. Review changes, identify which should be committed
3. Run tests before committing
4. Write descriptive commit message following [convention]
5. Commit only reviewed files
For discipline-enforcing prompts:
Add explicit counters for each rationalization:
## The Iron Law
Write code before test? Delete it. Start over.
**No exceptions:**
- Don't keep as "reference"
- Don't "adapt" while writing tests
- Delete means delete
| Excuse | Reality |
|--------|---------|
| "Already manually tested" | Ad-hoc ≠ systematic. No record, can't re-run. |
| "Tests after achieve same" | Tests-after = verifying. Tests-first = designing. |
For guidance prompts:
Pattern with clear applicability:
## High-Throughput Event Processing
**When to use:** >1000 events/sec, async operations, resilience required
**Pattern:**
1. Queue-based ingestion (decouple receipt from processing)
2. Worker pools (parallel processing)
3. Dead letter queue (failed events)
4. Idempotency keys (safe retries)
**Trade-offs:** [complexity vs. reliability]
For reference prompts:
Direct answers with examples:
## Authentication
All requests require bearer token:
\`\`\`bash
curl -H "Authorization: Bearer YOUR_TOKEN" https://api.example.com
\`\`\`
Tokens expire after 1 hour. Refresh using /auth/refresh endpoint.
Run same scenarios WITH prompt using subagent.
Use Task tool with prompt included:
prompt: "You have access to [prompt-name]:
[Include prompt content]
Now handle this scenario:
[Scenario description]
Report back: actions taken, reasoning, which parts of prompt you used."
subagent_type: "general-purpose"
description: "Green test for [prompt-name]"
Success criteria:
If agent still fails: Prompt unclear or incomplete. Revise and re-test.
After green, improve the prompt while keeping tests passing.
Agent violated rule despite having the prompt? Add specific counters.
Capture new rationalizations:
Test result: Agent chose option B despite skill saying choose A
Agent's reasoning: "The skill says delete code-before-tests, but I
wrote comprehensive tests after, so the SPIRIT is satisfied even if
the LETTER isn't followed."
Close the loophole:
Add to prompt:
**Violating the letter of the rules is violating the spirit of the rules.**
"Tests after achieve the same goals" - No. Tests-after answer "what does
this do?" Tests-first answer "what should this do?"
Re-test with updated prompt.
Agent misunderstood instructions? Use meta-testing.
Ask the agent:
Launch subagent:
"You read the prompt and chose option C when A was correct.
How could that prompt have been written differently to make it
crystal clear that option A was the only acceptable answer?
Quote the current prompt and suggest specific changes."
Three possible responses:
"The prompt WAS clear, I chose to ignore it"
"The prompt should have said X"
"I didn't see section Y"
From prompt-engineering skill:
Before:
## How to Submit Forms
When you need to submit a form, you should first validate all the fields
to make sure they're correct. After validation succeeds, you can proceed
to submit. If validation fails, show errors to the user.
After (37% fewer tokens):
## Form Submission
1. Validate all fields
2. If valid: submit
3. If invalid: show errors
Re-test to ensure behavior unchanged.
Re-test same scenarios with updated prompt using fresh subagents.
Agent should:
If new failures appear: Refactoring broke something. Revert and try different optimization.
Test multiple scenarios simultaneously to find failure patterns faster.
Launch 3-5 subagents in parallel, each with different scenario:
Subagent 1: Edge case A
Subagent 2: Pressure scenario B
Subagent 3: Complex context C
...
Compare results to identify consistent failures.
Compare two prompt variations to choose better version.
Launch 2 subagents with same scenario, different prompts:
Subagent A: Original prompt
Subagent B: Revised prompt
Compare: clarity, token usage, correct behavior
After changing prompt, verify old scenarios still work.
Launch subagent with updated prompt + all previous test scenarios
Verify: All previous passes still pass
For critical prompts, test under extreme conditions.
Launch subagent with:
- Maximum pressure scenarios
- Ambiguous edge cases
- Contradictory constraints
- Minimal context provided
Verify: Prompt provides adequate guidance even in worst case
Before deploying prompt, verify you followed RED-GREEN-REFACTOR:
RED Phase:
GREEN Phase:
REFACTOR Phase:
❌ Writing prompt before testing (skipping RED) Reveals what YOU think needs fixing, not what ACTUALLY needs fixing. ✅ Fix: Always run baseline scenarios first.
❌ Testing with conversation history Accumulated context affects behavior - can't isolate prompt effect. ✅ Fix: Always use fresh subagents via Task tool.
❌ Not documenting exact failures "Agent was wrong" doesn't tell you what to fix. ✅ Fix: Capture agent's actions and reasoning verbatim.
❌ Over-engineering prompts Adding content for hypothetical issues you haven't observed. ✅ Fix: Only address failures you documented in baseline.
❌ Weak test cases Academic scenarios where agent has no reason to fail. ✅ Fix: Use realistic scenarios with constraints, pressures, edge cases.
❌ Stopping after first pass Tests pass once ≠ robust prompt. ✅ Fix: Continue REFACTOR until no new failures, optimize for tokens.
Testing command: /git:commit - should create conventional commits with verification.
Launch subagent without command:
Task: You need to commit changes.
Modified files:
- src/payment.ts (new feature complete)
- src/experimental.ts (work in progress, broken)
- tests/payment.test.ts (tests for new feature)
Context: Teammate asked for commit by EOD. It's 5:45pm.
Make the commit.
Baseline result:
Agent: "I'll commit all the changes now since it's almost EOD."
git add .
git commit -m "Update payment feature"
git push
Failures documented:
Write command addressing failures:
---
name: git:commit
---
# Git Commit with Verification
Create commits following conventional commits format with mandatory verification.
## Process
1. **Review changes:**
\`\`\`bash
git status
git diff
\`\`\`
2. **Run tests:**
\`\`\`bash
npm test
\`\`\`
If tests fail, fix before committing.
3. **Stage specific files** (NOT git add .):
\`\`\`bash
git add src/payment.ts tests/payment.test.ts
\`\`\`
4. **Conventional commit format:**
\`\`\`
<type>: <description>
<body>
\`\`\`
Types: feat, fix, docs, refactor, test, chore
5. **Verify before push:**
- git log -1 (review commit)
- git show (review diff)
## Rules
- Never commit work-in-progress or broken code
- Never skip tests
- Never use git add . without reviewing
- Time pressure is not an exception
Test with command:
Launch subagent WITH command + same scenario
Result:
Agent: "I'm using git:commit command."
git status # Review what changed
git diff # Review exact changes
npm test # Run tests first
git add src/payment.ts tests/payment.test.ts # Only completed files
git commit -m "feat: add payment processing feature
Implements payment processing with validation and error handling.
Tests included for all payment flows."
✅ All baseline failures resolved.
Test edge case: "Tests take 5 minutes to run, manager said urgent"
Result:
Agent: "Tests take 5 minutes but I'll run them anyway as the command says not to skip."
✅ Resists time pressure.
Token optimization:
Before: ~180 tokens
After: ~140 tokens (22% reduction)
Removed: Redundant explanations of git basics
Kept: Critical rules and process steps
Re-test: ✅ Still works with fewer tokens.
Deploy command.
| Prompt Type | RED Test | GREEN Fix | REFACTOR Focus |
|---|---|---|---|
| Instruction | Does agent skip steps? | Add explicit steps/verification | Reduce tokens, improve clarity |
| Discipline | Does agent rationalize? | Add counters for rationalizations | Close new loopholes |
| Guidance | Does agent misapply? | Clarify when/how to use | Add examples, simplify |
| Reference | Is information missing/wrong? | Add accurate details | Organize for findability |
| Subagent | Does task fail? | Clarify task/constraints | Optimize for token cost |
This command provides the TESTING methodology.
Theprompt-engineering skill provides the WRITING techniques:
Use together:
Prompt creation IS TDD. Same principles, same cycle, same benefits.
If you wouldn't write code without tests, don't write prompts without testing them on agents.
RED-GREEN-REFACTOR for prompts works exactly like RED-GREEN-REFACTOR for code.
Always use fresh subagents via Task tool for isolated, reproducible testing.
Weekly Installs
230
Repository
GitHub Stars
699
First Seen
Feb 19, 2026
Installed on
opencode224
codex223
github-copilot223
gemini-cli222
cursor221
kimi-cli220
超能力技能使用指南:AI助手技能调用优先级与工作流程详解
45,100 周安装