brainstorming-research-ideas by orchestra-research/ai-research-skills
npx skills add https://github.com/orchestra-research/ai-research-skills --skill brainstorming-research-ideas用于发现下一个研究想法的结构化框架。本技能提供十个互补的构思视角,帮助研究者从模糊的好奇心转向具体、可论证的研究提案。每个框架针对不同的认知模式——可以单独使用,也可以组合使用以进行全面探索。
在以下情况请勿使用此技能:
scientific-skills:literature-review)研究想法源于两种不同的模式。了解你处于哪种模式可以避免一个常见失败:构建缺乏真实问题的解决方案,或追逐没有可行方法的问题。
问题先行(痛点 → 方法):
方案先行(新能力 → 应用):
工作流程:
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
自检:
每个研究问题都处于特定的抽象层次。刻意向上或向下移动阶梯,可以揭示在当前层次看不到的想法。
| 方向 | 行动 | 成果 |
|---|---|---|
| 向上移动(泛化) | 将具体结果转化为更广泛的原则 | 框架性论文、理论贡献 |
| 向下移动(实例化) | 在具体约束下测试通用范式 | 实证性论文、令人惊讶的失败分析 |
| 横向移动(类比) | 将相同抽象层次应用于相邻领域 | 交叉融合、迁移性论文 |
工作流程:
示例:
突破通常来自解决被广泛接受但看似冲突的目标之间的张力。这些矛盾不是缺陷——它们是研究机会。
常见的研究张力:
| 张力对 | 研究机会 |
|---|---|
| 性能 ↔ 效率 | 我们能否用 10 倍少的计算量达到 SOTA? |
| 隐私 ↔ 效用 | 联邦/加密方法能否缩小准确性的差距? |
| 通用性 ↔ 专用性 | 微调何时能胜过提示,为什么? |
| 安全性 ↔ 能力 | 对齐能否提升而非削弱能力? |
| 可解释性 ↔ 性能 | 机制性洞察能否促成更好的架构? |
| 规模 ↔ 可访问性 | 小模型能否复制涌现行为? |
工作流程:
自检:
从其他学科借鉴结构性想法是最具生成性的研究启发法之一。许多基础技术就是这样出现的——注意力机制借鉴了认知科学,遗传算法借鉴了生物学,对抗训练借鉴了博弈论。
有效类比的要求:
对 ML 研究高产出的源领域:
| 源领域 | 可迁移概念 |
|---|---|
| 神经科学 | 注意力、记忆巩固、分层处理 |
| 物理学 | 基于能量的模型、相变、重整化 |
| 经济学 | 机制设计、拍卖理论、激励对齐 |
| 生态学 | 种群动态、生态位竞争、共同进化 |
| 语言学 | 组合性、语用学、语法归纳 |
| 控制理论 | 反馈循环、稳定性、自适应调节 |
工作流程:
强有力的想法通常来自在新条件下重新审视旧问题。硬件、规模、数据可用性或法规方面的进步可能使先前的假设失效,并使以前不切实际的方法变得可行。
需要监控的变化类别:
| 变化类型 | 示例 | 研究意义 |
|---|---|---|
| 计算 | GPU 速度提升 10 倍 | 曾被斥为过于昂贵的方法变得可行 |
| 规模 | 万亿令牌数据集 | 在小规模下失败的统计论证现在可能成立 |
| 法规 | 欧盟 AI 法案、GDPR | 创造了对合规替代方案的需求 |
| 工具 | 新框架、API | 降低了复杂方法的实现门槛 |
| 失败 | 高调的系统故障 | 暴露了现有方法的空白 |
| 文化 | 新的用户行为 | 改变了哪些问题最重要 |
工作流程:
理解一个方法在何处失效通常与展示它在何处有效同样有价值。边界探测系统地揭示了公认技术失效的条件。
需要探测的边界类型:
工作流程:
自检:
在接受复杂性之前,先问一个更简单的方法是否足够。领域有时会过度关注复杂的解决方案,而一个精简的基线表现却很有竞争力。
不必要复杂性的警告信号:
工作流程:
贡献框架:
从多个视角审视一个系统,可以揭示不同类别的研究问题。每个利益相关者看到不同的摩擦、风险和机会。
利益相关者视角:
| 利益相关者 | 关键问题 |
|---|---|
| 最终用户 | 这个可用吗?哪些错误是不可接受的?延迟容忍度是多少? |
| 开发者 | 这个可调试吗?维护负担是什么?如何组合? |
| 理论家 | 为什么这有效?形式化保证是什么?空白在哪里? |
| 对抗者 | 这如何被利用?攻击面是什么? |
| 伦理学家 | 谁受到伤害?嵌入了哪些偏见?谁被排除在外? |
| 监管者 | 这个可审计吗?决策可以解释吗?有问责制吗? |
| 运营者 | 成本是多少?如何扩展?故障模式是什么? |
工作流程:
新颖性通常来自重组或模块化。创新往往不在于新的原语,而在于组件如何排列或分离。
组合(结合现有技术):
分解(拆解整体系统):
工作流程:
一个强大的研究想法应该能用两句话向聪明的非专家解释清楚。这个测试强制要求目的清晰,并强化价值主张。
两句话模板:
第一句(问题):“[领域] 目前在 [具体问题] 方面存在困难,这很重要,因为 [具体后果]。” 第二句(洞察):“我们通过 [关键机制] 采用 [方法],这有效是因为 [原因]。”
如果你无法填写此模板:
校准问题:
使用这个端到端的工作流程,从空白页到排名研究想法。
目标:产生 10-20 个候选想法,无需过滤。
目标:缩小到 3-5 个最强的想法。
对每个候选想法应用以下过滤器:
| 过滤器 | 问题 | 淘汰标准 |
|---|---|---|
| 解释测试(F10) | 我能用两句话陈述这个吗? | 如果不能 → 想法尚不清晰 |
| 问题先行检查(F1) | 问题是真实且重要的吗? | 如果没有人因此受困扰 → 放弃 |
| 简洁性测试(F7) | 复杂性是否合理? | 如果更简单的方法有效 → 简化或放弃 |
| 利益相关者检查(F8) | 谁受益?谁可能反对? | 如果没有明确的受益者 → 放弃 |
| 可行性 | 我能否用可用资源执行这个? | 如果明显不可行 → 暂存以备后用 |
目标:将最佳想法转化为具体的研究计划。
完成清单:
不确定从哪个框架开始?使用此决策指南:
| 你的情况 | 从以下开始 |
|---|---|
| “我不知道该研究哪个领域” | 寻找张力(F3) → 什么改变了(F5) |
| “我有一个模糊的领域但没有具体想法” | 抽象阶梯(F2) → 失败分析(F6) |
| “我有一个想法但不确定它好不好” | 解释测试(F10) → 简洁性测试(F7) |
| “我有一个好想法但需要新角度” | 交叉融合(F4) → 利益相关者轮换(F8) |
| “我想将现有工作组合成新东西” | 组合/分解(F9) |
| “我找到了一个很酷的技术并想应用它” | 问题先行检查(F1) → 利益相关者轮换(F8) |
| “我想挑战传统智慧” | 失败分析(F6) → 简洁性测试(F7) |
| 陷阱 | 症状 | 解决方法 |
|---|---|---|
| 有新颖性但无影响力 | “没有人做过 X”,但没有人需要 X | 应用问题先行检查(F1) |
| 默认渐进式 | 想法在基准测试上提升 2% | 攀登抽象阶梯(F2) |
| 复杂性崇拜 | 方法有 8 个组件,每个贡献微乎其微 | 应用简洁性测试(F7) |
| 回音室效应 | 所有想法都来自阅读相同的 10 篇论文 | 使用交叉融合(F4) |
| 过时的假设 | “这以前试过,没成功”(5 年前) | 应用什么改变了(F5) |
| 单一视角偏见 | 只考虑 ML 工程师的观点 | 使用利益相关者轮换(F8) |
| 过早收敛 | 未探索替代方案就致力于第一个想法 | 运行完整发散阶段 |
当研究者请求帮助头脑风暴研究想法时:
关键原则:
每周安装次数
77
代码库
GitHub 星标数
5.6K
首次出现
Feb 23, 2026
安全审计
安装于
opencode72
gemini-cli70
github-copilot70
amp70
codex70
kimi-cli70
Structured frameworks for discovering the next research idea. This skill provides ten complementary ideation lenses that help researchers move from vague curiosity to concrete, defensible research proposals. Each framework targets a different cognitive mode—use them individually or combine them for comprehensive exploration.
Do NOT use this skill when :
scientific-skills:literature-review)Research ideas originate from two distinct modes. Knowing which mode you are in prevents a common failure: building solutions that lack real problems, or chasing problems without feasible approaches.
Problem-First (pain point → method):
Solution-First (new capability → application):
Workflow :
Self-Check :
Every research problem sits at a particular level of abstraction. Deliberately moving up or down the ladder reveals ideas invisible at your current level.
| Direction | Action | Outcome |
|---|---|---|
| Move Up (generalize) | Turn a specific result into a broader principle | Framework papers, theoretical contributions |
| Move Down (instantiate) | Test a general paradigm under concrete constraints | Empirical papers, surprising failure analyses |
| Move Sideways (analogize) | Apply same abstraction level to adjacent domain | Cross-pollination, transfer papers |
Workflow :
Example :
Breakthroughs often come from resolving tensions between widely accepted but seemingly conflicting goals. These contradictions are not bugs—they are the research opportunity.
Common Research Tensions :
| Tension Pair | Research Opportunity |
|---|---|
| Performance ↔ Efficiency | Can we match SOTA with 10x less compute? |
| Privacy ↔ Utility | Can federated/encrypted methods close the accuracy gap? |
| Generality ↔ Specialization | When does fine-tuning beat prompting, and why? |
| Safety ↔ Capability | Can alignment improve rather than tax capability? |
| Interpretability ↔ Performance | Do mechanistic insights enable better architectures? |
| Scale ↔ Accessibility | Can small models replicate emergent behaviors? |
Workflow :
Self-Check :
Borrowing structural ideas from other disciplines is one of the most generative research heuristics. Many foundational techniques emerged this way—attention mechanisms draw from cognitive science, genetic algorithms from biology, adversarial training from game theory.
Requirements for a Valid Analogy :
High-Yield Source Fields for ML Research :
| Source Field | Transferable Concepts |
|---|---|
| Neuroscience | Attention, memory consolidation, hierarchical processing |
| Physics | Energy-based models, phase transitions, renormalization |
| Economics | Mechanism design, auction theory, incentive alignment |
| Ecology | Population dynamics, niche competition, co-evolution |
| Linguistics | Compositionality, pragmatics, grammatical induction |
| Control Theory | Feedback loops, stability, adaptive regulation |
Workflow :
Strong ideas often come from revisiting old problems under new conditions. Advances in hardware, scale, data availability, or regulations can invalidate prior assumptions and make previously impractical approaches viable.
Categories of Change to Monitor :
| Change Type | Example | Research Implication |
|---|---|---|
| Compute | GPUs 10x faster | Methods dismissed as too expensive become feasible |
| Scale | Trillion-token datasets | Statistical arguments that failed at small scale may now hold |
| Regulation | EU AI Act, GDPR | Creates demand for compliant alternatives |
| Tooling | New frameworks, APIs | Reduces implementation barrier for complex methods |
| Failure | High-profile system failures | Exposes gaps in existing approaches |
| Cultural | New user behaviors | Shifts what problems matter most |
Workflow :
Understanding where a method breaks is often as valuable as showing where it works. Boundary probing systematically exposes the conditions under which accepted techniques fail.
Types of Boundaries to Probe :
Workflow :
Self-Check :
Before accepting complexity, ask whether a simpler approach suffices. Fields sometimes over-index on elaborate solutions when a streamlined baseline performs competitively.
Warning Signs of Unnecessary Complexity :
Workflow :
Contribution Framing :
Viewing a system from multiple perspectives reveals distinct classes of research questions. Each stakeholder sees different friction, risk, and opportunity.
Stakeholder Perspectives :
| Stakeholder | Key Questions |
|---|---|
| End User | Is this usable? What errors are unacceptable? What is the latency tolerance? |
| Developer | Is this debuggable? What is the maintenance burden? How does it compose? |
| Theorist | Why does this work? What are the formal guarantees? Where are the gaps? |
| Adversary | How can this be exploited? What are the attack surfaces? |
| Ethicist | Who is harmed? What biases are embedded? Who is excluded? |
| Regulator | Is this auditable? Can decisions be explained? Is there accountability? |
| Operator | What is the cost? How does it scale? What is the failure mode? |
Workflow :
Novelty often emerges from recombination or modularization. Innovation frequently lies not in new primitives, but in how components are arranged or separated.
Composition (combining existing techniques):
Decomposition (breaking apart monolithic systems):
Workflow :
A strong research idea should be defensible in two sentences to a smart non-expert. This test enforces clarity of purpose and sharpens the value proposition.
The Two-Sentence Template :
Sentence 1 (Problem): "[Domain] currently struggles with [specific problem], which matters because [concrete consequence]." Sentence 2 (Insight): "We [approach] by [key mechanism], which works because [reason]."
If You Cannot Fill This Template :
Calibration Questions :
Use this end-to-end workflow to go from blank page to ranked research ideas.
Goal : Produce 10-20 candidate ideas without filtering.
Goal : Narrow to 3-5 strongest ideas.
Apply these filters to each candidate:
| Filter | Question | Kill Criterion |
|---|---|---|
| Explain-It Test (F10) | Can I state this in two sentences? | If no → idea is not yet clear |
| Problem-First Check (F1) | Is the problem genuine and important? | If no one suffers from this → drop it |
| Simplicity Test (F7) | Is the complexity justified? | If a simpler approach works → simplify or drop |
| Stakeholder Check (F8) | Who benefits? Who might object? | If no clear beneficiary → drop it |
| Feasibility | Can I execute this with available resources? | If clearly infeasible → park it for later |
Goal : Turn the top idea into a concrete research plan.
Completion Checklist :
Not sure which framework to start with? Use this decision guide:
| Your Situation | Start With |
|---|---|
| "I don't know what area to work in" | Tension Hunting (F3) → What Changed (F5) |
| "I have a vague area but no specific idea" | Abstraction Ladder (F2) → Failure Analysis (F6) |
| "I have an idea but I'm not sure it's good" | Explain-It Test (F10) → Simplicity Test (F7) |
| "I have a good idea but need a fresh angle" | Cross-Pollination (F4) → Stakeholder Rotation (F8) |
| "I want to combine existing work into something new" | Composition/Decomposition (F9) |
| "I found a cool technique and want to apply it" | Problem-First Check (F1) → Stakeholder Rotation (F8) |
| "I want to challenge conventional wisdom" | Failure Analysis (F6) → Simplicity Test (F7) |
| Pitfall | Symptom | Fix |
|---|---|---|
| Novelty without impact | "No one has done X" but no one needs X | Apply Problem-First Check (F1) |
| Incremental by default | Idea is +2% on a benchmark | Climb the Abstraction Ladder (F2) |
| Complexity worship | Method has 8 components, each helping marginally | Apply Simplicity Test (F7) |
| Echo chamber | All ideas come from reading the same 10 papers | Use Cross-Pollination (F4) |
| Stale assumptions | "This was tried and didn't work" (5 years ago) | Apply What Changed (F5) |
| Single-perspective bias | Only considering the ML engineer's view | Use Stakeholder Rotation (F8) |
| Premature convergence | Committed to first idea without exploring alternatives |
When a researcher asks for help brainstorming research ideas:
Key Principles :
Weekly Installs
77
Repository
GitHub Stars
5.6K
First Seen
Feb 23, 2026
Security Audits
Gen Agent Trust HubPassSocketPassSnykPass
Installed on
opencode72
gemini-cli70
github-copilot70
amp70
codex70
kimi-cli70
AI 代码实施计划编写技能 | 自动化开发任务分解与 TDD 流程规划工具
50,900 周安装
Next.js App Router国际化方案:next-intl多语言路由配置指南
144 周安装
阿里云 AnalyticDB MySQL 数据库技能冒烟测试指南与自动化脚本
144 周安装
AI技能模板 - 高效构建AI助手技能,支持主流开发工具集成
145 周安装
iOS游戏开发终极指南:SpriteKit、SceneKit、RealityKit架构、性能优化与故障排除
142 周安装
Anime.js v4 中文指南:轻量级 JavaScript 动画库安装、核心概念与 API 详解
145 周安装
Claude Code 插件设置模式详解:YAML配置与本地状态管理
146 周安装
| Run full Diverge phase |