重要前提
安装AI Skills的关键前提是:必须科学上网,且开启TUN模式,这一点至关重要,直接决定安装能否顺利完成,在此郑重提醒三遍:科学上网,科学上网,科学上网。查看完整安装教程 →
evo-memory by evoscientist/evoskills
npx skills add https://github.com/evoscientist/evoskills --skill evo-memory一个持久化学习层,用于在构思和实验周期中积累研究知识。维护两个记忆存储库,并实现三种进化机制,将学习到的模式反馈到未来的研究中。
idea-tournament 并需要更新构思记忆experiment-pipeline 并需要更新记忆研究是迭代的。每个周期——从构思到实验——都会产生知识,这些知识应该为下一个周期提供信息。没有持久化记忆,每个新项目都从零开始,重复错误并重新发现模式。
Evo-memory 通过维护两个结构化的记忆存储库和三种进化机制来解决这个问题,这些机制跨周期提取、分类和注入知识。
位置 : /memory/ideation-memory.md
记录您对研究方向的认知——哪些领域有前景,哪些是死胡同。
两个部分 :
| 部分 | 包含内容 | 示例条目 |
|---|
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
| 可行方向 | 在先前周期中显示出前景的方向 | "用于少样本分类的对比学习——已确认可行,在锦标赛周期 2 中排名前三" |
| 不成功方向 | 尝试过但失败的方向,包含失败分类 | "用于实时视频的自回归生成——根本性失败:延迟约束与自回归解码不兼容" |
每个条目记录 : 方向名称、一句话摘要、证据(哪个周期,什么结果)、分类(可行 / 实施失败 / 根本性失败)、日期。
如何使用 : idea-tournament 在阶段 1 开始时读取 M_I。论文使用基于嵌入的检索和余弦相似度,选择前 k_I 个最相似的项目(实验中 k_I=2)。先前周期的可行方向可以用于生成新的树分支。不成功的方向在剪枝时使用——根本性失败被剪除;实施失败可能会重试。
位置 : /memory/experiment-memory.md
记录您对研究策略的认知——哪些技术方法和配置在实践中有效。
论文将 M_E 定义为存储“可复用的数据处理和模型训练策略”。ESE 共同总结了 (i) 数据处理策略 和 (ii) 模型训练策略。我们扩展了此定义,增加了两个实用的部分(架构和调试),以实现全面覆盖。
两个核心部分(来自论文)+ 两个实用扩展 :
| 部分 | 来源 | 包含内容 | 示例条目 |
|---|---|---|---|
| 数据处理策略 | 论文(核心) | 预处理、增强和数据处理的模式 | "对于噪声传感器数据:在归一化之前应用中值滤波器可将训练不稳定性降低约 40%" |
| 模型训练策略 | 论文(核心) | 超参数、训练技巧和训练计划 | "在 10% 的步数中进行学习率预热可防止 Transformer 微调中的早期发散" |
| 架构策略 | 扩展 | 设计选择、模块配置和结构模式 | "残差连接对于插入 Transformer 中超过 10 层的模块至关重要" |
| 调试策略 | 扩展 | 解决实验失败的诊断模式 | "当损失在 50% 的训练后停滞时:检查梯度范数——裁剪阈值可能设置得过于激进" |
每个条目记录 : 策略名称、上下文(何时使用此策略)、证据(哪个周期,什么结果)、通用性(领域特定或广泛适用)、日期。
如何使用 : experiment-pipeline 在每个周期开始时读取 M_E。论文使用基于嵌入的检索和余弦相似度,选择前 k_E 个最相似的项目(实验中 k_E=1)。来自先前周期的相关策略为超参数选择、数据处理决策和调试方法提供信息,从而减少所需的尝试次数。
触发条件 : 在 idea-tournament 完成后(阶段 3 的方向摘要可用)。
目的 : 从锦标赛结果中提取有前景的研究方向,并将其存储在 M_I 中供未来周期使用。
论文提示 : 使用 references/paper-prompts.md 中的 IDE 提示作为主要提取机制。根据原始研究方向填写 {user_goal},根据 /direction-summary.md 填写 {top_ranked_ideas},然后逐步推理提示。输出(包含标题、核心思想、为何有前景、要求、验证计划的 DIRECTION SUMMARY)直接用于以下步骤。
流程 :
/memory/ideation-memory.md 读取当前的 M_I关键原则 : 存储方向,而非具体想法。像“用于结构化数据的对比学习”这样的方向可以在未来周期中衍生出许多具体想法。像“在分子数据集上使用图增强的 SimCLR”这样的具体想法过于狭窄,无法复用。
完整流程请参见 references/ide-protocol.md。
触发条件(两种情况,遵循论文):
目的 : 对方法失败的原因进行分类,并相应更新 M_I。这是最关键的进化机制,因为它可以防止未来周期重复死胡同方向。
论文提示 : 使用 references/paper-prompts.md 中的 IVE 提示作为主要分类机制。根据 /research-proposal.md 填写 {research_proposal},根据阶段轨迹日志填写 {execution_report},然后逐步推理提示。该提示将失败分类为 FAILED(NoExecutableWithinBudget)、FAILED(WorseThanBaseline) 或 NOT_FAILED。
运行论文提示后 :
五个诊断问题(针对 WorseThanBaseline 情况):
如果 3 个或更多答案指向一种类型,则分类为该类型。如果结果分散,则分类为实施失败(更保守——允许重试)。
重试升级规则 : 如果一个方向在不同周期中被分类为“实施失败” 3 次,则升级为仔细重新评估——三次独立的实施失败可能表明该方向比看起来更难。考虑重新分类为根本性失败。
完整流程和工作示例请参见 references/ive-protocol.md。
触发条件 : 在 experiment-pipeline 成功后——所有 4 个阶段完成且满足关卡条件。
目的 : 从成功的实验运行中提炼可复用的策略,并将其存储在 M_E 中供未来周期使用。
论文提示 : 使用 references/paper-prompts.md 中的 ESE 提示作为主要提取机制。根据 /research-proposal.md 填写 {research_proposal},根据所有 4 个阶段的轨迹日志填写 {trajectories},然后逐步推理提示。该提示输出 DATA SUMMARY 和 MODEL SUMMARY,它们对应我们的数据处理策略和模型训练策略部分。
流程 :
泛化指南 : 如果一个策略解决的是普遍性挑战(训练不稳定、过拟合、收敛缓慢),而不是领域特定特征,那么它就是广泛适用的。如有疑问,请将上下文与策略一起记录,让未来的用户判断适用性。
完整流程请参见 references/ese-protocol.md。
当开始一个新的研究周期(加载 idea-tournament 或 experiment-pipeline)时:
/memory/ideation-memory.md 和 /memory/experiment-memory.mdidea-tournament:使用 M_I 的可行方向来生成树分支。在剪枝时使用 M_I 的不成功方向(仅限根本性失败)。experiment-pipeline:使用 M_E 策略为超参数范围、训练计划和调试方法提供信息。不要盲目应用旧策略。 上下文很重要。一个在图像分类中有效的策略可能不适用于文本生成。始终将记录下来的上下文与当前问题进行核对。
检索方法 : 论文使用基于嵌入的余弦相似度进行检索。在实践中,通过读取每个条目的摘要/上下文和检索标签,然后判断与当前目标的相关性来进行这种语义比较。如果您的环境中提供了自动化的嵌入工具,请使用这些工具来处理更大的记忆存储库。
/memory/ideation-memory.mdFailure Classification: Fundamental 的不成功方向:标记为剪枝。示例注入:"先前周期确认‘自回归实时视频生成’是根本性失败(O(n) 延迟)。剪除任何匹配此模式的树分支。"/memory/experiment-memory.md定期审查两个记忆存储库,并移除或归档不再相关的条目:
每个记忆文件都维护一个 Last Updated 字段和一个周期计数器。当条目被修改(不仅仅是追加)时,在进化报告中注明更改内容。这创建了您的研究知识如何演变的审计轨迹。
每次进化机制触发后,生成一份报告保存到 /memory/evolution-reports/cycle_N_type.md:
在更新和使用记忆时优先遵循这些规则:
存储前先抽象 : 存储方向和策略,而不是具体的实验细节。"对比学习改进少样本分类" 可以在许多项目中复用;"为 CIFAR-10 上的 ResNet-50 设置 lr=0.001" 则不行。目标是可迁移的知识,而不是实验记录本。
失败的方向比成功的更有价值 : 知道不该尝试什么比知道什么有效更能节省时间。成功案例发表在论文中——每个人都可以访问。失败案例很少被分享,这使得您的失败记忆成为独特的竞争优势。
实施失败不等于方向失败 : 最常见的进化错误是因为实现有缺陷而将一个好的方向标记为失败。IVE 的存在正是为了做出这种区分。如有疑问,分类为实施失败——重试一个好想法比永久丢弃它成本更低。
记忆不清理就会衰减 : 10 个周期前在不同数据上有效的策略可能不再相关。积累陈旧的条目会增加噪音,使得更难找到有用的策略。积极清理——一个更小、更精选的记忆比一个庞大、嘈杂的记忆更有价值。
交叉授粉胜过深度专业化 : 来自一个领域的 M_E 策略通常可以迁移到另一个领域。学习率预热在 NLP、视觉和语音中都有帮助。在开始新的实验流程之前,请审查整个 M_E,而不仅仅是领域特定的条目。
进化报告是给人看的 : 撰写研究人员(而不仅仅是 AI 代理)能够理解和采取行动的报告。包含足够的上下文,以便 6 个月后阅读报告的人能理解为什么做出了更改,而不仅仅是什么被更改了。
Evo-memory 如何连接到流程中的其他技能:
| 触发条件 | 来源技能 | 机制 | 更新的记忆 |
|---|---|---|---|
| 锦标赛完成 | idea-tournament | IDE | M_I(可行方向) |
| 在预算内无可执行代码,或方法性能低于基线 | experiment-pipeline | IVE | M_I(不成功方向) |
| 流程成功 | experiment-pipeline | ESE | M_E(数据处理 + 模型训练;可选架构 + 调试) |
| 新周期开始 | idea-tournament | 读取 (top-k_I=2) | 读取 M_I 用于生成/剪枝 |
| 新周期开始 | experiment-pipeline | 读取 (top-k_E=1) | 读取 M_E 用于策略指导 |
| 主题 | 参考文件 | 何时使用 |
|---|---|---|
| IDE 流程详情 | ide-protocol.md | 完成 idea-tournament 后 |
| IVE 流程详情 | ive-protocol.md | 在 experiment-pipeline 失败后(无可执行代码或方法性能不佳) |
| ESE 流程详情 | ese-protocol.md | 在 experiment-pipeline 成功后 |
| 论文的实际提示 | paper-prompts.md | 参考确切的 IDE/IVE/ESE 提示设计 |
| 记忆数据结构 | memory-schema.md | 理解 M_I 和 M_E 格式 |
| 构思记忆模板 | ideation-memory-template.md | 初始化 M_I |
| 实验记忆模板 | experiment-memory-template.md | 初始化 M_E |
| 进化报告模板 | evolution-report-template.md | 记录记忆更新 |
每周安装量
72
代码仓库
GitHub 星标数
109
首次出现
10 天前
安全审计
安装于
gemini-cli72
amp72
cline72
github-copilot72
kimi-cli72
opencode72
A persistent learning layer that accumulates research knowledge across ideation and experimentation cycles. Maintains two memory stores and implements three evolution mechanisms that feed learned patterns back into future research.
idea-tournament and needs to update Ideation Memoryexperiment-pipeline and needs to update memoryResearch is iterative. Each cycle — from ideation through experimentation — generates knowledge that should inform the next cycle. Without persistent memory, every new project starts from scratch, repeating mistakes and rediscovering patterns.
Evo-memory solves this by maintaining two structured memory stores and three evolution mechanisms that extract, classify, and inject knowledge across cycles.
Location : /memory/ideation-memory.md
Records what you've learned about research DIRECTIONS — which areas are promising and which are dead ends.
Two sections :
| Section | What It Contains | Example Entry |
|---|---|---|
| Feasible Directions | Directions that showed promise in prior cycles | "Contrastive learning for few-shot classification — confirmed feasible, top-3 in tournament cycle 2" |
| Unsuccessful Directions | Directions that were tried and failed, with failure classification | "Autoregressive generation for real-time video — fundamental failure: latency constraint incompatible with autoregressive decoding" |
Each entry records : Direction name, one-sentence summary, evidence (which cycle, what results), classification (feasible / implementation failure / fundamental failure), date.
How it's used : idea-tournament reads M_I at the start of Phase 1. The paper uses embedding-based retrieval with cosine similarity, selecting the top-k_I most similar items (k_I=2 in experiments). Feasible directions from prior cycles can seed new tree branches. Unsuccessful directions are used during pruning — fundamental failures are pruned; implementation failures may be retried.
See assets/ideation-memory-template.md for the template.
Location : /memory/experiment-memory.md
Records what you've learned about research STRATEGIES — which technical approaches and configurations work in practice.
The paper defines M_E as storing "reusable data processing and model training strategies." ESE jointly summarizes (i) a data processing strategy and (ii) a model training strategy. We extend this with two additional practical sections (architecture and debugging) for comprehensive coverage.
Two core sections (from paper) + two practical extensions :
| Section | Source | What It Contains | Example Entry |
|---|---|---|---|
| Data Processing Strategies | Paper (core) | Preprocessing, augmentation, and data handling patterns | "For noisy sensor data: median filter before normalization reduces training instability by ~40%" |
| Model Training Strategies | Paper (core) | Hyperparameters, training tricks, and training schedules | "Learning rate warmup for 10% of steps prevents early divergence in transformer fine-tuning" |
| Architecture Strategies | Extension | Design choices, module configurations, and structural patterns | "Residual connections are critical for modules inserted deeper than 10 layers in transformers" |
| Debugging Strategies | Extension | Diagnostic patterns that resolved experiment failures | "When loss plateaus after 50% of training: check gradient norm — clipping threshold may be too aggressive" |
Each entry records : Strategy name, context (when to use this), evidence (which cycle, what results), generality (domain-specific or broadly applicable), date.
How it's used : experiment-pipeline reads M_E at the start of each cycle. The paper uses embedding-based retrieval with cosine similarity, selecting the top-k_E most similar items (k_E=1 in experiments). Relevant strategies from prior cycles inform hyperparameter choices, data processing decisions, and debugging approaches, reducing the number of attempts needed.
See assets/experiment-memory-template.md for the template.
Trigger : After idea-tournament completes (Phase 3 direction summary is available).
Purpose : Extract promising research directions from the tournament results and store them in M_I for future cycles.
Paper Prompt : Use the IDE prompt from references/paper-prompts.md as the primary extraction mechanism. Fill in {user_goal} from the original research direction and {top_ranked_ideas} from /direction-summary.md, then reason through the prompt step by step. The output (DIRECTION SUMMARY with Title, Core idea, Why promising, Requirements, Validation plan) feeds directly into the steps below.
Process :
/memory/ideation-memory.mdKey principle : Store directions, not ideas. A direction like "contrastive learning for structured data" can spawn many specific ideas across future cycles. A specific idea like "SimCLR with graph augmentations on molecular datasets" is too narrow to be reusable.
See references/ide-protocol.md for the full process.
Trigger (two conditions, following the paper):
Purpose : Classify WHY the method failed and update M_I accordingly. This is the most critical evolution mechanism because it prevents future cycles from repeating dead-end directions.
Paper Prompt : Use the IVE prompt from references/paper-prompts.md as the primary classification mechanism. Fill in {research_proposal} from /research-proposal.md and {execution_report} from the stage trajectory logs, then reason through the prompt step by step. The prompt classifies the failure as FAILED(NoExecutableWithinBudget), FAILED(WorseThanBaseline), or NOT_FAILED.
After running the paper prompt :
Five diagnostic questions (for WorseThanBaseline cases):
If 3+ answers point to one type, classify as that type. If split, classify as implementation failure (more conservative — allows retry).
Retry escalation rule : If a direction has been classified as "implementation failure" 3 times across different cycles, escalate to a careful re-evaluation — three separate implementation failures may indicate the direction is harder than it appears. Consider reclassifying as fundamental.
See references/ive-protocol.md for the full process and worked examples.
Trigger : After experiment-pipeline succeeds — all 4 stages complete and gates met.
Purpose : Distill reusable strategies from the successful experiment run and store them in M_E for future cycles.
Paper Prompt : Use the ESE prompt from references/paper-prompts.md as the primary extraction mechanism. Fill in {research_proposal} from /research-proposal.md and {trajectories} from all 4 stage trajectory logs, then reason through the prompt step by step. The prompt outputs DATA SUMMARY and MODEL SUMMARY, which map to our Data Processing Strategies and Model Training Strategies sections.
Process :
Generalization guidelines : A strategy is broadly applicable if it addresses a general challenge (training instability, overfitting, slow convergence) rather than a domain-specific characteristic. When in doubt, record the context alongside the strategy and let future users judge applicability.
See references/ese-protocol.md for the full process.
When starting a new research cycle (loading idea-tournament or experiment-pipeline):
/memory/ideation-memory.md and /memory/experiment-memory.mdidea-tournament: Use M_I feasible directions to seed tree branches. Use M_I unsuccessful directions (fundamental failures only) during pruning.experiment-pipeline: Use M_E strategies to inform hyperparameter ranges, training schedules, and debugging approaches.Don't blindly apply old strategies. Context matters. A strategy that worked for image classification may not work for text generation. Always check the recorded context against the current problem.
Retrieval method : The paper uses embedding-based cosine similarity for retrieval. In practice, perform this semantic comparison by reading each entry's Summary/Context and Retrieval Tags, then judging relevance to the current goal. If automated embedding tools are available in your environment, use those instead for larger memory stores.
/memory/ideation-memory.mdFailure Classification: Fundamental: flag for pruning. Example injection: "Prior cycle confirmed 'Autoregressive real-time video generation' is a fundamental failure (O(n) latency). Prune any tree branch matching this pattern."/memory/experiment-memory.mdPeriodically review both memory stores and remove or archive entries that are no longer relevant:
Each memory file maintains a Last Updated field and a cycle counter. When entries are modified (not just appended), note what changed in the evolution report. This creates an audit trail of how your research knowledge evolves.
After each evolution mechanism triggers, generate a report saved to /memory/evolution-reports/cycle_N_type.md:
See assets/evolution-report-template.md for the template.
Prioritize these rules when updating and using memory:
Abstract before storing : Store directions and strategies, not specific experiment details. "Contrastive learning improves few-shot classification" is reusable across many projects; "set lr=0.001 for ResNet-50 on CIFAR-10" is not. The goal is transferable knowledge, not a lab notebook.
Failed directions are more valuable than successful ones : Knowing what NOT to try saves more time than knowing what worked. Success stories are published in papers — everyone can access them. Failure stories are rarely shared, making your failure memory a unique competitive advantage.
Implementation failures are not direction failures : The most common evolution mistake is marking a good direction as failed because the implementation was buggy. IVE exists specifically to make this distinction. When in doubt, classify as implementation failure — it's cheaper to retry a good idea than to permanently discard it.
Memory decays without pruning : A strategy that worked 10 cycles ago on different data may no longer be relevant. Accumulating stale entries adds noise that makes it harder to find useful strategies. Prune actively — a smaller, curated memory is more valuable than a large, noisy one.
Cross-pollination beats deep specialization : Strategies from M_E in one domain often transfer to another. Learning rate warmup helps in NLP AND vision AND speech. Review the full M_E before starting a new experiment pipeline, not just domain-specific entries.
The evolution report is for humans : Write reports that a researcher — not just an AI agent — can understand and act on. Include enough context that someone reading the report 6 months later understands WHY the change was made, not just WHAT changed.
How evo-memory connects to other skills in the pipeline:
| Trigger | Source Skill | Mechanism | Memory Updated |
|---|---|---|---|
| Tournament completed | idea-tournament | IDE | M_I (feasible directions) |
| No executable code within budget, or method underperforms baseline | experiment-pipeline | IVE | M_I (unsuccessful directions) |
| Pipeline succeeded | experiment-pipeline | ESE | M_E (data processing + model training; optionally architecture + debugging) |
| New cycle starts | idea-tournament | Read (top-k_I=2) |
| Topic | Reference File | When to Use |
|---|---|---|
| IDE process details | ide-protocol.md | After completing idea-tournament |
| IVE process details | ive-protocol.md | After experiment-pipeline failure (no executable code or method underperforms) |
| ESE process details | ese-protocol.md | After experiment-pipeline succeeds |
| Paper's actual prompts | paper-prompts.md | Reference for exact IDE/IVE/ESE prompt design |
| Memory data structures | memory-schema.md | Understanding M_I and M_E formats |
| Ideation memory template |
Weekly Installs
72
Repository
GitHub Stars
109
First Seen
10 days ago
Security Audits
Gen Agent Trust HubPassSocketPassSnykPass
Installed on
gemini-cli72
amp72
cline72
github-copilot72
kimi-cli72
opencode72
超能力技能使用指南:AI助手技能调用优先级与工作流程详解
53,700 周安装
通用项目发布工具 - 多语言更新日志自动生成 | 支持Node.js/Python/Rust/Claude插件
62 周安装
Edge Pipeline Orchestrator:自动化金融交易策略流水线编排工具
62 周安装
Python ROI 计算器:投资回报率、营销ROI、盈亏平衡分析工具
62 周安装
Salesforce Hyperforce 2025架构指南:云原生、零信任安全与开发最佳实践
62 周安装
PowerShell 2025 重大变更与迁移指南:版本移除、模块停用、WMIC替代方案
62 周安装
2025安全优先Bash脚本编写指南:输入验证、命令注入与路径遍历防护
62 周安装
| M_I read for seeding/pruning |
| New cycle starts | experiment-pipeline | Read (top-k_E=1) | M_E read for strategy guidance |
| ideation-memory-template.md |
| Initializing M_I |
| Experiment memory template | experiment-memory-template.md | Initializing M_E |
| Evolution report template | evolution-report-template.md | Documenting memory updates |