重要前提
安装AI Skills的关键前提是:必须科学上网,且开启TUN模式,这一点至关重要,直接决定安装能否顺利完成,在此郑重提醒三遍:科学上网,科学上网,科学上网。查看完整安装教程 →
audio-voice-recovery by pproenca/dot-skills
npx skills add https://github.com/pproenca/dot-skills --skill audio-voice-recovery全面的音频法证和语音恢复指南,提供 CSI 级别的能力,用于从低质量、低音量或受损的音频录音中恢复语音。包含 8 个类别共 45 条规则,按影响优先级排序,以指导音频增强、法证分析和转录工作流程。
在以下情况下参考本指南:
| 优先级 | 类别 | 影响 | 前缀 | 规则数量 |
|---|---|---|---|---|
| 1 | 信号保存与分析 | 关键 | signal- | 5 |
| 2 | 噪声分析与估计 | 关键 | noise- | 5 |
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
| 3 | 频谱处理 | 高 | spectral- | 6 |
| 4 | 语音分离与增强 | 高 | voice- | 7 |
| 5 | 时域处理 | 中高 | temporal- | 5 |
| 6 | 转录与识别 | 中 | transcribe- | 5 |
| 7 | 法证认证 | 中 | forensic- | 5 |
| 8 | 工具集成与自动化 | 低中 | tool- | 7 |
signal-preserve-original - 绝不修改原始录音signal-lossless-format - 使用无损格式进行处理signal-sample-rate - 保持原生采样率signal-bit-depth - 处理时使用最大位深度signal-analyze-first - 处理前先进行分析noise-profile-silence - 从静音片段提取噪声特征noise-identify-type - 降噪前识别噪声类型noise-adaptive-estimation - 对非平稳噪声使用自适应估计noise-snr-assessment - 测量处理前后的信噪比noise-avoid-overprocessing - 避免过度处理和音乐噪声伪影spectral-subtraction - 对平稳噪声应用谱减法spectral-wiener-filter - 使用维纳滤波器进行最优噪声估计spectral-notch-filter - 应用陷波滤波器处理单音干扰spectral-band-limiting - 对语音应用频带限制spectral-equalization - 使用法证均衡恢复可懂度spectral-declip - 在其他处理前修复削波音频voice-rnnoise - 使用 RNNoise 进行实时机器学习降噪voice-dialogue-isolate - 对复杂背景使用源分离voice-formant-preserve - 在音高调整期间保留共振峰voice-dereverb - 对房间回声应用去混响voice-enhance-speech - 使用 AI 语音增强服务快速获得结果voice-vad-segment - 使用 VAD 进行针对性处理voice-frequency-boost - 提升特定音素的频率区域temporal-dynamic-range - 使用动态范围压缩保持电平一致性temporal-noise-gate - 应用噪声门静音非语音片段temporal-time-stretch - 使用时间拉伸提高可懂度temporal-transient-repair - 修复瞬态损伤(咔嗒声、爆音、信号丢失)temporal-silence-trim - 导出前修剪静音并归一化transcribe-whisper - 使用 Whisper 进行抗噪转录transcribe-multipass - 对困难音频使用多轮转录transcribe-segment - 分割音频进行针对性转录transcribe-confidence - 跟踪不确定词汇的置信度分数transcribe-hallucination - 检测并过滤 ASR 幻觉forensic-enf-analysis - 使用 ENF 分析进行时间戳验证forensic-metadata - 提取并验证音频元数据forensic-tampering - 检测音频篡改和拼接forensic-chain-custody - 记录证据的保管链forensic-speaker-id - 提取说话人特征用于识别tool-ffmpeg-essentials - 掌握核心 FFmpeg 音频命令tool-sox-commands - 使用 SoX 进行高级音频操作tool-python-pipeline - 构建 Python 音频处理流水线tool-audacity-workflow - 使用 Audacity 进行可视化分析和手动编辑tool-install-guide - 安装音频法证工具链tool-batch-automation - 自动化批处理工作流程tool-quality-assessment - 测量音频质量指标| 工具 | 用途 | 安装命令 |
|---|---|---|
| FFmpeg | 格式转换,滤波 | brew install ffmpeg |
| SoX | 噪声分析,效果处理 | brew install sox |
| Whisper | 语音转录 | pip install openai-whisper |
| librosa | Python 音频分析 | pip install librosa |
| noisereduce | 机器学习降噪 | pip install noisereduce |
| Audacity | 可视化编辑 | brew install audacity |
使用捆绑的脚本来生成客观基线、创建工作流程计划并验证结果。
scripts/preflight_audio.py - 生成法证飞行前检查报告 (JSON 或 Markdown)。scripts/plan_from_preflight.py - 根据飞行前检查报告创建工作流程计划模板。scripts/compare_audio.py - 比较基线与处理后音频的客观指标。使用示例:
# 1) 分析并捕获基线指标
python3 skills/.experimental/audio-voice-recovery/scripts/preflight_audio.py evidence.wav --out preflight.json
# 2) 生成工作流程计划模板
python3 skills/.experimental/audio-voice-recovery/scripts/plan_from_preflight.py --preflight preflight.json --out plan.md
# 3) 比较基线 vs 处理后指标
python3 skills/.experimental/audio-voice-recovery/scripts/compare_audio.py \
--before evidence.wav \
--after enhanced.wav \
--format md \
--out comparison.md
使飞行前检查与 SWGDE 数字音频增强最佳实践 (20-a-001) 和 SWGDE 法证音频最佳实践 (08-a-001) 保持一致。建立客观的基线状态并规划工作流程,确保处理过程不会引入削波、伪影或虚假的“完成”信心。使用 scripts/preflight_audio.py 捕获基线指标,并将报告与案件文件一同保存。
在处理前捕获并记录:
流程:
scripts/plan_from_preflight.py 生成计划草案,并根据具体案件决策完成它。失败模式防护措施:
# 1. 分析原始文件 (运行飞行前检查并捕获基线指标)
python3 skills/.experimental/audio-voice-recovery/scripts/preflight_audio.py evidence.wav --out preflight.json
# 2. 创建带校验和的工作副本
cp evidence.wav working.wav
sha256sum evidence.wav > evidence.sha256
# 3. 应用增强
ffmpeg -i working.wav -af "\
highpass=f=80,\
adeclick=w=55:o=75,\
afftdn=nr=12:nf=-30:nt=w,\
equalizer=f=2500:t=q:w=1:g=3,\
loudnorm=I=-16:TP=-1.5:LRA=11\
" enhanced.wav
# 4. 转录
whisper enhanced.wav --model large-v3 --language en
# 5. 验证原始文件未改变
sha256sum -c evidence.sha256
# 6. 验证改进 (客观比较 + A/B 试听)
python3 skills/.experimental/audio-voice-recovery/scripts/compare_audio.py \
--before evidence.wav \
--after enhanced.wav \
--format md \
--out comparison.md
阅读各个参考文件以获取详细解释和代码示例:
| 文件 | 描述 |
|---|---|
| AGENTS.md | 包含所有规则的完整编译指南 |
| references/_sections.md | 类别定义和排序 |
| assets/templates/_template.md | 新规则模板 |
| metadata.json | 版本和参考信息 |
每周安装次数
58
代码仓库
GitHub 星标数
88
首次出现
2026 年 2 月 5 日
安全审计
安装于
codex54
gemini-cli53
github-copilot52
cursor52
opencode52
kimi-cli51
Comprehensive audio forensics and voice recovery guide providing CSI-level capabilities for recovering voice from low-quality, low-volume, or damaged audio recordings. Contains 45 rules across 8 categories, prioritized by impact to guide audio enhancement, forensic analysis, and transcription workflows.
Reference these guidelines when:
| Priority | Category | Impact | Prefix | Rules |
|---|---|---|---|---|
| 1 | Signal Preservation & Analysis | CRITICAL | signal- | 5 |
| 2 | Noise Profiling & Estimation | CRITICAL | noise- | 5 |
| 3 | Spectral Processing | HIGH | spectral- | 6 |
| 4 | Voice Isolation & Enhancement | HIGH | voice- | 7 |
| 5 | Temporal Processing | MEDIUM-HIGH | temporal- | 5 |
| 6 | Transcription & Recognition | MEDIUM | transcribe- | 5 |
| 7 | Forensic Authentication | MEDIUM | forensic- | 5 |
| 8 | Tool Integration & Automation | LOW-MEDIUM | tool- | 7 |
signal-preserve-original - Never modify original recordingsignal-lossless-format - Use lossless formats for processingsignal-sample-rate - Preserve native sample ratesignal-bit-depth - Use maximum bit depth for processingsignal-analyze-first - Analyze before processingnoise-profile-silence - Extract noise profile from silent segmentsnoise-identify-type - Identify noise type before reductionnoise-adaptive-estimation - Use adaptive estimation for non-stationary noisenoise-snr-assessment - Measure SNR before and afternoise-avoid-overprocessing - Avoid over-processing and musical artifactsspectral-subtraction - Apply spectral subtraction for stationary noisespectral-wiener-filter - Use Wiener filter for optimal noise estimationspectral-notch-filter - Apply notch filters for tonal interferencespectral-band-limiting - Apply frequency band limiting for speechspectral-equalization - Use forensic equalization to restore intelligibilityvoice-rnnoise - Use RNNoise for real-time ML denoisingvoice-dialogue-isolate - Use source separation for complex backgroundsvoice-formant-preserve - Preserve formants during pitch manipulationvoice-dereverb - Apply dereverberation for room echovoice-enhance-speech - Use AI speech enhancement services for quick resultstemporal-dynamic-range - Use dynamic range compression for level consistencytemporal-noise-gate - Apply noise gate to silence non-speech segmentstemporal-time-stretch - Use time stretching for intelligibilitytemporal-transient-repair - Repair transient damage (clicks, pops, dropouts)temporal-silence-trim - Trim silence and normalize before exporttranscribe-whisper - Use Whisper for noise-robust transcriptiontranscribe-multipass - Use multi-pass transcription for difficult audiotranscribe-segment - Segment audio for targeted transcriptiontranscribe-confidence - Track confidence scores for uncertain wordstranscribe-hallucination - Detect and filter ASR hallucinationsforensic-enf-analysis - Use ENF analysis for timestamp verificationforensic-metadata - Extract and verify audio metadataforensic-tampering - Detect audio tampering and splicesforensic-chain-custody - Document chain of custody for evidenceforensic-speaker-id - Extract speaker characteristics for identificationtool-ffmpeg-essentials - Master essential FFmpeg audio commandstool-sox-commands - Use SoX for advanced audio manipulationtool-python-pipeline - Build Python audio processing pipelinestool-audacity-workflow - Use Audacity for visual analysis and manual editingtool-install-guide - Install audio forensic toolchain| Tool | Purpose | Install |
|---|---|---|
| FFmpeg | Format conversion, filtering | brew install ffmpeg |
| SoX | Noise profiling, effects | brew install sox |
| Whisper | Speech transcription | pip install openai-whisper |
| librosa | Python audio analysis | pip install librosa |
| noisereduce | ML noise reduction | pip install noisereduce |
Use the bundled scripts to generate objective baselines, create a workflow plan, and verify results.
scripts/preflight_audio.py - Generate a forensic preflight report (JSON or Markdown).scripts/plan_from_preflight.py - Create a workflow plan template from the preflight report.scripts/compare_audio.py - Compare objective metrics between baseline and processed audio.Example usage:
# 1) Analyze and capture baseline metrics
python3 skills/.experimental/audio-voice-recovery/scripts/preflight_audio.py evidence.wav --out preflight.json
# 2) Generate a workflow plan template
python3 skills/.experimental/audio-voice-recovery/scripts/plan_from_preflight.py --preflight preflight.json --out plan.md
# 3) Compare baseline vs processed metrics
python3 skills/.experimental/audio-voice-recovery/scripts/compare_audio.py \
--before evidence.wav \
--after enhanced.wav \
--format md \
--out comparison.md
Align preflight with SWGDE Best Practices for the Enhancement of Digital Audio (20-a-001) and SWGDE Best Practices for Forensic Audio (08-a-001). Establish an objective baseline state and plan the workflow so processing does not introduce clipping, artifacts, or false "done" confidence. Use scripts/preflight_audio.py to capture baseline metrics and preserve the report with the case file.
Capture and record before processing:
Procedure:
scripts/plan_from_preflight.py and complete it with case-specific decisions.Failure-pattern guardrails:
# 1. Analyze original (run preflight and capture baseline metrics)
python3 skills/.experimental/audio-voice-recovery/scripts/preflight_audio.py evidence.wav --out preflight.json
# 2. Create working copy with checksum
cp evidence.wav working.wav
sha256sum evidence.wav > evidence.sha256
# 3. Apply enhancement
ffmpeg -i working.wav -af "\
highpass=f=80,\
adeclick=w=55:o=75,\
afftdn=nr=12:nf=-30:nt=w,\
equalizer=f=2500:t=q:w=1:g=3,\
loudnorm=I=-16:TP=-1.5:LRA=11\
" enhanced.wav
# 4. Transcribe
whisper enhanced.wav --model large-v3 --language en
# 5. Verify original unchanged
sha256sum -c evidence.sha256
# 6. Verify improvement (objective comparison + A/B listening)
python3 skills/.experimental/audio-voice-recovery/scripts/compare_audio.py \
--before evidence.wav \
--after enhanced.wav \
--format md \
--out comparison.md
Read individual reference files for detailed explanations and code examples:
| File | Description |
|---|---|
| AGENTS.md | Complete compiled guide with all rules |
| references/_sections.md | Category definitions and ordering |
| assets/templates/_template.md | Template for new rules |
| metadata.json | Version and reference information |
Weekly Installs
58
Repository
GitHub Stars
88
First Seen
Feb 5, 2026
Security Audits
Gen Agent Trust HubPassSocketPassSnykPass
Installed on
codex54
gemini-cli53
github-copilot52
cursor52
opencode52
kimi-cli51
AI界面设计评审工具 - 全面评估UI/UX设计质量、检测AI生成痕迹与优化用户体验
58,500 周安装
spectral-declipvoice-vad-segmentvoice-frequency-boost - Boost frequency regions for specific phonemestool-batch-automationtool-quality-assessment - Measure audio quality metrics| Audacity | Visual editing | brew install audacity |