voice-ai-integration by qodex-ai/ai-agent-skills
npx skills add https://github.com/qodex-ai/ai-agent-skills --skill voice-ai-integration构建智能的语音驱动 AI 应用程序,使其能够理解口语并通过音频自然地响应,创造无缝的语音优先用户体验。
语音 AI 系统结合了三种关键能力:
实现示例请参见 examples/speech_recognition_providers.py:
实现示例请参见 examples/text_to_speech_providers.py:
示例请参见 :
广告位招租
在这里展示您的产品或服务
触达数万 AI 开发者,精准高效
VoiceAssistantRealTimeVoiceProcessor 示例请参见 examples/realtime_voice_processor.py:
class SmartHomeVoiceAgent:
def __init__(self):
self.voice_assistant = VoiceAssistant()
self.devices = {
"lights": SmartLights(),
"temperature": SmartThermostat(),
"security": SecuritySystem()
}
async def handle_voice_command(self, audio_input):
# 从语音获取文本
command_text = await self.voice_assistant.process_voice_input(audio_input)
# 解析意图
intent = parse_smart_home_intent(command_text)
# 执行命令
if intent.action == "turn_on_lights":
self.devices["lights"].turn_on(intent.room)
elif intent.action == "set_temperature":
self.devices["temperature"].set(intent.value)
# 语音确认
response = f"I've {intent.action_description}"
audio_output = await self.voice_assistant.synthesize_response(response)
return audio_output
class VoiceMeetingRecorder:
def __init__(self):
self.processor = RealTimeVoiceProcessor()
self.transcripts = []
async def record_and_transcribe_meeting(self, duration_seconds=3600):
audio_stream = self.processor.stream_audio_input()
buffer = []
chunk_duration = 30 # 每 30 秒转录一次
for audio_chunk in audio_stream:
buffer.append(audio_chunk)
if sum(len(chunk) for chunk in buffer) >= chunk_duration * 16000:
# 转录块
transcript = transcribe_audio_whisper(buffer)
self.transcripts.append({
"timestamp": datetime.now(),
"text": transcript
})
buffer = []
return self.transcripts
解决方案:
解决方案:
解决方案:
每周安装量
86
代码仓库
GitHub 星标数
5
首次出现
2026 年 1 月 22 日
安全审计
安装于
opencode64
gemini-cli63
codex63
cursor62
github-copilot59
cline55
Build intelligent voice-enabled AI applications that understand spoken language and respond naturally through audio, creating seamless voice-first user experiences.
Voice AI systems combine three key capabilities:
See examples/speech_recognition_providers.py for implementations:
See examples/text_to_speech_providers.py for implementations:
See examples/voice_assistant.py for VoiceAssistant:
See examples/realtime_voice_processor.py for RealTimeVoiceProcessor:
class SmartHomeVoiceAgent:
def __init__(self):
self.voice_assistant = VoiceAssistant()
self.devices = {
"lights": SmartLights(),
"temperature": SmartThermostat(),
"security": SecuritySystem()
}
async def handle_voice_command(self, audio_input):
# Get text from voice
command_text = await self.voice_assistant.process_voice_input(audio_input)
# Parse intent
intent = parse_smart_home_intent(command_text)
# Execute command
if intent.action == "turn_on_lights":
self.devices["lights"].turn_on(intent.room)
elif intent.action == "set_temperature":
self.devices["temperature"].set(intent.value)
# Confirm with voice
response = f"I've {intent.action_description}"
audio_output = await self.voice_assistant.synthesize_response(response)
return audio_output
class VoiceMeetingRecorder:
def __init__(self):
self.processor = RealTimeVoiceProcessor()
self.transcripts = []
async def record_and_transcribe_meeting(self, duration_seconds=3600):
audio_stream = self.processor.stream_audio_input()
buffer = []
chunk_duration = 30 # Transcribe every 30 seconds
for audio_chunk in audio_stream:
buffer.append(audio_chunk)
if sum(len(chunk) for chunk in buffer) >= chunk_duration * 16000:
# Transcribe chunk
transcript = transcribe_audio_whisper(buffer)
self.transcripts.append({
"timestamp": datetime.now(),
"text": transcript
})
buffer = []
return self.transcripts
Solutions :
Solutions :
Solutions :
Weekly Installs
86
Repository
GitHub Stars
5
First Seen
Jan 22, 2026
Security Audits
Gen Agent Trust HubPassSocketPassSnykPass
Installed on
opencode64
gemini-cli63
codex63
cursor62
github-copilot59
cline55
AI 代码实施计划编写技能 | 自动化开发任务分解与 TDD 流程规划工具
50,900 周安装