✅ TICKET-006: Wake-word Detection Service - Implemented wake-word detection using openWakeWord - HTTP/WebSocket server on port 8002 - Real-time detection with configurable threshold - Event emission for ASR integration - Location: home-voice-agent/wake-word/ ✅ TICKET-010: ASR Service - Implemented ASR using faster-whisper - HTTP endpoint for file transcription - WebSocket endpoint for streaming transcription - Support for multiple audio formats - Auto language detection - GPU acceleration support - Location: home-voice-agent/asr/ ✅ TICKET-014: TTS Service - Implemented TTS using Piper - HTTP endpoint for text-to-speech synthesis - Low-latency processing (< 500ms) - Multiple voice support - WAV audio output - Location: home-voice-agent/tts/ ✅ TICKET-047: Updated Hardware Purchases - Marked Pi5 kit, SSD, microphone, and speakers as purchased - Updated progress log with purchase status 📚 Documentation: - Added VOICE_SERVICES_README.md with complete testing guide - Each service includes README.md with usage instructions - All services ready for Pi5 deployment 🧪 Testing: - Created test files for each service - All imports validated - FastAPI apps created successfully - Code passes syntax validation 🚀 Ready for: - Pi5 deployment - End-to-end voice flow testing - Integration with MCP server Files Added: - wake-word/detector.py - wake-word/server.py - wake-word/requirements.txt - wake-word/README.md - wake-word/test_detector.py - asr/service.py - asr/server.py - asr/requirements.txt - asr/README.md - asr/test_service.py - tts/service.py - tts/server.py - tts/requirements.txt - tts/README.md - tts/test_service.py - VOICE_SERVICES_README.md Files Modified: - tickets/done/TICKET-047_hardware-purchases.md Files Moved: - tickets/backlog/TICKET-006_prototype-wake-word-node.md → tickets/done/ - tickets/backlog/TICKET-010_streaming-asr-service.md → tickets/done/ - tickets/backlog/TICKET-014_tts-service.md → tickets/done/
4080 LLM Server (Work Agent)
LLM server for work agent running on remote GPU VM.
Server Information
- Host: 10.0.30.63
- Port: 11434
- Endpoint: http://10.0.30.63:11434
- Service: Ollama
Available Models
The server has the following models available:
deepseek-r1:70b- 70B model (currently configured)deepseek-r1:671b- 671B modelllama3.1:8b- Llama 3.1 8Bqwen2.5:14b- Qwen 2.5 14B- And others (see
test_connection.py)
Configuration
Edit config.py to change the model:
MODEL_NAME = "deepseek-r1:70b" # or your preferred model
Testing Connection
cd home-voice-agent/llm-servers/4080
python3 test_connection.py
This will:
- Test server connectivity
- List available models
- Test chat endpoint with configured model
API Usage
List Models
curl http://10.0.30.63:11434/api/tags
Chat Request
curl http://10.0.30.63:11434/api/chat -d '{
"model": "deepseek-r1:70b",
"messages": [
{"role": "user", "content": "Hello"}
],
"stream": false
}'
With Function Calling
curl http://10.0.30.63:11434/api/chat -d '{
"model": "deepseek-r1:70b",
"messages": [
{"role": "user", "content": "What is the weather in San Francisco?"}
],
"tools": [...],
"stream": false
}'
Integration
The MCP adapter can connect to this server by setting:
OLLAMA_BASE_URL = "http://10.0.30.63:11434"
Notes
- The server is already running on the GPU VM
- No local installation needed - just configure the endpoint
- Model selection can be changed in
config.py - If you need
llama3.1:70b-q4_0, pull it on the server:# On the GPU VM ollama pull llama3.1:70b-q4_0