- Deploy Music Assistant on Pi (10.0.0.199:8095) with host networking for
Chromecast mDNS discovery, Spotify + SMB library support
- Switch primary LLM from Ollama to Claude Sonnet 4 (Anthropic API),
local models remain as fallback
- Add model info tag under each assistant message in dashboard chat,
persisted in conversation JSON
- Rewrite homeai-agent/setup.sh: loads .env, injects API keys into plists,
symlinks plists to ~/Library/LaunchAgents/, smoke tests services
- Update install_service() in common.sh to use symlinks instead of copies
- Open UFW ports on Pi for Music Assistant (8095, 8097, 8927)
- Add ANTHROPIC_API_KEY to openclaw + bridge launchd plists
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Character schema v2: background, dialogue_style, appearance, skills, gaze_presets
with automatic v1→v2 migration. LLM-assisted character creation via Character MCP
server. Two-tier memory system (personal per-character + general shared) with
budget-based injection into LLM system prompt. Per-character TTS voice routing via
state file — Wyoming TTS server reads active config to route between Kokoro (local)
and ElevenLabs (cloud PCM 24kHz). Dashboard: memories page, conversation history,
character profile on cards, auto-TTS engine selection from character config.
Also includes VTube Studio expression bridge and ComfyUI API guide.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
- Pull qwen2.5:7b model (~4.7GB) with native tool-calling support
- Configure OpenClaw to use qwen2.5:7b as primary model
- Fix HASS_TOKEN file (remove trailing comment)
- Verify tool calling works end-to-end with HA skill
- Test home-assistant skill: turn_on/turn_off lights
- Update TODO.md with completed Phase 4 tasks
- Add PHASE4_COMPLETION.md documentation
Tool calling now working:
✓ qwen2.5:7b returns proper tool_calls array
✓ OpenClaw parses and executes commands
✓ Home Assistant skill controls entities
✓ HA API connectivity verified