Build an embodied voice AI agent with long-term memory using Honcho
Reachy Mini is Hugging Face and Pollen Robotics’ open-source robot for human-robot interaction. This guide integrates Honcho for persistent, multi-user memory with OpenAI’s Realtime API for voice.
Real-time memory: Honcho’s async API is designed for live voice interactions. Messages persist in the background without blocking audio, and the dialectic API returns user context fast enough for mid-conversation tool calls.
The robot calls Honcho mid-conversation via OpenAI function calling — fast enough for real-time voice:
Tool
Purpose
recall
Query Honcho about the user (“What’s their name?”)
create_conclusion
Save important facts to long-term memory
see
Capture and analyze camera feed
Copy
# Recall - ask Honcho's dialectic API (returns in ~200-500ms)result = await user_peer.aio.chat( "What do I know about this user?", session=session, reasoning_level="medium")# Create conclusion - save a factawait user_peer.conclusions_of(user_id).aio.create([ {"content": "Their name is Alice"}])
Face recognition identifies returning users. When a new face is detected, the agent:
Flushes pending transcripts to the previous user’s session
Switches Honcho context to the new user
Fetches a briefing from Honcho’s dialectic API
Reconnects OpenAI with fresh context and triggers a greeting
Copy
# Get briefing when user is recognizedbriefing = await user_peer.aio.chat( "What should I know about this user? Name, interests, recent topics.", session=session, reasoning_level="low")
SYSTEM_PROMPT = """You are Reachy, a friendly robot. Keep responses concise.You have a recall tool for memory. ALWAYS use it before claiming you don'tknow something about the user. Never say "Nice to meet you" if you've met before."""