Reachy Mini · Python app · by SantiPa
Marginalia
A small voice in the margin of the page. Ask Reachy Mini about any book, character, motif, or piece of lore — from Hamlet and Mrs Dalloway to Confessions of a Justified Sinner, Dune, Borges' Ficciones, and the Silmarils. Reachy reads each answer aloud — or twist either antenna for a silent reading.
Preview only. The real conversation lives at
http://0.0.0.0:8042 when the app is running on
your robot.
How to use it on your robot
- Open the dashboard. Launch the Reachy Mini Control App on your laptop and connect to your robot (or the bundled simulator).
-
Discover Apps → search
marginalia. Click Install. The dashboard pulls this Space, sets up a private virtualenv, and registers the app's entry point. -
Hit Start. The browser opens at
http://0.0.0.0:8042automatically. - Ask a question. Reachy sways its head while it thinks, nods the moment the first token lands, then starts speaking the answer paragraph by paragraph — the first words come through the speaker about a second after the LLM finishes, while the rest of the answer is still being synthesised in the background. The head dips, antennas perk up, and the body sways with the audio. Each answer comes with a How to read this panel of tips tailored to the work, plus Further reading for amplification.
- Prefer to read in silence? Twist either antenna and Reachy will switch to read mode — text only on the page, no speaker. Twist again to bring the voice back. The pill at the top of the in-app UI mirrors the same toggle if your hands are on the keyboard.
What every answer contains
- The answer itself. 3-6 paragraphs, in measured prose, with inline source citations [1] [2] [3].
- How to read this. Concrete, idiosyncratic tips calibrated to this author or work — Shakespeare doesn't read the same way as Dune, and Dune doesn't read the same way as Borges.
- Further reading. Links to canonical, freely-accessible sources: Project Gutenberg, the Folger Shakespeare Library, the Stanford Encyclopedia of Philosophy, Poetry Foundation, the Internet Archive, In Our Time, the LRB and NYRB archives.
- In the margin. Wikipedia search hits with intro extracts and thumbnail portraits/covers, so you can glance at the source material before clicking through.
- An embodied reading. The robot speaks the answer paragraph by paragraph in a natural neural voice (Microsoft Edge TTS — pick from US, UK, IE, AU, male or female from a small dropdown), and the head, antennas, and body move in time with the audio waveform: dips on emphasised words, antenna lifts on loud syllables. Twist an antenna any time to switch to silent read mode.
How it works
- Wikipedia search. Your question goes through a single combined Wikipedia API call that returns titles, intro extracts, full URLs, and thumbnail images for the top hits — one round-trip, no key required.
-
LLM, streamed. The question and the search
excerpts are sent to a Hugging Face hosted model
(
meta-llama/Llama-3.3-70B-Instructby default) viaInferenceClient.chat_completion(stream=True). No extra API key — the app uses the cachedhf auth logintoken already on your machine. - Server-Sent Events. Tokens stream back to the browser, which renders markdown progressively with a deep-emerald drop-cap, gilded citation underlines, and chapter-style ornaments separating the three sections.
- Embodied reading. A 50 Hz state machine drives the head, antennas, and body together: idle sway → thinking sway with curling antennas → a single nod the moment the first token arrives → an attentive bob while the rest streams in → audio-reactive head and antenna motion while the answer is read aloud, so the robot moves with the voice rather than on a fixed loop.
-
Voice. The cleaned answer prose is split
into paragraph chunks. Each chunk is synthesised through
edge-tts (Microsoft's free neural
voices — Ava, Andrew, Sonia, Ryan, …) and
pushed to the robot speaker as soon as it's ready, so
the next chunk renders while the previous one is
already playing. Fallbacks: HF Inference TTS,
pyttsx3, gTTS. Twisting either antenna above ~23° toggles spoken vs silent mode and is acknowledged with a small head nod.
Try without a robot
The Reachy Mini Control App ships with a built-in MuJoCo simulator — the same install path works without hardware. Pick the simulation connection mode in the desktop app, then install Marginalia from Discover Apps as usual.
More: install guide · SDK on GitHub