A wearable biomedical device that reads the silent electrical language of your muscles and speaks it back — in your own voice.
EMG · CH4 (MASSETER) · RAW SIGNAL · 2kSPS
When we speak, our brain sends precise electrical signals to over a dozen facial and neck muscles — the masseter, the digastric, the risorius, the zygomaticus major. Even when no sound comes out, those signals still fire. That is the fundamental insight behind Silent Speech Interface research, and it is the foundation of TrueVoice.
Surface electromyography (sEMG) can measure these micro-electrical patterns at the skin surface. With enough channels, simultaneous sampling, and modern AI, those patterns can be decoded into the phonemes and words a person intended to say — and then spoken aloud in that person's own cloned voice.
This is not science fiction. It is an active frontier of biomedical research, and it is what TrueVoice is building toward.
Concept Vision
Eight electrode patches on the jaw, cheek, and neck. The main unit clips to the collar. The phone speaks. No throat buzzer — just the person's own voice, restored.
My life partner was 21 when throat cancer took her voice. What replaced it was a plastic device that buzzed when pressed to the neck. It let her be understood. It did not let her be heard — not really, not as herself.
She passed away in 2005. I'm not sharing this to tell a compelling founder story. I'm sharing it because it's the reason this project exists, and because she deserves to be the reason — not a footnote in a pitch deck.
Twenty-one years later, the technology finally exists to do this properly. TrueVoice is not a product looking for a market. It's a long-overdue answer to a problem I watched someone I loved face alone.
— Founder, JP Media R&D Lab · Nova Scotia, Canada
TrueVoice is not guesswork. Every architectural decision is grounded in published academic research. Here are the foundational papers informing our hardware and AI design.
University of the Basque Country
The most directly applicable paper to TrueVoice. Specifically targeting laryngectomized speakers, this study identified that a minimum of 8 bipolar electrode channels are required for phoneme-level classification — and mapped the exact muscles to target.
Key Findings We Implemented
IEEE Transactions on Instrumentation & Measurement
Validated that CNN architectures with transfer learning can achieve significant accuracy improvements on limited-channel EMG data. Demonstrated that 8 optimally-placed channels can approximate full-array performance.
Key Findings We Implemented
Meta has invested heavily in wrist-based EMG for AR/VR gesture control. However, wrist EMG captures gross hand motor signals — a far simpler problem than speech articulation. Jaw and neck EMG is more complex and more informative. TrueVoice operates in harder, richer territory.
As of 2025, there is no commercially available EMG silent speech interface on the market. No product exists that combines 8-channel EMG acquisition with on-device feature extraction, BLE streaming, and personal voice cloning — at a price accessible outside of academic grants. TrueVoice is building that product.
A five-stage pipeline transforms invisible muscle movement into audible, personalized speech.
Eight Ag/AgCl bipolar electrode pairs adhere to the jaw, cheek, and neck at muscle sites identified by the 2025 Basque Country research. Each pair feeds raw differential EMG into the ADS1298 — a 24-bit medical-grade analog front end sampling simultaneously at 2,000 samples/second per channel.
The ESP32-S3 reads the ADS1298 over SPI and computes log-Mel spectrogram features in real time — compressing the raw 256 kbps stream down to approximately 10–20 kbps of feature vectors for stable BLE transmission.
On the phone, a binary gating model distinguishes intentional speech from eating or swallowing artifacts. Only when speech is detected does the primary CNN/Transformer phoneme classifier activate.
Raw phoneme sequences are resolved by a lightweight on-device language model that uses context to select the most likely intended word before passing to synthesis.
The phone speaks the output using the user's personal voice clone — recorded before voice loss. End-to-end latency approximately 300–400ms. Their words. Their voice.
A 50×50mm 4-layer PCB designed to medical-grade signal acquisition standards.
Hardware Concept
TrueVoice SSI v1.0 — 50×50mm · 4-layer JLC3313 · ADS1298 + ESP32-S3 · ENIG finish
Complete schematic, BOM, netlist, placement grid, and routing playbook verified through multiple independent audit passes.
Footprint library, DRC custom rules, netclasses, layer stackup, and placement coordinates locked for KiCad 8.
Laying copper in KiCad. EMG differential pairs, DRL/REF routing, SPI corridor, and power planes in sequence.
JLC3313 controlled-impedance stackup. First physical prototype boards.
Register readback, lead-off detection, raw EMG signal capture and validation.
EMG dataset collection, phoneme classifier training, voice cloning integration, iOS/Android deployment.
TrueVoice is an open-architecture project. We are actively looking for collaboration with speech-language pathologists, oncology teams, accessibility advocates, embedded engineers, and AI researchers.
Most critically — we are looking for patients who have lost or are losing their voice and would be willing to help train and validate the system. Without real users, the AI cannot learn. Your participation is the most important contribution possible.
Reach Out to the Lab