logo
ResearchBunny Logo
A high-performance speech neuroprosthesis

Medicine and Health

A high-performance speech neuroprosthesis

F. R. Willett, E. M. Kunz, et al.

Intracortical microelectrode arrays decoded attempted speech from a participant with ALS, achieving a 9.1% WER on a 50-word vocabulary and 23.8% WER on a 125,000-word vocabulary—the first large‑vocabulary demonstration—at 62 words per minute, nearing conversational speed. The study also reveals spatially intermixed articulator tuning and persistent phoneme representations. Research conducted by Francis R. Willett, Erin M. Kunz, Chaofei Fan, Donald T. Avansino, Guy H. Wilson, Eun Young Choi, Foram Kamdar, Matthew F. Glasser, Leigh R. Hochberg, Shaul Druckmann, Krishna V. Shenoy, and Jaimie M. Henderson.... show more
Abstract
Speech brain–computer interfaces (BCIs) could restore rapid communication by decoding neural activity evoked by attempted speech into text or sound. Prior demonstrations have not reached accuracies sufficient for unconstrained sentences from large vocabularies. Here, using intracortical microelectrode arrays that record spiking activity, a participant with amyotrophic lateral sclerosis who cannot speak intelligibly achieved a 9.1% word error rate (WER) on a 50-word vocabulary (2.7× fewer errors than previous state-of-the-art) and a 23.8% WER on a 125,000-word vocabulary—the first demonstration, to our knowledge, of large-vocabulary decoding. Attempted speech was decoded at 62 words per minute (WPM), 3.4× faster than the previous record and approaching natural conversation speed (~160 WPM). We also show two encouraging aspects of the neural code for speech: spatially intermixed tuning to speech articulators enabling accurate decoding from a small cortical region, and a detailed articulatory representation of phonemes that persists years after paralysis. These results chart a feasible path toward restoring rapid communication to people with paralysis who can no longer speak.
Publisher
Nature
Published On
Aug 23, 2023
Authors
Francis R. Willett, Erin M. Kunz, Chaofei Fan, Donald T. Avansino, Guy H. Wilson, Eun Young Choi, Foram Kamdar, Matthew F. Glasser, Leigh R. Hochberg, Shaul Druckmann, Krishna V. Shenoy, Jaimie M. Henderson
Tags
brain–computer interface
speech decoding
intracortical microelectrode arrays
word error rate
large-vocabulary decoding
articulatory representation
amyotrophic lateral sclerosis
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny