A short (~100p) book-length overview of 2017-era progress on Silent speech interface technology, across many modalities. Includes an overview of how human speech production works, from neurophysiology through to facial actuation.
This book was a nice way to get an overview of the different modalities, but its portrayal of their relative progress seems already outdated / untrustworthy: the deep learning revolution was only just beginning to enter this research space when this book was published. For instance, this book doesn’t (and couldn’t) include AlterEgo (2018, SEMG + DL) or Prajwal et al’s 2022 transformer-based Visual speech recognition work, which seem like an enormous leap over the rates presented.
At the end of the book, I found myself interested in more coverage of “non-audible murmur” microphones (see page 43)—they might be quiet enough for Tools for thought purposes. See e.g. Fukumoto, M. (2018). SilentVoice: Unnoticeable Voice Input by Ingressive Speech. Proceedings of the 31st Annual ACM Symposium on User Interface Software and Technology, 237–246