A comparison of EMG-to-Speech Conversion for Isolated and Continuous Speech

Konferenz: Speech Communication - 13. ITG-Fachtagung Sprachkommunikation
10.10.2018 - 12.10.2018 in Oldenburg, Deutschland

Tagungsband: ITG-Fb. 282: Speech Communication

Seiten: 5Sprache: EnglischTyp: PDF

Persönliche VDE-Mitglieder erhalten auf diesen Artikel 10% Rabatt

Autoren:
Diener, Lorenz; Bredehoeft, Sebastian; Schultz, Tanja (Cognitive Systems Lab, University of Bremen, Bremen, Germany)

Inhalt:
This paper presents initial results of performing EMG-to-Speech conversion within our new EMG-to-Speech corpus. This new corpus consists of parallel facial array sEMG and read audible speech signals recorded from multiple speakers. It contains different styles of utterances – continuous sentences, isolated words, and isolated consonant-vowel combinations – which allows us to evaluate the performance of EMG-to-Speech conversion when trying to convert these different styles of utterance as well as the effect of training systems on one style to convert another. We find that our system deals with isolated-word/consonant-vowel utterances better than with continuous speech. We also find that it is possible to use a model trained on one style to convert utterances from another – however, performance suffers compared to training within that style, especially when going from isolated to continuous speech.