Mon-1-9-1 Toward Silent Paralinguistics: Speech-to-EMG – Retrieving Articulatory Muscle Activity from Speech

Catarina Botelho(INESC-ID/Instituto Superior Técnico, University of Lisbon, Portugal), Lorenz Diener(University of Bremen), Dennis Küster(Cognitive Systems Lab (CSL), University of Bremen), Kevin Scheck(Cognitive Systems Lab (CSL), University of Bremen), Shahin Amiriparian(University of Augsburg), Björn Schuller(University of Augsburg / Imperial College London), Tanja Schultz(Universität Bremen), Alberto Abad(INESC-ID/IST) and Isabel Trancoso(INESC-ID / IST Univ. Lisbon)
Abstract: Electromyographic (EMG) signals recorded during speech production encode information on articulatory muscle activity and also on the facial expression of emotion, thus representing a speech-related biosignal with strong potential for paralinguistic applications. In this work, we estimate the electrical activity of the muscles responsible for speech articulation directly from the speech signal. To this end, we first perform a neural conversion of speech features into electromyographic time domain features, and then attempt to retrieve the original EMG signal from the time domain features. We propose a feed forward neural network to address the first step of the problem (speech features to EMG features) and a neural network composed of a convolutional block and a bidirectional long short-term memory block to address the second problem (true EMG features to EMG signal). We observe that four out of the five originally proposed time domain features can be estimated reasonably well from the speech signal. Further, the five time domain features are able to predict the original speech-related EMG signal with a concordance correlation coefficient of 0.663. We further compare our results with the ones achieved on the inverse problem of generating acoustic speech features from EMG features.
Student Information

Student Events

Travel Grants