Mon-2-9-8 Towards an ASR error robust Spoken Language Understanding System

Weitong Ruan(Amazon Alexa), Yaroslav Nechaev(Amazon Alexa), Luoxin Chen(Amazon Alexa), Chengwei Su(Amazon Alexa) and Imre Kiss(Amazon Alexa)
Abstract: A modern Spoken Language Understanding (SLU) system usually contains two sub-systems, Automatic Speech Recognition (ASR) and Natural Language Understanding (NLU), where ASR transforms voice signal to text form and NLU provides intent classification and slot filling from the text. In practice, such decoupled ASR/NLU design facilitates fast model iteration for both components. However, this makes downstream NLU susceptible to errors from the upstream ASR, causing significant performance degradation. Therefore, dealing with such errors is a major opportunity to improve overall SLU model performance. In this work, we first propose a general evaluation criterion that requires an ASR error robust model to perform well on both transcription and ASR hypothesis. Then robustness training techniques for both classification task and NER task are introduced. Experimental results on two datasets show that our proposed approaches improve model robustness to ASR errors for both tasks.
Student Information

Student Events

Travel Grants