In Spoken Language Understanding (SLU), the ability to detect out-of-domain (OOD) input dialog plays a very important role (e.g., voice assistance and chatbot systems). However, most of the existing OOD detection methods rely heavily on manually labeled OOD data. Manual labeling of the OOD data for a dynamically changing and evolving area is time-consuming and not immediately possible. It limits the feasibility of these models in practical applications. So, to solve this problem, we are considering the scenario of having no OOD labeled data (i.e., zero-shot learning). To achieve this goal, we have used the intent focused semantic parsing, extracted with the help of Transformer-based techniques [e.g., BERT (Devlin et al., 2018)]. The two main components of the intent-focused semantic parsing are - (a) the sentence-level intents and (b) token-level intent classes, which show the relation of slot tokens with intent classes. Finally, we combine both information and use a One Class Neural Network (OC-NN) based zero-shot classifier. Our devised system has shown better results compared to the state-of-the-art on four publicly available datasets.
CITATION STYLE
Kumar, N., & Baghel, B. K. (2021). Intent Focused Semantic Parsing and Zero-Shot Learning for Out-of-Domain Detection in Spoken Language Understanding. IEEE Access, 9, 165786–165794. https://doi.org/10.1109/ACCESS.2021.3133657
Mendeley helps you to discover research relevant for your work.