Joint Modelling of Slot Filling and Intent Detection in Constrained Resource Scenarios
DOI:
https://doi.org/10.54097/fcis.v5i2.13124Keywords:
Slot Filling, Intent Detection, Graph Neural NetworksAbstract
In the era of pervasive smart devices, natural language understanding (NLU) holds a pivotal role for facilitating intelligent interactions and decision-making. Core to NLU are slot filling and intent recognition, essential tasks for comprehending user input. While joint modelling of these tasks has gained prominence, the challenges of realizing efficient joint models on resource-constrained devices have emerged as significant. These devices possess limited computational capacity and real-time requirements, necessitating lightweight and efficient models. In this study, we explore the design of a resource-efficient joint model for slot filling and intent recognition. Through leveraging BERT, , graph neural networks, and mask mechanisms, our model achieves the dual goals of semantic slot prediction and intent classification. We focus on model design, training, and real-time inference, aiming to contribute to the paradigm of resource-constrained natural language understanding. Our investigation demonstrates the efficacy of our approach, even when working with a reduced dataset, underscoring the model's applicability to real-world scenarios with limited resources.
Downloads
References
Jeong, M., & Lee, G. G. (2008). Triangular-chain conditional random fields. IEEE Transactions on Audio, Speech, and Language Processing, 16(7), 1287–1302.
Guo, D., Tur, G., Yih, W. T., & Zweig, G. (2014). Joint semantic utterance classification and slot filling with recursive neural networks. Proceedings of the IEEE Spoken Language Technology Workshop (SLT'14, pp. 554–559.
Raedt, L. D., Hammer, B., Hitzler, P., & Maass, W. (2008). Recurrent Neural Networks - Models, Capacities, and Applications. In: Internationales Begegnungs- und Forschungszentrum fuer Informatik (IBFI). Schloss Dagstuhl, Germany. 08041
Luo, X., Zhou, W., Wang, W., Zhu, Y., & Deng, J. (2018). Attention-based relation extraction with bidirectional gated recurrent unit and highway network in the analysis of geological data. IEEE Access, 6, 5705–5715.
Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735–1780.
Tang, H., Ji, D., & Zhou, Q. (2020). End-to-end masked graph-based CRF for joint slot filling and intent detection. Neurocomputing, 413, 348–359.
Wang, S., et al. (2023) GNN-SL: Sequence Labeling Based on Nearest Examples via GNN. In: Findings of the Association for Computational linguistics (ACL 2023). Toronto. 12679-12692.
Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 1: 4171-4186.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., & Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30.
Hu, Z., Dong, Y., Wang, K., & Sun, Y. (2020). Heterogeneous graph transformer. Proceedings of the Web Conference 2020. 2704–2710.


