An Analysis of Knowledge Distillation Methods for Sentiment Analysis: Limitations and Future Prospects

Authors

  • Qirui Li

DOI:

https://doi.org/10.54097/tgjwa583

Keywords:

Knowledge Distillation, Sentiment Analysis, Cross-Modal Distillation, Graph-Based Distillation, Ensemble Knowledge Distillation.

Abstract

With the advancement of sentiment analysis, the integration of multi-teacher and multi-modal models has grown increasingly prominent. However, the direct fusion of these diverse models poses significant challenges. This paper presents a comprehensive review of knowledge distillation techniques designed to address these challenges. This paper systematically surveys several key approaches: for multi-teacher scenarios, this paper examines ensemble knowledge distillation (EKD) as a method for effective model compression; for multi-modal applications, this paper explores cross-modal distillation (CMD) and graph-based distillation to effectively handle the intrinsic heterogeneity between data modalities. Besides, this paper analysis identifies critical limitations in current methods, including the static nature of knowledge extraction and constraints of graph distillation approaches. To address these shortcomings, this paper proposes several forward-looking research directions, such as incorporating meta-learning for dynamic weight adjustment, using attention mechanisms for adaptive teacher selection, and replacing conventional Graph Neural Networks (GNNs) with Hypergraph Neural Networks (HNNs) to capture higher-order inter-modal relationships. By critically examining existing methods and a clear outlook on future trends, this paper provides valuable insights and a roadmap for researchers in the field.

Downloads

Download data is not yet available.

References

[1] M. Kamyab, G. Liu, Michael. A, Attention-Based CNN and Bi-LSTM Model Based on TF-IDF and GloVe Word Embedding for Sentiment Analysis. Appl. Sci. 11, 11255 (2021). DOI: https://doi.org/10.3390/app112311255

[2] D. Kounadis, O. Schrüfer, A. Derington, H. Wierstorf, F. Eyben, F. Burkhardt, B. Schuller, wav2vec 2.0: A framework for self-supervised learning of speech representations. Advances in neural information processing systems. 33, 12449-12460, (2020).

[3] G. Hinton, O. Vinyals, J. Dean, Distilling the Knowledge in a Neural Network. NIPS workshop. (2015).

[4] Y. Tsai, S. Bai, P. Liang, J. Kolter, L. Morency, R. Salakhutdinov, Multimodal Transformer for Unaligned Multimodal Language Sequences. In proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, July (2019), 6558-6569. DOI: https://doi.org/10.18653/v1/P19-1656

[5] D. Kim, P. Kang, Cross-modal distillation with audio–text fusion for fine-grained emotion classification using BERT and Wav2vec 2.0, Neurocomputing. 506, 168-183 (2022). DOI: https://doi.org/10.1016/j.neucom.2022.07.035

[6] Y. Li, Y. Wang, Z. Cui, Decoupled Multimodal Distilling for Emotion Recognition. In proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, September 13 (2025), 6631-6640. DOI: https://doi.org/10.1109/CVPR52729.2023.00641

[7] Z. Luo, J. Hsieh, L. Jiang, J. Niebles, F. Li, Graph Distillation for Action Detection with Privileged Modalities. In proceedings of the European Conference on Computer Vision, Sep (2018), 166-183. DOI: https://doi.org/10.1007/978-3-030-01264-9_11

[8] H. Li, H. Zhong, C. Xu, X. Liu, G. Wen, L. Liu, Global distilling framework with cognitive gravitation for multimodal emotion recognition. Neurocomputing, 622, 129306 (2025). DOI: https://doi.org/10.1016/j.neucom.2024.129306

[9] K. Huang, T. Feng, Y. Fu, T. Hsu, P. Yen, W. Tseng, K. Chang, H. Lee, Ensemble Knowledge Distillation of Self-Supervised Speech Models. ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing, Rhodes Island, Greece, June 4 (2023), 1-5. DOI: https://doi.org/10.1109/ICASSP49357.2023.10096445

[10] Y. Feng, H. You, Z. Zhang, R. Ji, Y. Gao, Hypergraph Neural Networks, In proceedings of the AAAI Conference on Artificial Intelligence. Jul 17 (2019), 3558-3565. DOI: https://doi.org/10.1609/aaai.v33i01.33013558

Downloads

Published

29-01-2026

Issue

Section

Articles

How to Cite

Li, Q. (2026). An Analysis of Knowledge Distillation Methods for Sentiment Analysis: Limitations and Future Prospects. Academic Journal of Science and Technology, 19(2), 408-417. https://doi.org/10.54097/tgjwa583