Bidirectional Deep Learning Model based on Attention Mechanism in English Cloze Tests

Authors

  • Haoshan Yuan

DOI:

https://doi.org/10.54097/4c4qk780

Keywords:

English Cloze Tests; Deep Learning; BiGRU; Attention Mechanisms.

Abstract

English cloze tests, as a common form of language assessment, aim to evaluate learners' comprehensive understanding of context, vocabulary, and grammar. However, the complex contextual relationships and long-distance dependencies within the questions pose significant challenges for machine learning models. Traditional rule-based or statistical methods struggle to effectively capture the intricate contextual information present in sentences. This study aims to develop a deep learning-based English cloze test answering system to enhance students' ability to tackle such questions. To address the limitations of traditional methods in handling complex contexts and long-distance dependencies, a model that combines bidirectional gated recurrent units (BiGRU) and attention mechanisms is proposed. This model is better equipped to capture the surrounding context of sentences and dynamically adjust attention to accurately predict the missing words. Additionally, integrating the embedding layer with BiGRU and attention mechanisms further improves model performance. Testing results based on the Children’s Book Test dataset are highly promising. Our model excels in key metrics such as accuracy, recall, F1 score, and Cohen's Kappa, achieving scores of 77.5%, 77.5%, 0.758, and 0.7649, respectively. Compared to traditional models, our approach demonstrates clear advantages in handling complex contexts and long sentences. This research provides new technical support for developing more intelligent English learning systems.

Downloads

Download data is not yet available.

References

[1] Li, Q., Liao, W., & Meng, J. (2022). A dual-channel DAC-RNN text classification model based on attention mechanism. Computer Engineering and Applications, 58(16), 157-163.

[2] Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.

[3] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you need. In Advances in Neural Information Processing Systems (pp. 5998-6008).

[4] Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.

[5] Zhou, P., Shi, W., Tian, J., Qi, Z., Li, B., Hao, H., & Xu, B. (2016). Attention-based bidirectional long short-term memory networks for relation classification. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) (pp. 207-212).

[6] Sun, C., Huang, Y., Qiu, X., & Huang, X. (2019). How to fine-tune BERT for text classification? In China National Conference on Chinese Computational Linguistics (pp. 194-206). Springer, Singapore.

[7] Zheng, Z. (2018). A study on the Word2Vec embedding model (Master's thesis). Liaoning Technical University. Yang, P., & Dong, W. (2020). A Chinese named entity recognition method based on BERT embedding. Computer Engineering, 46(04), 40-45+52.

[8] Hu, Q., Li, Q., & Wang, S. (2021). A comparative study of word embedding models in text sentiment analysis. Computer Knowledge and Technology, 17(36), 109-111.

[9] Krautscheid, L., King, R., Lembke, K., et al. (2024). Lecture capture strategies with embedded retrieval practices: Relationship with academic performance. Journal of Educational Technology Systems, 53(1), 30-45.

[10] Xie, T., Yang, J., & Liu, H. (2020). A Chinese entity recognition model based on BERT-BiLSTM-CRF. Computer Systems Applications, 29(07), 48-55.

[11] Wang, W., Sun, Y., Qi, Q., et al. (2019). A text sentiment classification model based on BiGRU-attention neural network. Computer Applications Research, 36(12), 3558-3564.

[12] Zia, S., Azhar, M., Lee, B., Tahir, A., Ferzund, J., Murtaza, F., & Ali, M. (2023). Recognition of printed Urdu script in Nastaleeq font using a CNN-BiGRU-GRU based encoder-decoder framework. Intelligent Systems with Applications, 18.

[13] Xu, K., Wang, S., Li, Z. C., et al. (2020). Biomedical named entity recognition based on BiGRU network combining multi-headed attention mechanism. Computer Applications and Software, 37(05), 151-155+232.

[14] Liu, J., & Gu, F. Y. (2022). Unbalanced text sentiment analysis of online public opinion based on a hybrid method of BERT and BiLSTM. Journal of Intelligence, 41(04), 104-110.

[15] Zhao, H., Fu, Z., & Zhao, F. (2022). A study on sentiment analysis of Weibo based on BERT and hierarchical attention. Computer Engineering and Applications, 58(05), 156-162.

[16] Zhao, H., Fu, Z., & Zhao, F. (2022). A study on sentiment analysis of Weibo based on BERT and hierarchical attention. Computer Engineering and Applications, 58(05), 156-162.

[17] Zou, Z., Guo, H., & Gao, Y. (2007). A method for segmenting English strings. Computer Applications Research, (07), 52-54.

[18] Rao, H. (2019). Linking and discourse marking: A dual evolutionary model exemplified by "including." Chinese Language, (03), 311-318+383.

Downloads

Published

15-12-2024

How to Cite

Yuan, H. (2024). Bidirectional Deep Learning Model based on Attention Mechanism in English Cloze Tests. Highlights in Science, Engineering and Technology, 122, 77-87. https://doi.org/10.54097/4c4qk780