Part-of-Speech Tagging for Traditional Chinese Medicine Clinical Pulse Feeling and Palpation Information Based on Hidden Markov Model Prompt-Tuning

Authors

  • Linfeng Dai

DOI:

https://doi.org/10.54097/rp1kne02

Keywords:

Prompt Tuning, Part-of-Speech Tagging, Hidden Markov Model, Pre-training Language Model, Traditional Chinese Medicine Clinical Pulse Feeling and Palpation Information.

Abstract

 This paper proposes a Traditional Chinese Medicine Clinical Pulse Feeling and Palpation Information part-of-speech tagging method based on Hidden Markov Model Prompt-Tuning, aimed at addressing the challenges of scarce annotated samples and the issue of words with multiple parts of speech in Traditional Chinese Medicine Clinical Pulse Feeling and Palpation Information part-of-speech tagging tasks. By designing specific prompt templates, this method transforms the downstream task into a cloze task for the pretrained language model during its pretraining phase, thereby bridging the gap between pretraining and fine-tuning objectives and enhancing the knowledge representation ability of the pretrained language model for downstream tasks. This approach also integrates the part-of-speech information of adjacent words into the prompt templates and uses the Hidden Markov Model to model the dependencies between adjacent words' parts of speech, effectively improving the overall performance of the model in the Traditional Chinese Medicine Clinical Pulse Feeling and Palpation Information part-of-speech tagging task, particularly in identifying words with multiple parts of speech. Furthermore, empirical evidence demonstrates the outstanding performance of the proposed method in this part-of-speech tagging task, achieving an F1 score of up to 85.52%.

Downloads

Download data is not yet available.

References

. Zheng, Hongxin. 2016. Basic Theories of Traditional Chinese Medicine. China Press of Traditional Chinese Medicine.

. Li, Candong. 2016. Diagnostics of Traditional Chinese Medicine. China Press of Traditional Chinese Medicine.

. Brill E. Transformation-based error-driven learning and natural language processing: A case study in part-of-speech tagging[J]. Computational linguistics, 1995, 21(4): 543-565.

. Wang Y, Jiang W, Jiang Y, et al. On Corpus Construction and Joint Modeling for Clinical Pulse Feeling and Palpation Information Extraction and Lexical Analysis of Traditional Chinese Medicine[C]//Proceedings of the 22nd Chinese National Conference on Computational Linguistics. 2023: 352-363.

. Zhang H, Yang Z, Fu Y, et al. BERT 4EVER@ EvaHan 2022: Ancient Chinese Word Segmentation and Part-of-Speech Tagging based on Adversarial Learning and Continual Pre-training[C]//Proceedings of the Second Workshop on Language Technologies for Historical and Ancient Languages. 2022: 150-154.

. Pathak D, Nandi S, Sarmah P. Part-of-speech tagger for Assamese using Ensembling approach[J]. ACM Transactions on Asian and Low-Resource Language Information Processing, 2023, 22(10): 1-22.

. Han J, Zhao S, Cheng B, et al. Generative prompt tuning for relation classification[J]. arXiv preprint arXiv:2210.12435, 2022.

. Chen, Deguang, Ma, Jinlin, Ma, Ziping, et al. 2021. A Survey of Pretraining Techniques in Natural Language Processing. Journal of Computer Science and Exploration, 15(8), 1359.

. Liu P, Yuan W, Fu J, et al. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing[J]. ACM Computing Surveys, 2023, 55(9): 1-35.

. Tian Y, Guo Y. Ancient chinese word segmentation and part-of-speech tagging using data augmentation[C]//Proceedings of the Second Workshop on Language Technologies for Historical and Ancient Languages. 2022: 146-149.

. Devlin J, Chang M W, Lee K, et al. Bert: Pre-training of deep bidirectional transformers for language understanding[J]. arXiv preprint arXiv:1810.04805, 2018.

. Wang Y, Li K, Jiang Y, et al. Four Diagnostic Description Extraction in Clinical Records of Traditional Chinese Medicine with Batch Data Oversampling[C]//Proceedings of the 21st Chinese National Conference on Computational Linguistics. 2022: 611-622.

. Lewis M, Liu Y, Goyal N, et al. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension[J]. arXiv preprint arXiv:1910.13461, 2019.

Downloads

Published

27-03-2024

Issue

Section

Articles

How to Cite

Part-of-Speech Tagging for Traditional Chinese Medicine Clinical Pulse Feeling and Palpation Information Based on Hidden Markov Model Prompt-Tuning. (2024). Academic Journal of Science and Technology, 10(1), 437-441. https://doi.org/10.54097/rp1kne02

Similar Articles

1-10 of 774

You may also start an advanced similarity search for this article.