Voice Print Recognition Check-in System Based on Resnet
DOI:
https://doi.org/10.54097/hset.v16i.2473Keywords:
Voiceprint recognition, Deep learning, Convolution network.Abstract
With the development of modern technology and the rise of artificial intelligence, the application scenarios of identity authentication technology are becoming more and more complex, especially the current situation of the spread of the novel coronavirus, the traditional identity authentication technology can no longer meet people's actual needs, and the society urgently needs a secure and convenient identity authentication technology. Voice print recognition technology is an identity recognition method that uses specific feature extraction methods to extract the features that can identify the speaker's identity from the original speech input, and then uses these features to identify the speaker's identity. Aiming at the above problems, this paper proposes a deep learning-based voicing recognition algorithm, which is based on the theoretical knowledge of deep learning. The research work includes the following aspects: providing convolutional network to extract features; The "speech feature template library" is established by massive training. Research on matching and recognition algorithm; Research on check-in system based on voice print recognition. Based on this algorithm, this paper designs and implements a voiceprint recognition check-in system with clear interactive interface. The system has the functions of adding members, refreshing the voice library, checking in with the voice print, clearing the results and so on. The system interface will display the data of the voice print library and the historical check-in record, as well as the current recognition results, accuracy and check-in time. The average recognition rate of the system is about 95%, which can meet the requirement of practical application.
Downloads
References
Rabiner, Lawrence, and Biinghwang Juang. "An introduction to hidden Markov models." IEEE ASSP Magazine 3.1 (1986): 4-16. .
Campbell W M. A SVM/HMM system for speaker recognition [C] / 2003 IEEE, International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP'03). IEEE, 2003, 2: II-209.
Matsui T, Furui S. Comparison of text-independent speaker recognition methods using VQ-distortion and discrete/continuous HMM's [J]. IEEE Transactions on speech and audio processing, 1994, 2(3): 456-459.
Wan, L., Wang, Q., Papir, A., & Moreno, I. L. (2017). Generalized end-to-end loss for speaker verification. ArXiv preprint arXiv: 1710.10467.
Kumar, R., Yeruva, V., & Ganapathy, S. (2018). On Convolutional LSTM Modeling for Joint Wake-Word Detection and Text Dependent Speaker Verification. Proc. Interspeech 2018, 1121-1125.
Lei Y, Burget L, Scheffer N. A noise robust i-vector extractor using vector taylor series for speaker recognition [C] /2013 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 2013: 6788-6791.
Tanprasert C, Wutiwiwatchai C, Sae-Tang S. Text-dependent speaker identification using neural network on distinctive Thai tone marks. InIJCNN'99. International Joint Conference on Neural Networks. Proceedings (Cat. No. 99CH36339) 1999 Jul 10 (Vol.5, pp.2950-2953). IEEE.
Downloads
Published
Issue
Section
License

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.







