Multi-classification of Human Action Based on ResNet
DOI:
https://doi.org/10.54097/hset.v23i.3203Keywords:
Human Action classification, machine learning, ResNet.Abstract
With the development of machine learning and other related technologies, more and more excellent research on human action recognition and classification has been proposed, which significantly promotes the application of this technology in the actual situation. This paper mainly focuses on the characteristics of ResNet18, ResNet50, ResNet101, and ResNet152 in 15 human action recognition and classification tasks respectively. First, adjust and enhance the sample images of all training sets and test sets, and adjust the overall parameters according to the characteristics of the input image, so that all images can be normalized and high quality can be guaranteed at the same time; Then use the above four ResNet models to study and test with an epoch of 200, and explore the characteristics of the four models in this project by comparing the accuracy, loss, confusion matrix graphic, F1-score, and the number of epoch experienced in reaching the steady state of accuracy. The results show that in the overall effect, with the increase of the number of layers of the ResNet models, the accuracy changes in a positive correlation, and the epoch number experienced by the accuracy reaching the steady state changes in a negative correlation.
Downloads
References
Lin, C. -H., et al. "A Lightweight Fine-Grained Action Recognition Network for Basketball Foul Detection," 2021 IEEE International Conference on Consumer Electronics-Taiwan (ICCE-TW), 2021.
Mukhanbet, A. A., et al. "Hybrid Architecture of Face and Action Recognition Systems for Proctoring on a Graphic Processor," 2021 IEEE International Conference on Smart Information Systems and Technologies (SIST), 2021.
Sarhan, N., et al. "Transfer Learning for Videos: From Action Recognition To Sign Language Recognition," 2020 IEEE International Conference on Image Processing (ICIP), 2020.
Widyadhana, I. H., et al. "Development of Video Based Action Recognition System For Item Taking From Shelf," 2021 8th International Conference on Advanced Informatics: Concepts, Theory and Applications (ICAICTA), 2021.
Liu, J., et al. "Action Recognition by Multiple Features and Hyper-Sphere Multi-class SVM," 2010 20th International Conference on Pattern Recognition, 2010.
Wang, P., et al. "Application of K-Nearest Neighbor (KNN) Algorithm for Human Action Recognition," 2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), 2021.
Chuan, C., et al. "Human Action Recognition Based on Action Forests Model Using Kinect Camera," 2016 30th International Conference on Advanced Information Networking and Applications Workshops (WAINA), 2016.
Ijjina, E. P., et al. "Human Action Recognition Based on Recognition of Linear Patterns in Action Bank Features Using Convolutional Neural Networks," 2014 13th International Conference on Machine Learning and Applications, 2014.
Yang, Y., et al. "Human Action Recognition Based on Skeleton and Convolutional Neural Network," 2019 Photonics & Electromagnetics Research Symposium - Fall (PIERS - Fall), 2019.
He, K., Zhang, X., Ren, S., et al. "Deep residual learning for image recognition," Proceedings of the IEEE conference on computer vision and pattern recognition. 2016
Kaggle, "Human Action Recognition (HAR) Dataset," 2022, https://www.kaggle.com/datasets/meetnagadia/human-action-recognition-har-dataset.
Yu, Q., et al. "Semantic segmentation of intracranial hemorrhages in head CT scans." 2019 IEEE 10th International Conference on Software Engineering and Service Science (ICSESS). IEEE, 2019.
Farooq, M., et al. "Covid-resnet: A deep learning framework for screening of covid19 from radiographs." arXiv preprint arXiv:2003.14395, 2020.
Ariza-Lopez, F. J., et al. "Complete Control of an Observed Confusion Matrix," IGARSS 2018 - 2018 IEEE International Geoscience and Remote Sensing Symposium, 2018.
Downloads
Published
Issue
Section
License

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.







