Research on Object Detection Based on YOLO and Its Application in Tactile Paving Recognition
DOI:
https://doi.org/10.54097/wps8rx55Keywords:
YOLO Model; Tactile Paving Recognition; Object Detection.Abstract
The YOLO model, as an advanced real-time object detection algorithm, has made significant progress in the field of object detection since its inception. It achieves fast and accurate object detection by transforming the object detection task into a regression problem and using a single neural network to predict multiple bounding boxes and category probabilities. The YOLO model has significant advantages in processing speed, detection accuracy, and real-time performance, and is widely used in fields such as autonomous driving, intelligent monitoring, and medical image analysis. This paper adopts a tactile paving detection method based on YOLOv8 and CBAM attention mechanism, and improves the model's ability to extract tactile paving features by introducing CBAM attention mechanism and loss function. This method shows good adaptability when dealing with the tactile paving environment under different lighting and weather conditions, can effectively identify diverse obstacles, and provides a strong guarantee for the travel safety of the visually impaired. In conclusion, YOLO model has broad prospects in object detection and tactile paving detection. The tactile paving detection method based on YOLOv8 and CBAM attention mechanism adopted in this paper provides new ideas and methods for the development of tactile paving detection technology, and has important practical application value. In the future, with the continuous optimization of performance and technological innovation of YOLO series models, further improvement of hardware performance, and the fusion of YOLO and other sensor data to form a multimodal detection system, YOLO models are expected to play an important role in more practical application scenarios.
Downloads
References
[1] Wei Y, Tian Q, Guo J, et al. multi-vehicle detection algorithm through combining Harr and HOG features [J]. Mathematics and Computers in Simulation, 2019, 155:130-145.
[2] Girshick R, Donahue J, Darrell T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]. Proceedings of the IEEE conference on computer vision and pattern recognition. 2014: 580-587.
[3] Girshick R. Fast r-cnn[C]. Proceedings of the IEEE international conference on computer vision. 2015: 1440-1448.
[4] Ren S, He K, Girshick R, et al. Faster r-cnn: Towards real-time object detection with region proposal networks[J]. Advances in neural information processing systems, 2015, 28.
[5] Redmon J, Divvala S, Girshick R, et al. You only look once: Unified, real-time object detection[C]. Proceedings of the IEEE conference on computer vision and pattern recognition. 2016: 779-788.
[6] Redmon J, Farhadi A. YOLO9000: better, faster, stronger[C]. Proceedings of the IEEE conference on computer vision and pattern recognition. 2017: 7263-7271.
[7] Redmon J, Farhadi A. YOLOv3: An incremental improvement[J]. arXiv preprintarXiv: 180402767, 2018.
[8] Bochkovskiy A, Wang C-Y, Liao H-Y M. YOLOv4: Optimal speed and accuracy of object detection[J]. arXiv preprint arXiv:200410934, 2020.
[9] Cong Z, Li X. Track obstacle detection algorithm based on YOLOv3[C]. 202013th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI). IEEE, 2020: 12-17.
[10] Benjumea A, Teeti I, Cuzzolin F, et al. YOLO-Z: Improving small object detection in YOLOv5 for autonomous vehicles[J]. arXiv preprint arXiv:211211798, 2021.
[11] Guan L, Jia L, Xie Z, et al. A lightweight framework for obstacle detection in the railway image based on fast region proposal and improved YOLO-tiny network[J]. IEEE Transactions on Instrumentation and Measurement, 2022, 71: 1-16.
[12] He C, Saha P. Investigating YOLO Models Towards Outdoor Obstacle Detection for Visually Impaired People[J]. arXiv preprint arXiv:231207571, 2023.
[13] Yuki I, Premachandra C, Sumathipala S, et al. HSV conversion based tactile paving detection for developing walking support system to visually handicapped people[C]. 2019 IEEE 23rd International Symposium on Consumer Technologies (ISCT). IEEE, 2019: 138-142.
[14] Yamanaka Y, Takaya E, Kurihara S. Tactile Tile Detection Integrated with Ground Detection using an RGB-Depth Sensor[C]. ICAART (2). 2020: 750-757.
[15] Ito Y, Premachandra C, Sumathipala S, et al. Tactile paving detection by dynamic thresholding based on HSV space analysis for developing a walking support system[J]. IEEE Access, 2021, 9: 20358-20367.
[16] Aktaş A, Doğan B, Demir Ö. Tactile paving surface detection with deep learning methods[J]. Journal of the Faculty of Engineering and Architecture of Gazi University, 2020, 35(3): 1685-700.
[17] Chen Qianqian, Wang Huan, Zhu Min Tactile paving and zebra crossing detection based on YOLOv5 [J] Information Technology and Informatization, 2022, (07): 10-14.
[18] Yang K, Bergasa L M, Romera E, et al. Semantic perception of curbs beyond travers ability for real-world navigation assistance systems[C]. 2018 IEEE International Conference on Vehicular Electronics and Safety (ICVES). IEEE,2018: 1-7.
[19] Cao Z, Xu X, Hu B, et al. Rapid detection of blind roads and crosswalks by using a light weight semantic segmentation network[J]. IEEE Transactions on Intelligent Transportation Systems, 2020, 22(10): 6188-6197.
[20] Xia Y, Li Y, Ye Q, et al. Image segmentation for blind lanes based on improved SegNet model[J]. Journal of Electronic Imaging, 2023, 32(1): 013038-013038.
Downloads
Published
Issue
Section
License
Copyright (c) 2024 Highlights in Science, Engineering and Technology

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.







