Document Type : Original/Review Paper

Authors

1 Department of Electrical Engineering, Faculty of Engineering, Bu-Ali Sina University, Hamedan, Iran.

2 Department of Computer Engineering, Faculty of Engineering, Bu-Ali Sina University, Hamedan, Iran.

3 School of Computer and Electrical Engineering, University of Tehran, Tehran, Iran.

Abstract

Providing a dataset with a suitable volume and high accuracy for training deep neural networks is considered to be one of the basic requirements in that a suitable dataset in terms of the number and quality of images and labeling accuracy can have a great impact on the output accuracy of the trained network. The dataset presented in this article contains 3000 images downloaded from online Iranian car sales companies, including Divar and Bama sites, which are manually labeled in three classes: car, truck, and bus. The labels are in the form of 5765 bounding boxes, which characterize the vehicles in the image with high accuracy, ultimately resulting in a unique dataset that is made available for public use.The YOLOv8s algorithm, trained on this dataset, achieves an impressive final precision of 91.7% for validation images. The Mean Average Precision (mAP) at a 50% threshold is recorded at 92.6%. This precision is considered suitable for city vehicle detection networks. Notably, when comparing the YOLOv8s algorithm trained with this dataset to YOLOv8s trained with the COCO dataset, there is a remarkable 10% increase in mAP at 50% and an approximately 22% improvement in the mAP range of 50% to 95%.

Keywords

Main Subjects

[1] Buch, N., S.A. Velastin, and J. Orwell, "A review of computer vision techniques for the analysis of urban traffic," IEEE Transactions on Intelligent Transportation Systems, vol. 12, no. 3, pp. 920-939, 2011.
 
[2] Yang, Z. and L.S. Pun-Cheng, "Vehicle detection in intelligent transportation systems and its applications under varying environments: A review," Image and Vision Computing, vol. 69, pp. 143-154, 2018.
 
[3] Reis, D., Kupec, J., Hong, J., & Daoudi, A., "Real-Time Flying Object Detection with YOLOv8," arXiv preprint arXiv:2305.09972, 2023.
 
[4] Girshick, R., Donahue, J., Darrell, T., & Malik, J., "Rich feature hierarchies for accurate object detection and semantic segmentation," in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 580-587, 2014.
 
[5] Du, J., "Understanding of object detection based on CNN family and YOLO," in Journal of Physics: Conference Series, IOP Publishing, 2018.
 
[6] Ren, S., He, K., Girshick, R., & Sun, J., "Faster R-CNN: Towards real-time object detection with region proposal networks," Advances in Neural Information Processing Systems, vol. 28, 2015.
 
[7] Girshick, R., "Fast R-CNN," in Proceedings of the IEEE International Conference on Computer Vision, 2015.
 
[8] He, K., Gkioxari, G., Dollár, P., & Girshick, R., "Mask R-CNN," in Proceedings of the IEEE International Conference on Computer Vision, pp. 2961-2969, 2017.
 
[9] Redmon, J., Divvala, S., Girshick, R., & Farhadi, A., "You Only Look Once: Unified, Real-Time Object Detection," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779-788, 2016.
 
[10] Redmon, J. and A. Farhadi, "YOLO9000: better, faster, stronger," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
 
[11] Redmon, J. and A. Farhadi, "YOLOv3: An incremental improvement," arXiv preprint arXiv:1804.02767, 2018.
[12] Bochkovskiy, A., C.-Y. Wang, and H.-Y.M. Liao, "YOLOv4: Optimal speed and accuracy of object detection," arXiv preprint arXiv:2004.10934, 2020.
 
[13] Nasehi, M., Ashourian, M., & Emami, H. "Vehicle Type, Color and Speed Detection Implementation by Integrating VGG Neural Network and YOLO algorithm utilizing Raspberry Pi Hardware, " Journal of AI and Data Mining, pp.579-588, 2022.
 
[14] Team, M., "YOLOv6: A fast and accurate target detection framework is open source," June 23, 2022. [Online].Available: https://github.com/meituan/YOLOv6.
 
 [15] Wang, C.Y., Bochkovskiy, A., Liao, H.Y., "YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors," in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2023, pp. 7464-7475.
 
[16] Asgarian Dehkordi, R., and H. Khosravi. "Vehicle type recognition based on dimension estimation and bag of word classification." Journal of AI and Data Mining, vol. 8, no. 3, pp. 427-438, 2020.
 
[17] Oltean, G., Florea, C., Orghidan, R., & Oltean, V., "Towards real time vehicle counting using YOLO-tiny and fast motion estimation," in 2019 IEEE 25th International Symposium for Design and Technology in Electronic Packaging (SIITME) IEEE, Cluj-Napoca, Romania, October 2019, pp. 240-243.
 
[18] Rahman, Z., A.M. Ami, and M.A. Ullah, "A real-time wrong-way vehicle detection based on YOLO and centroid tracking," in 2020 IEEE Region 10 Symposium (TENSYMP), IEEE, Dhaka, Bangladesh, 2020, pp. 916-920.
 
[19] Al-qaness, M. A., Abbasi, A. A., Fan, H., Ibrahim, R. A., Alsamhi, S. H., & Hawbani, A., "An improved YOLO-based road traffic monitoring system," Computing, vol. 103, no. 2, pp. 211-230, 2021.
 
[20] Kim, J.-a., J.-Y. Sung, and S.-h. Park, "Comparison of Faster-RCNN, YOLO, and SSD for real-time vehicle type recognition," in 2020 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), IEEE, Seoul, Korea (South) 2020, pp. 1-4.
 
[21] Zhu, E., M. Xu, and D.C. Pi, "Vehicle Type Recognition Algorithm Based on Improved Network in Network," Complexity, pp. 1-10, 2021.
 
[22] Sang, J., Wu, Z., Guo, P., Hu, H., Xiang, H., Zhang, Q., & Cai, B., "An improved YOLOv2 for vehicle detection," Sensors, vol. 18, p. 4272, 2018.
 
[23] Gholamalinejad, H., and Hossein Khosravi. "Irvd: A large-scale dataset for classification of iranian vehicles in urban streets." Journal of AI and Data Mining, vol.9, no. 1, pp. 1-9, 2021.
 
[24] Carrasco, D. P., Rashwan, H. A., García, M. Á., & Puig, D., "T-YOLO: Tiny vehicle detection based on YOLO and multi-scale convolutional neural networks," IEEE Access, vol. 11, pp. 22430-22440, 2021.
 
[25] Miao, Y., Liu, F., Hou, T., Liu, L., & Liu, Y., "A nighttime vehicle detection method based on YOLO v3," in 2020 Chinese Automation Congress (CAC), IEEE, Shanghai, China, November 2020, pp. 6617-6621.
 
[26] Huang, S., Y. He, and X.-a. Chen, "M-YOLO: A Nighttime Vehicle Detection Method Combining MobileNet v2 and YOLO v3," in Journal of Physics: Conference Series, IOP Publishing, Shanghai, China, 2021, pp. 6617-6621.
 
[27] Goel, S., Baghel, A., Srivastava, A., Tyagi, A., & Nagrath, P., "Detection of emergency vehicles using modified YOLO algorithm," in Intelligent Communication, Control and Devices: Proceedings of ICICCD 2018, Springer Singapore, 2020, pp. 671-687.
 
[28] Baghel, A., Srivastava, A., Tyagi, A., Goel, S., & Nagrath, P., "Analysis of Ex-YOLO algorithm with other real-time algorithms for emergency vehicle detection," in Proceedings of First International Conference on Computing, Communications, and Cyber-Security (IC4S 2019), Springer Singapore, 2020, pp. 607-618.
 
[29] Geiger, A., P. Lenz, and R. Urtasun, "Are we ready for autonomous driving? the KITTI vision benchmark suite," in 2012 IEEE Conference on Computer Vision and Pattern Recognition, IEEE, Providence, RI, USA, 2012, pp. 3354-3361.
 
[30] Krause, J., Gebru, T., Deng, J., Li, L. J., & Fei-Fei, L., "Learning features and parts for fine-grained recognition," in 2014 22nd International Conference on Pattern Recognition, IEEE, Stockholm, Sweden, August 2014, pp. 26-33.
 
[31] Yang, L., Luo, P., Chang Loy, C., & Tang, X., "A large-scale car dataset for fine-grained categorization and verification," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 3973-3981.
 
[32] Dong, Z., Wu, Y., Pei, M., & Jia, Y., "Vehicle type classification using a semisupervised convolutional neural network," IEEE Transactions on Intelligent Transportation Systems, vol. 16, no. 4, pp. 2247-2256, 2015.
 
[33] Wen, L., Du, D., Cai, Z., Lei, Z., Chang, M. C., Qi, H., ... & Lyu, S., "UA-DETRAC: A new benchmark and protocol for multi-object detection and tracking," Computer Vision and Image Understanding, vol. 193, p. 102907, 2020.
 
[34] Siahkali, F., Alavi, S. A., & Masouleh, M. T., "SIVD: Dataset of Iranian Vehicles for Real-Time Multi-Camera Video Tracking and Recognition," in 2022 8th Iranian Conference on Signal Processing and Intelligent Systems (ICSPIS), IEEE, Behshahr, Iran, December 2022, pp. 1-7.
 
[35] Gholamalinejad, H., & Khosravi, H., "Irvd: A large-scale dataset for classification of Iranian vehicles in urban streets," Journal of AI and Data Mining, vol. 9, no. 1, pp. 1-9, 2021.
 
[36] Divar, https://divar.ir/, 2021.
 
[37] Bama, https://bama.ir/.
 
[38] Shorten, C. and T.M. Khoshgoftaar, "A survey on image data augmentation for deep learning," Journal of Big Data, vol. 6, no. 1, pp. 1-48, 2019.
 
[39] Zoph, B., Cubuk, E. D., Ghiasi, G., Lin, T. Y., Shlens, J., & Le, Q. V., "Learning data augmentation strategies for object detection," in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, pp. 566-583.
 
[[40] Kaur, P., B.S. Khehra, and E.B.S. Mavi, "Data augmentation for object detection: A review," in 2021 IEEE International Midwest Symposium on Circuits and Systems (MWSCAS), IEEE, Lansing, MI, USA, 2021, pp. 537-543.
 
[41] Shin, H.-C., K.-I. Lee, and C.-E. Lee, "Data augmentation method of object detection for deep learning in maritime image," in 2020 IEEE International Conference on Big Data and Smart Computing (BigComp), IEEE, Busan, Korea (South), 2020, pp. 463-466