Fast and Accurate Deep Learning Architecture on Vehicle Type Recognition
Main Article Content
Abstract
Vehicle Type Recognition has a significant problem that happens when people need to search for vehicle data from a video surveillance system at a time when a license plate does not appear in the image. This paper proposes to solve this problem with a deep learning technique called Convolutional Neural Network (CNN), which is one of the latest advanced machine learning techniques. In the experiments, researchers collected two datasets of Vehicle Type Image Data (VTID I & II), which contained 1,310 and 4,356 images, respectively. The first experiment was performed with 5 CNN architectures (MobileNets, VGG16, VGG19, Inception V3, and Inception V4), and the second experiment with another 5 CNNs (MobileNetV2, ResNet50, Inception ResNet V2, Darknet-19, and Darknet-53) including several data augmentation methods. The results showed that MobileNets, when combine with the brightness augmented method, significantly outperformed other CNN architectures, producing the highest accuracy rate at 95.46%. It was also the fastest model when compared to other CNN networks.
Keywords: Vehicle Type Image Recognition; image classification; Convolutional Neural Network; deep learning; pattern recognition; image recognition
*Corresponding author: Tel.: (+66) 43-754359
E-mail: olarik.s@msu.ac.th
Article Details
Copyright Transfer Statement
The copyright of this article is transferred to Current Applied Science and Technology journal with effect if and when the article is accepted for publication. The copyright transfer covers the exclusive right to reproduce and distribute the article, including reprints, translations, photographic reproductions, electronic form (offline, online) or any other reproductions of similar nature.
The author warrants that this contribution is original and that he/she has full power to make this grant. The author signs for and accepts responsibility for releasing this material on behalf of any and all co-authors.
Here is the link for download: Copyright transfer form.pdf
References
Li, J., Zhao, W. and Guo, H., 2009. Vehicle type recognition based on Harris Corner Detector. Proceedings of the Second International Conference on Transportation Engineering, Chengdu, China, July 25-27, 2009, 3320-3325.
Zhang, B., 2013. Reliable classification of vehicle types based on cascade classifier ensembles. IEEE Transaction on Intelligent Transportation Systems, 14, 322-332.
Clady, X., Negri, P., Milgram, M. and Poulenard, R., 2008. Multi-class vehicle type recognition system. IAPR Workshop on Artificial Neural Networks in Pattern Recognition, Springer, Berlin, 228-239.
Dong, Z., Wu, Y., Pei, M. and Jia, Y., 2015. Vehicle type classification using a semisupervised convolutional neural network. IEEE Transaction on Intelligent Transportation Systems, 16(4), 2247-2256.
Huttunen, H., Yancheshmeh, F.S. and Chen, K. 2016. Car type recognition with deep neural networks. Proceedings of the Intelligent Vehicles Symposium (IV), Gothenburg, Sweden, 2016, 1115-1120.
Bautista, C.M., Dy, C.A., Mañalac, M.I., Orbe, R.A. and Cordel, M. 2016. Convolutional neural network for vehicle detection in low resolution traffic videos. Proceedings of the 2016 IEEE Region 10 Symposium (TENSYMP), Bali, Indonesia, 2016, 277-281.
Krizhevsky, A., Sutskever, I. and Hinton, G.E., 2012. Imagenet classification with deep convolutional neural networks. Proceedings of the 25th International Conference on Neural Information Processing Systems, December, 2012, 1097-1105.
Simonyan, K. and Zisserman, A., 2014. Very deep convolutional networks for large-scale image recognition. [online] Available at: http://arxiv.org/abs/1409.1556.
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V. and Rabinovich, A., 2015. Going deeper with convolutions. Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 2015, 1-9.
Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M. and Adam, H., 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. [online] Available at: http://arxiv.org/abs/1704.04861.
Boonsirisumpun, N. and Puarungroj, W., 2018. Loei Fabric Weaving Pattern Recognition Using Deep Neural Network. Proceedings of the 15th International Joint Conference on Computer Science and Software Engineering (JCSSE), Nakhonpathom, Thailand, July 11-13, 2018, 1-9.
LeCun, Y., Boser, B., Denker, J.S., Henderson, D., Howard, R.E., Hubbard, W. and Jackel, L.D., 1989. Backpropagation applied to handwritten zip code recognition. Neural Computing. 1(4), 541-551.
LeCun, Y. and Bengio, Y., 1995. Convolutional networks for images, speech, and time series. In: M.A. Arbib, ed. The Handbook of Brain Theory and Neural Networks. Cambridge: MIT Press, pp. 1-14.
Ioffe, S. and Szegedy, C., 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. Proceedings of Machine Learning Research, 2015, 448-456.
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J. and Wojna, Z., 2016. Rethinking the inception architecture for computer vision. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 2016, 2818-2826.
Szegedy, C., Ioffe, S., Vanhoucke, V. and Alemi, A., 2017. Inception-v4, Inception-Resnet and the impact of residual connections on learning. Proceedings of the Thirty-First AAAI (Association for the Advancement of Artificial Intelligence) Conference on Artificial Intelligence, San Francisco, CA, USA, 2017, 4278-4284.
Yoo, B., Choi, Y. and Choi, H., 2018. Fast Depthwise Separable Convolution for Embedded Systems. International Conference on Neural Information Processing, Siem Reap, Cambodia, December 13-16, 2018, 656-665.
Sifre, L. and Mallat, S., 2014. Rigid-Motion Scattering for Image Classification. Ph.D. ENS University, Paris, France.
Sandler, M., Howard, A., Zhu, M., Zhmoginov, A. and Chen, L.C., 2018. Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, 4510-4520.
He, K., Zhang, X., Ren, S. and Sun, J., 2016. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, 770-778.
Redmon, J., Divvala, S., Girshick, R. and Farhadi, A., 2016. You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2017, 779-788.
Redmon, J. and Farhadi, A., 2017. YOLO9000: Better, faster, stronger. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, 7263-7271.
Redmon, J. and Farhadi, A., 2018. Yolov3: An Incremental Improvement. [online] Available at: https://arxiv.org/pdf/1804.02767.pdf.
Shorten, C. and Khoshgoftaar, T.M., 2019. A survey on image data augmentation for deep learning. Journal of Big Data, 6, 60, https://doi.org/10.1186/s40537-019-0197-0
Mikołajczyk, A. and Grochowski, M., 2018. Data augmentation for improving deep learning in image classification problem. 2018 International Interdisciplinary PhD Workshop (IIPhDW), 117-122.