Image Classification of Transfer Slips and Optical Character Recognition

Main Article Content

Chayanon Issaard
Sajjaporn Waijanya
Nuttachot Promrit
Manu Markmanee

Abstract

 


This article presents the classification of transfer slip images and Optical Character Recognition (OCR) use of automatic learning techniques that mimic the human neural network. The approach utilizes a Convolutional Neural Network (CNN) algorithm specifically designed for classifying transfer slip images, comprising three 2D convolutional layers combined with Max Pooling to prevent overfitting during the classification process. Subsequently, image segmentation is performed using YOLOv5 to identify the bank names, followed by OCR using Tesseract OCR technology to read the information on the transfer slips. The dataset was randomly split into a training set (80%) and test set (20%). Within the training set, 20% of the data was used as a validation set. The model was trained using a batch size of 32 and 50 epochs, achieving a classification accuracy of 99%. After classification, the segmented images were used to identify the bank names, and once identified, the images were subjected to OCR to extract the text. The results were displayed in a human-readable format and exported in JSON format.

Article Details

How to Cite
Issaard, C., Waijanya, S., Promrit, N., & Markmanee, M. (2025). Image Classification of Transfer Slips and Optical Character Recognition. Journal of Science Ladkrabang, 34(2), 1–15. retrieved from https://li01.tci-thaijo.org/index.php/science_kmitl/article/view/261945
Section
Research article

References

Arslnd, Ö., & Uymaz, S. A. (2022). Classification of invoice images by using convolutional neural networks. Journal of Advanced Research in Natural and Applied Sciences, 8(1), 8-25.

Hien, N. N., Thanh, D. N. H., Erkan, U., & Tavares, J. M. R. S. (2022). Image noise removal method based on thresholding and regularization techniques. IEEE Access, 10, 71584-71595. https://doi.org/10.1109/ACCESS.2022.3188315

Iamsamang, W., & Kongkachandra, R. (2013). Automatic Thai handwritten characters segmentation based on linear regression analysis and decision tree learning. Thai Journal of Science and Technology, 2, 167-174. (in Thai)

Lestari, I. N. T., & Mulyana, D. I. (2022). Implementation of OCR (optical character recognition) using Tesseract in detecting character in quotes text images. Journal of Applied Engineering and Technological Science, 4(1), 58-63. https://doi.org/10.37385/jaets.v4i1.905

Memon, J., Sami, M., Khan, R. A., & Uddin, M. (2020). Handwritten optical character recognition (OCR): A comprehensive systematic literature review (SLR). IEEE Access, 8, 142642-142668. https://doi.org/10.1109/ACCESS.2020.3012542

Park, J., Lee, E., Kim, Y., Kang, I., Koo, H. I., & Cho, N. I. (2020). Multi-lingual optical character recognition system using the reinforcement learning of character segmenter. IEEE Xplore, 8, 174437-174446. https://doi.org/10.1109/ACCESS.2020.3025769

Rezkiani, K., Nurtanio, I., & Syafaruddin. (2022). Logo detection using You Only Look Once (YOLO) method. Proceedings of the 2nd International Conference on Electronic and Electrical Engineering and Intelligent System (ICE3IS) (pp. 227-232). IEEE. https://doi.org/10.1109/ICE3IS56585.2022.10010121

Saravanan, C. (2010). Color image to grayscale image conversion. Proceedings of the 2010 Second International Conference on Computer Engineering and Applications (pp. 196-199). IEEE. https://doi.org/10.1109/ICCEA.2010.192

Tanprasert, C., Sinthupinyo, W., Dubey, P., & Tanprasert, T. (1999). Improved mixed Thai & English OCR using two-step neural net classification. NECTEC Technical Journal, 1, 41-46.

Vaidya, M., Rule, P. K., Kumar, H., Jain, A., & Kamble, A. R. (2019). Automating data entry forms for banks using OCR and CNN. International Journal for Applied Sciences and Engineering Technology, 7, 890-893.