Image Classification of Transfer Slips and Optical Character Recognition
Main Article Content
Abstract
This article presents the classification of transfer slip images and Optical Character Recognition (OCR) use of automatic learning techniques that mimic the human neural network. The approach utilizes a Convolutional Neural Network (CNN) algorithm specifically designed for classifying transfer slip images, comprising three 2D convolutional layers combined with Max Pooling to prevent overfitting during the classification process. Subsequently, image segmentation is performed using YOLOv5 to identify the bank names, followed by OCR using Tesseract OCR technology to read the information on the transfer slips. The dataset was randomly split into a training set (80%) and test set (20%). Within the training set, 20% of the data was used as a validation set. The model was trained using a batch size of 32 and 50 epochs, achieving a classification accuracy of 99%. After classification, the segmented images were used to identify the bank names, and once identified, the images were subjected to OCR to extract the text. The results were displayed in a human-readable format and exported in JSON format.
Article Details

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
References
Arslnd, Ö., & Uymaz, S. A. (2022). Classification of invoice images by using convolutional neural networks. Journal of Advanced Research in Natural and Applied Sciences, 8(1), 8-25.
Hien, N. N., Thanh, D. N. H., Erkan, U., & Tavares, J. M. R. S. (2022). Image noise removal method based on thresholding and regularization techniques. IEEE Access, 10, 71584-71595. https://doi.org/10.1109/ACCESS.2022.3188315
Iamsamang, W., & Kongkachandra, R. (2013). Automatic Thai handwritten characters segmentation based on linear regression analysis and decision tree learning. Thai Journal of Science and Technology, 2, 167-174. (in Thai)
Lestari, I. N. T., & Mulyana, D. I. (2022). Implementation of OCR (optical character recognition) using Tesseract in detecting character in quotes text images. Journal of Applied Engineering and Technological Science, 4(1), 58-63. https://doi.org/10.37385/jaets.v4i1.905
Memon, J., Sami, M., Khan, R. A., & Uddin, M. (2020). Handwritten optical character recognition (OCR): A comprehensive systematic literature review (SLR). IEEE Access, 8, 142642-142668. https://doi.org/10.1109/ACCESS.2020.3012542
Park, J., Lee, E., Kim, Y., Kang, I., Koo, H. I., & Cho, N. I. (2020). Multi-lingual optical character recognition system using the reinforcement learning of character segmenter. IEEE Xplore, 8, 174437-174446. https://doi.org/10.1109/ACCESS.2020.3025769
Rezkiani, K., Nurtanio, I., & Syafaruddin. (2022). Logo detection using You Only Look Once (YOLO) method. Proceedings of the 2nd International Conference on Electronic and Electrical Engineering and Intelligent System (ICE3IS) (pp. 227-232). IEEE. https://doi.org/10.1109/ICE3IS56585.2022.10010121
Saravanan, C. (2010). Color image to grayscale image conversion. Proceedings of the 2010 Second International Conference on Computer Engineering and Applications (pp. 196-199). IEEE. https://doi.org/10.1109/ICCEA.2010.192
Tanprasert, C., Sinthupinyo, W., Dubey, P., & Tanprasert, T. (1999). Improved mixed Thai & English OCR using two-step neural net classification. NECTEC Technical Journal, 1, 41-46.
Vaidya, M., Rule, P. K., Kumar, H., Jain, A., & Kamble, A. R. (2019). Automating data entry forms for banks using OCR and CNN. International Journal for Applied Sciences and Engineering Technology, 7, 890-893.