IEEE Access (Jan 2025)
Robust Classification of Smartphone Captured Handwritten Document Images Using Deep Learning
Abstract
Document classification is a challenging research problem in Document image analysis, majorly the presence of degradations like low illumination, blur, and shadows. In this paper, a deep learning-based two-level hierarchical approach is proposed for classifying smartphone-captured handwritten document images into three classes. The model is devised based on the pre-trained weights transferred from XceptionNet and the additional convolutional layers for feature extraction focused on high-frequency details from document images. A set of synthetically generated datasets concerning all three classes with two to three levels of degradation severity is considered for training the proposed deep learning model along with the real images labelled based on reference datasets. The dataset comprises blur and low illumination images with three levels of severity, and shadow images with two levels of severity. About 2,841 images from classes with 783 for blur, 828 for low illumination, and 1,230 for shadow were collected. The proposed model is compared with nine state-of-the-art deep-learning classification models. The proposed model showcases robust performance under varying levels of degradation severity levels by achieving 100% sensitivity towards shadow documents, while maintaining a high F1 score of 98.79% for Level 1. In case of low illumination conditions, again the model produces 100% precision, sensitivity, and specificity at Level 3 towards testsets. The results emphasize the model’s robustness to perform reliably, particularly under shadow and low illumination conditions.
Keywords