Next Article in Journal
An Advanced Strategy for Addressing Heterogeneity in SDN-IoT Networks for Ensuring QoS
Previous Article in Journal
Underwater Target Recognition via Cayley-Klein Measure and Shape Prior Information in Hyperspectral Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review of Document Image Enhancement Based on Document Degradation Problem

School of Opto-Electronic and Communication Engineering, Xiamen University of Technology, Xiamen 361000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(13), 7855; https://doi.org/10.3390/app13137855
Submission received: 7 May 2023 / Revised: 21 June 2023 / Accepted: 26 June 2023 / Published: 4 July 2023
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Document image enhancement methods are often used to improve the accuracy and efficiency of automated document analysis and recognition tasks such as character recognition. These document images could be degraded or damaged for various reasons including aging, fading handwriting, poor lighting conditions, watermarks, etc. In recent years, with the improvement of computer performance and the continuous development of deep learning, many methods have been proposed to enhance the quality of these document images. In this paper, we review six tasks of document degradation, namely, background texture, page smudging, fading, poor lighting conditions, watermarking, and blurring. We summarize the main models for each degradation problem as well as recent work, such as the binarization model that can be used to deal with the degradation of background textures, lettering smudges. When facing the problem of fading, a model for stroke connectivity can be used, while the other three degradation problems are mostly deep learning models. We discuss the current limitations and challenges of each degradation task and introduce the common public datasets and metrics. We identify several promising research directions and opportunities for future research.

1. Introduction

Document image enhancement is one of the pre-processing steps for document analysis and recognition. The document images are severely damaged or degraded due to long storage time, poor storage environment, and problems such as page smudges and fading of handwriting [1,2].Or, during the digitization process, the degradation of document images is caused by poor lighting conditions, shadows, camera distortion noise, blur, etc. [3,4,5]. Degraded document images have low visual quality and legibility. They could contain handwritten or machine-printed text or a mixture of both. In addition, they could contain multiple handwriting styles and different languages. To solve this problem, many document image enhancement methods have been proposed. For example, the quality of the image is improved by removing the degradation effect present in the image to restore its original content [6,7]. However, relying only on traditional methods, the task results for these degraded documents still need to be more significant, and it is challenging to automate document analysis. To achieve good results, it is necessary to manually enhance these images, and when the amount of data is relatively large, file image analysis becomes more difficult. Deep learning has been proposed and applied to different computer vision and image processing tasks, such as object detection [8,9], semantic segmentation [10], face detection and dataset creation [11,12,13], image enhancement [14,15,16], etc. The data show that compared with traditional methods, methods based on deep learning achieve more gratifying results. Similarly, the method of dealing with document image enhancement has also received great attention. Thus, it is essential to develop methods that can automatically enhance the visual quality and legibility of these images. The purpose of this survey is to review current methods for degraded document image enhancement and to analyze and discuss the advantages and disadvantages of the leading models for various degradation tasks, focusing on the limitations of the model processing effects and then discussing the directions for future research.
We have structured the rest of the paper as follows. Section 1 summarizes six degradation problems of documents, illustrated in Figure 1. In Section 2, we summarize the models to solve the corresponding problem and review the related work. In Section 3, we present the common public datasets and metrics and discuss the current limitations and challenges of each degradation task. Section 4 summarizes the current data problems in the field of document image enhancement and gives an outlook on the future of the document image enhancement field.
We make several vital contributions in this paper: one is reviewing the research progress of deep-learning-based document image enhancement methods, especially in the past five years, to help readers better understand the current development in this field. On the other hand, we summarize six image degradation problems, namely, background texture, page smudge, word fading, uneven lighting, watermark, and blur, and analyze the characteristics, advantages, and disadvantages of the leading models to help researchers choose the right approach according to their needs.

2. Document Image Degradation Problems

This section describes six degradation problems of document image enhancement, namely, background texture, page smudging, fading, poor lighting conditions, watermarking, and blurring. Figure 2 shows some image examples for each of the degradation problems. The image on the left is the input, and the right image is the output of each problem.

2.1. Background Texture

The texture is the description of the spatial distribution pattern of the image pixel gray level, reflecting the texture of the item, such as roughness, smoothness, granularity, randomness, and normalization. However, with the change in document storage time, sometimes the background texture will affect the readability of the text, resulting in the quality of the text being reduced. Figure 2a shows an example of background texture removal.

2.2. Page Smudging

Document image files may be incompletely preserved due to damage caused by the natural environment and human factors, resulting in various degradation issues like water impregnation, ink stains, ink mark penetration, and yellowing of pages, among others. These problems can be attributed to the task of handling page smudges in the degradation process. Figure 2b shows an example of addressing page smudges.

2.3. Handwriting Fading

The problem of handwriting fading mainly affects historical documents. Ancient historical books are precious pieces of cultural heritage with significant scientific and cultural values. The digitization of ancient books is a necessary approach to document protection and cultural inheritance. However, manual processing of these massive documents is time-consuming and laborious, and error-prone. Therefore, it is necessary to automatically process the images of historical documents with the help of computers. Due to the severe pollution of historical documents, messy writing format, and many other problems, especially the fading of the handwriting caused by these degradation problems, the text recognition of ancient historical documents could be better. Therefore, image analysis and recognition focusing on historical documents are challenging. Figure 2c is a restoration example of the fading problem.

2.4. Poor Lighting Conditions

In the process of document digitization, in the past, scanners were commonly used digital instruments, and scanning resulted in high-quality documents. However, with the development of science and technology, the types of digital image acquisition equipment are increasing, such as digital cameras, mobile phones, cameras, and so on. More people now use their cell phone cameras to get digital copies of their files. However, this approach may encounter problems, such as uneven lighting, excessive shadows, or too dark. In addition, for some old documents or antique literature, there may be wormholes, translucency, or unclear handwriting. This interference with the readability of the acquisition of documents has a harmful impact. The unevenness approach focuses on evaluating shadows on the document image and trying to remove them, making the document clean again and readable. Figure 2d shows an example of shadow removal under uneven lighting.

2.5. Watermark Removal

The digital age has led to more accessible communication of digital images and digital information. However, in the process of communication, there are various intentional and unintentional content modifications, and such editing techniques sometimes infringe on intellectual property rights. Moreover, the existence of a watermark is significant for the integrity of original content and copyright protection. Some files may contain one or more watermarks that block the document text, causing problems that make the document difficult to read. Similar to the case of document image denoising, it is difficult to deal with such problems because the watermark color may be the same as or darker than the document text. Therefore, we need a way to recover degraded version files. Watermark removal methods focus on eliminating watermarks and increasing document images’ visual quality and readability. We also need to use the correct watermark removal method. Figure 2e shows a watermark-removed image sample.

2.6. Deblur

Converting documents to digital images makes it difficult to control some unpredictable degradation factors fully. Among them, how to manage image clarity is an important issue. How to control fuzziness is one of the important concerns. One of the most common examples is the blurring of document images that can be introduced during filming. For example, moving the focus of the lens and camera vibration can all result in blurred images. Figure 2f shows an example of a blurry document image.
Image deblurring technology aims to restore the fuzzy image to the original image. As one of the important research areas in the field of image restoration, there have been many excellent achievements in this regard. The traditional method is mainly the filtering method, generally used to directly restore clear images through deconvolution [18]. With the extensive application of deep learning in recent years, scholars have proposed many effective deblurring models to achieve better results in document image deblurring.

3. Document Image Degradation Methods

In this section, we describe document image enhancement methods based primarily on deep learning and discuss their features, challenges, and limitations. Most of these methods solve multiple degradation problems. Below, we describe these methods in more detail.

3.1. Background Texture Problem

Methods for processing document images’ background texture can be divided into two kinds.
The first method is based on binarization. Image binarization converts the acquired image into a binary image, and the converted image performs better in text recognition. Binarization can be divided into threshold-based and numerical-based binarization. Threshold-based binarization methods, such as the Otsu algorithm [19], separate the foreground from the background of an image by finding a threshold value that maximizes the variance between the classes. It has the advantage of a fast computation speed for dealing with background texture problems, but the processing effect is still not ideal. Sauvola [20], Niblack [21], and other local threshold algorithms have better separation effects and applicability performance than their global counterparts. However, some problems exist because the thresholding performance depends heavily on the sliding window size. In addition, the parameter tuning method [22], the mixed-threshold algorithm of the clustering method [23], and the advantages of the global threshold and local threshold algorithms are integrated to better segment the foreground and background and reduce the impact of the original image background texture on the document image. Numerical theory-based methods such as the conditional random field (CRF)-based Howe algorithm, [24], which uses the Laplace operator to minimize the target energy function, can have relatively stable results in most cases but has high requirements for image quality. Ref. [25] estimated local threshold values using high-contrast image pixels to binarize the document image.
The second method is based on pixel classification. For example, Bezmaternykh et al. proposed a model [26] based on U-Net. They combined it with the structural advantages of CNN for historical document images to improve the ability of the model to obtain context information. However, manual parameter adjustment and data enhancement models are needed, and the model’s generalization ability needs to be improved. In 2021, Xiong et al. proposed an FD-Net model [27], as shown in Figure 3. It is a full dilation convolutional network, replacing all the downsampled and upsampled layers with dilation convolution. It can enhance the receptive field and segmentation accuracy. However, a grid effect, i.e., a discontinuity in the convolution kernel, may arise, which is unfavorable for the pixel-by-pixel prediction task.

3.2. Page Smudging Problem

Problems include ink penetration, prominent smudges, and yellowing of paper in the original document due to improper preservation. We attribute this type of problems to page smudging in degradation tasks. In addition to the traditional binarization method, the following models have better effects on this kind of problem.
In 2018, a hierarchical deep supervised network (DSN) architecture [28] was proposed by Vo et al. It applied the deep supervised network model to the document image binarization issues for the first time, which could better distinguish the foreground from background noise to predict text pixels with different feature levels. However, the large number of convolutional layers leads to the problem of processing small amounts of data as well as long processing times. In the same year, Bhowmik et al. proposed a game theory-inspired binarization technique [29], which designs a two-player, non-zero-sum algorithm at the pixel level to extract local information from document images, and then feeds the obtained results into the K-means algorithm for pixel classification. The pre-processing step and the post-processing step of the method also contribute to the enhancement of the overall algorithm, eliminating the intensity variations that often occur in the background and refining the results. In 2019, Gallego et al. proposed a selective auto-coding approach [30] that performs a transformation step on the input signal using a convolutional auto-encoder and adjusts these transformation parameters during the training process to achieve model optimization, and finally applies a global threshold to binarize the document, as shown in Figure 4. For different input image sizes and when the convolutional network uses a different number of convolutional kernels, the model can have good binarization results, and the convolutional kernel size does not have a large impact on the binarization results. However, the number of model parameters needs to be increased due to the use of step size instead of the pooling layer in the downsampling process. In the same year, Zhao et al. introduced conditional generative adversarial networks (cGAN) to solve the core problem of multi-scale information combinations in binarization tasks [31], as shown in Figure 5. The model has a good effect on degradation problems such as page smudging, but the drawback is that the training strategy of the sub-generator still needs to be modified, and the update efficiency and scalability of the model still need to be improved. Peng et al. used a multi-resolution attention model to infer the probability of text areas and then feed it into a convolutional random field (ConvCRF) to obtain the final result [32]. This method uses distance reciprocal distortion measure loss to infer the relationship between text regions and background by ConvCRF, which has a stronger generalization ability. However, the training and inference speed of ConvCRF is slow compared to other models, and its parameters are difficult to learn.
In 2020, Souibgui et al. proposed the model of DE-GAN [33], which transforms the threshold setting task of processing text and background into an image-to-image transformation task. For the generator, a U-Net-based structure is used, and the addition of bypass links reduces information loss and allows the model to learn more advanced features. For the discriminator, a simple multilayer convolution structure is used. The model cannot only deal with degradation problems such as page smudging but also has better results in watermark and blur removal for document images. Although it has better results on different document enhancement tasks, when dealing with binarization tasks, it is more time-consuming to train it several times in processing the problem to obtain better results because of the need to manually change the variables in the loss function.

3.3. Handwriting Fading Problem

In digitizing many historical books, the document images have a degraded condition of faded handwriting. A series of models have been proposed for this situation in image processing.
In 2018, Jia et al. proposed an adaptive global threshold selection algorithm based on structural symmetric pixels (SSP) [34]. First, they proposed an adaptive global threshold selection algorithm for SSP extraction, which can remove bleeding backgrounds while preserving fuzzy characters. SSP extraction enables feature maps of arbitrary size to output a vector of fixed length. Then, the proposed stroke width estimation algorithm improves adaptiveness. Finally, an iterative algorithm is used to estimate the text stroke width and remove the noise at the same time. This method is more suitable for processing printed text images than handwritten text images. In 2019, Ref. [35] introduced an adversarial learning approach, as shown in Figure 6. This method constructs a Texture Augmentation Network that transfers the texture element of a degraded reference document image to a clean binary image. The available document dataset is expanded by adding various noisy textures to the same textual content to form multiple versions of the degraded image. Finally, the newly generated images are passed through the network in order to return clean versions. The model is effective in improving robustness, and it does not require paired data. The model is similar to the cycleGAN approach, but the results obtained are more advantageous, thanks to the ability of BiNet in the model to learn autonomously and to fully supervise the optimization results.
In 2021, Xiong et al. proposed an algorithm based on background estimation and energy minimization [36]. First, a disc-shaped structuring element whose radius is computed by the minimum entropy-based stroke width transform (SWT) is used to estimate and compensate for the document background. SWT can extract the stroke width to obtain the character candidate region and filter out the mis-checked candidates using a priori knowledge of shape and texture. Then, Laplacian energy-based segmentation is performed on the compensated document images. Finally, post-processing is implemented to preserve text stroke connectivity and eliminate isolated noise. This method is able to better preserve text strokes, but it does not have more advantages than the current advanced deep-learning-based methods. In the same year, the DP-LinkNet model [37] was proposed by Xiong et al. The model follows the LinkNet [38] and D-LinkNet [39] architectures with the addition of the hybrid dilated convolution (HDC) module and the spatial pyramid pooling (SPP) module, which expands the perceptual field and aggregates multiscale contextual features, and an SPP module that encodes the output of the HDC with multicore pooling, both of which enable the model to extract textual strokes features with deep semantic information. Compared to other neural network models, training is faster and does not require deeper network layers. In 2022, Suh et al. proposed a two-stage color document image enhancement and binarization method using generative adversarial neural networks [40]. The first stage uses a color-independent adversarial neural network to remove background information from local image blocks and extract colored foreground information for document image enhancement. The second stage uses a multiscale adversarial neural network to generate a local binarized resultant image and a global binarized resultant image of the document image. One of the model’s advantages is its ability to handle multi-color degradation of document images. The disadvantage is that the training is slow due to the need to generate images of different channels. In the same year, Souibgui et al. proposed an encoder–decoder architecture based on Vision Transformer [41], as shown in Figure 7. The degraded image is first divided into several patches, which are then fed into the encoder, where the patch is mapped to a potential representation of each token during the encoding process, where each token corresponds one-to-one. The token is then passed to the decoder for patch enhancement. The model uses a self-attentive mechanism to obtain high-level global information for better performance. Since the images are trained in pixel blocks, there may be incomplete extraction of information from edge junction regions.
Figure 6. The structure of the model constructed by Suh et al. [40]. Reproduced with permission from Suh et al., Pattern Recognition; published by Elsevier, 2022.
Figure 6. The structure of the model constructed by Suh et al. [40]. Reproduced with permission from Suh et al., Pattern Recognition; published by Elsevier, 2022.
Applsci 13 07855 g006
Figure 7. DocEnTr model structure constructed by Souibgui et al. [41]. The input image is split into patches, which are linearly embedded, and the position information is added to them. The resulting sequence of vectors is fed to a standard Transformer encoder to obtain the latent representations. These representations are fed to another Transformer representing the decoder to obtain the decoded vector, which is linearly projected to vectors of pixels representing the output image patches. Published under a Creative Commons Attribution 4.0 license.
Figure 7. DocEnTr model structure constructed by Souibgui et al. [41]. The input image is split into patches, which are linearly embedded, and the position information is added to them. The resulting sequence of vectors is fed to a standard Transformer encoder to obtain the latent representations. These representations are fed to another Transformer representing the decoder to obtain the decoded vector, which is linearly projected to vectors of pixels representing the output image patches. Published under a Creative Commons Attribution 4.0 license.
Applsci 13 07855 g007

3.4. The Problem of Poor Lighting Conditions

Due to the effects of uneven lighting, various shadows are produced in the captured images. The following is a proposed method for removing shadows to address the problem of uneven lighting.
In 2017, Bako et al. proposed a method to estimate the background and text colors in local image blocks [42]. First, the local background color estimates are matched to the global reference to generate a shadow map. This shadow map is then used to correct the image to produce a final unshadowed output. This method detects and removes shadows from document photos and other text-based items. However, there are some limitations; e.g., the result will produce slight color and brightness variations, or in cases with strong hard shadows, the method can leave small residual artifacts at the shadow boundaries. In 2018, Ref. [43] proposed the ST-CGAN model, which consists of two stacked CGANs, each with a generator and a discriminator, capable of performing two tasks, shadow detection and shadow removal. It is the first framework for end-to-end joint learning of shadow detection and shadow removal with superior performance on various datasets and on both tasks. However, it is more demanding on the dataset and has more a priori conditions. In the same year, Kligler et al. introduced a newer general approach to the document enhancement problem [4], the main idea being to modify any state-of-the-art algorithm to improve its results by providing it with new information (input). Finally, it was found that the algorithm can be used for visibility detection and has better results for document shadow removal. However, for slight shadow problems, the PSNR effect of the output image does not have an advantage. In 2019, Jung et al. proposed an efficient shadow removal algorithm for digitized documents [44], which uses the luminance value of each pixel to create a digitized input document and then estimates the shading artifact on the document by simulating the immersion process before correcting the illumination. After estimating the shading artifacts, the digitized document is reconstructed using the Lambertian surface model. However, for the task of eliminating specularly reflected light from the image, the model does not process very well. In 2020, Lin et al. proposed the BEDSR-Net model [17], the first work to propose using deep learning to remove document shadows, as shown in Figure 8 below. The authors designed a background estimation module to extract the global background color of a document, as well as to learn information about the spatial distribution of pixels, encode it into an attention map, and then use a shadow removal network (SR-Net) to complete the shadow removal work. However, the effect of removing images with low background brightness needs to be improved.
In 2021, CANet [45] was proposed by Chen et al., as shown in Figure 9. The innovation of this model is to achieve shadow removal by passing information from non-shadowed regions to shadowed regions in the feature space, but the poor precision and inconsistent color of the segmented blocks limit it. Ref. [46] proposed the LG-ShadowNet model for shadow removal by training on unpaired data, where shaded and unshadowed training images are completely different and have no correspondence. The model does not use particularly sophisticated methods to optimize the results but rather uses linear regression to transform the pixel values to achieve better results and reduce errors. However, the advantages are not very obvious in terms of training for paired data.

3.5. Watermark Removal Problem

Watermarking is the most common way to protect the copyright of an image. Many organizations will add watermarks to their product images to prevent them from being stolen by others. We must recover the real image from the watermarked image if the original image is lost. Watermark removal can be divided into two main methods: the first is a single-stage target detection method, and the second is a two-stage target detection method. Next, typical models of the two methods are listed.
The single-stage target detection method is as follows.
In 2019, Gangeh et al. designed and tested an encoder–decoder model with an eight-level jump connection [47] to improve the quality of scanned documents with better results for removing watermarks from documents, but experimental and quantitative evaluations were inadequate. In 2020, DE-GAN was proposed [33], a more flexible model that is useful for watermark removal and other document degradation problems such as deblurring. However, it is computationally complex and requires predetermined parameters to be adjusted for different images. In 2022, Ref. [48] proposed an end-to-end architecture based on GANs with the addition of a handwritten text recognizer, where text information patches exist in handwritten text recognition (HTR) form during training to increase the readability of the text and facilitate document watermark removal, as shown in Figure 10. However, the use of the results for unpaired data is not yet as advantageous.
The two-stage target detection method is as follows.
In 2019, Liu et al. proposed a novel two-stage separable deep learning (TSDL) framework [49] consisting of noise-free end-to-end adversary training (FEAT) and noise-aware decoder-only training (ADOT). A redundant multilayer feature encoding network (RMFEN) was designed as an encoder framework that can learn a robust watermarking pattern without seeing any noise, can resist high-intensity noise, and has better stability, greater performance, and faster convergence speed. However, it is necessary to set the parameters manually, resulting in the need to manually set the results of the algorithm according to the features of different images, which leads to a decrease in model efficiency when facing multiple datasets. In 2020, Jiang et al. proposed a visible watermark removal network architecture [50] based on conditional generative adversarial networks (CGAN) [51] and least squares generative adversarial networks (LSGAN) [52]. As shown in Figure 11, the proposed model can automatically detect the watermark location without manual checking, which improves the model efficiency, but the generalization ability of the model is weak.
In 2021, WDNet was proposed by Liu et al. [53], which is a network dedicated to watermark removal. It does not remove the watermark but separates the watermark and uses this watermark to build a larger dataset, which can better segment the real image and the watermark with higher accuracy. However, the segmentation accuracy for color images still needs to be improved. In 2023, Ge et al. proposed an end-to-end document image watermarking scheme using deep neural networks [54], as shown in Figure 12. A noise layer is added to simulate various attacks that may be encountered in reality, such as cropout, dropout, Gaussian blur, etc. The text-sensitive loss function aims to limit the embedding modification of characters. An embedding strength adjustment strategy is also proposed in the paper to improve the quality of watermarked images while reducing the loss of extraction accuracy. This method is designed specifically for document images, but the advantages of processing the natural image watermarking scheme are not great.

3.6. Deblur Problem

The process of acquiring document images (e.g., scanned images of archival contracts) in the batch may be subject to various disturbances, the most common of which are motion-type blurs such as camera shake and motion offset. Document image blurring not only makes the image content visually illegible but also leads to poor image text detection and recognition, so in-depth research on document image deblurring is needed.
In 2015, Hradis et al. proposed using convolutional neural networks to recover blurred text documents [55]. The trained network can reconstruct high-quality images directly from blurred inputs without assuming any specific blur and noise model. The performance of convolutional networks was improved on a large number of text documents and a combination of realistic bokeh and camera-dithering blur kernels. This method gives good results even if the noise level is higher than the trained model. However, the model still has room for improvement. In 2016, Pan et al. proposed a simple and effective blind image deblurring method based on the dark channel prior [56]. Based on the observation of the blurring process, the L0 regularization term is used to minimize the dark channel. In the recovery process, it is in favor of the clear image rather than the blurred image to improve the model’s deblurring performance. The model also has good results in non-uniform deblurring, but the accuracy of the processing effect on text images needs to be improved. In 2017, a model for cyclic consistency was proposed in Ref. [57]. It is a method for learning to transform images from a source domain X to a target domain Y without paired examples and is able to capture the relationship between high-level appearance structures. However, it has drawbacks; e.g., the accuracy is not high enough. In the same year, Nah et al. applied deep multiscale convolutional neural networks to the problem of deblurring [58]. This multiscale strategy decomposes the complex problem and recovers it step by step, first recovering large-scale information at low resolution and then recovering detailed information at high resolution. The problem is simplified while increasing the receptive field of the image as a way to improve the model’s ability to obtain context. However, the deblurring effect on real blurred images needs to be improved. In 2018, DeblurGAN is a GAN-based method for blind motion blur removal proposed by Kupyn et al. [59], who were inspired by SRGAN [60] and CGAN [51] and regarded image deblurring as a special class of image2image task, which is based on WGAN [61] and content loss to train the model with good performance in terms of processing speed and results. The model is mainly for the dataset with high-speed motion, and the effect for ordinary fuzzy data needs to be improved. In 2019, Lee et al. came out with a blind text image deblurring algorithm using a text-specific hybrid dictionary [62]. It is a text image prior based on sparse representation, which models the relationship between an intermediate latent image and a desired sharp image and can obtain better contextual information. However, it is not effective if the blurred text image has low contrast. In the same year, Lu et al. proposed an unsupervised method for domain-specific single-image deblurring based mainly on disentangled representation [63] by untangling content and blurred features in blurred images and adding KL divergence loss to prevent blurred features from encoding content information. To preserve the content structure of the original image, a blurring branch and cyclic-consistency loss are added to the framework, while the added perceptual loss helps to remove unrealistic artifacts from the blurred image. The innovation is precisely the attempt to separate content from ambiguous information. In 2020, a method using the Local Maximum Difference Prior (LMD) was proposed in Ref. [64] by introducing a linear operator to compute the LMD and using the L1 norm to constrain the terms involved in the LMD. The LMD method decomposes a complex multicomponent signal into several components. This method is more general. However, when the image contains non-Gaussian noise, the processing is not very effective. In 2021, Blur2sharp was proposed by Neji et al. as an end-to-end model based on cycleGan [65], as shown in Figure 13. This model can perform deblurring without knowledge of the blurring kernel and generates clear images. However, the model processing results are similar to those of other methods and are not particularly advantageous. In 2022, Gonwirat et al. proposed DeblurGAN-CNN [66], as shown in Figure 14, a combination of DeblurGAN and CNN models which was upgraded from the effect of the original DeblurGAN to solve many noise problems and enhance the recognition performance of handwritten character images.

4. Datasets and Metrics

4.1. Datasets

  • Jung’s dataset [44]
The dataset consists mainly of 159 digitized documents with illumination distortion, of which 109 images were captured using the cameras of two smartphones and the remaining 50 were captured using a scanner. The size of the camera-captured images was 3264 × 2448 and that of the scanned images was 3455 × 2464 (72 dpi).
2.
Real Document Shadow Removal Dataset(RDSRD) [17]
The images were captured using Sony RX100 m3 and a flashlight, manufactured by the Sony group of companies from Tokyo, Japan. These devices are all fixed on tripods The camera is triggered using a remote through WiFi to avoid touching the camera during capture. The dataset contains 540 photos, including images of the effects of paper, newspapers, and slides under different lighting conditions, covering 25 documents.
3.
Blurry document images (BMVC)Text dataset [55]
A total of 66,000 text images of size 300 × 300 were used for training and 94 images of size 512 × 512 were used for OCR testing. Each patch was extracted from a different document page and each fuzzy kernel used was unique.
4.
Bickley diary [67]
The images of the Bickley diary dataset were taken from a photocopy of a diary that was written about 100 years ago. It consists of 92 badly degraded handwritten travel diary documents, 7 of which have GT images. These images suffer from different types of degradation, such as water stains and ink bleed-through.
5.
SMADI (Synchromedia Multispectral Ancient Document Images Dataset) [68]
S-MS (Synchromedia Multispectral Ancient Documents): Multispectral imaging (MSI) represents an innovative and non-destructive technique for analyzing materials such as ancient documents. Ancient handwritten letters were collected in a multispectral image database. This database contains 30 multispectral images of authentic historical handwritten letters. These old documents are written in iron gall ink and are dated from the 17th to the 20th century. The original documents were borrowed from the National Library of Quebec and were imaged using an eighth-note CX MSI camera. Through this process, 8 images were produced for each document, resulting in a total of 240 images for the file.
6.
DIBCO and H-DIBCO
The DIBCO printed or handwritten document image dataset is mainly used for the task of binarization. It was introduced by the Document Image Binarization Competition, including DIBCO 2009 [69], 2011 [70], 2013 [71], 2017 [72], and 2019 [73] and H-DIBCO 2010 [74], 2012 [75], 2014 [76], 2016 [77], and 2018 [78] benchmark datasets, covering 36 machine-printed and 90 handwritten document images, as well as their corresponding ground truth images. The historical documents in these datasets originated from the Recognition and Enrichment of Archival Documents (READ) project. A total of 136 test images contain representative historical document degradation, such as severely damaged pages, ink bleed-through, page stains, text stroke fading, and background texture.
7.
DocImgEN
A batch of PDFs were downloaded from the IEEE Xplore database [79] and converted PDF pages to PNG images. Then, the image block size of 400 × 400 was cropped to the document watermark image. DocImgEN includes 230,000 training samples, 10,000 validation samples, and 10,000 test document images with English words.
8.
DocImgCN
A batch of PDFs were downloaded from the China National Knowledge Infrastructure (CNKI) [80] to prepare document images with Chinese characters. DocImgCN includes 230,000 training samples, 10,000 validation samples, and 10,000 test document images.
Table 1 shows a summary of the relevant contents of the dataset.

4.2. Metrics

In this section, for the quantitative evaluation and comparison of advanced algorithms, we introduce evaluation metrics: F-measure (FM), inverse distance distortion (DRD), pseudo-F-measure (p-FM), peak signal-to-noise ratio (PSNR), word error rate (CER), word error rate (WER), and structural similarity (SSIM).
  • F-Measure (FM) [71]: It is the summed average of accuracy and recall, which is a common evaluation criterion in the field of IR (information retrieval) and is often used to evaluate the classification models. The calculation formula is as follows:
    F β = β 2 + 1 P R β 2 · P + R
    where β is a parameter, P is precision, and R is recall. P = T P T P + F P , R = T P T P + F N , TP denotes true, FP denotes false positive, and FN denotes false negative.
When the parameter β = 1, it is the most common F1-measure.
F 1 = 2 × P R P + R
  • Distance Reciprocal Distortion (DRD) [71]: It is used to measure the visual distortion of the image in the binary document. The formula is as follows:
    D R D = k D R D k N U B N
    where D R D k is the distortion of the kth flipped pixel and NUBN is the number of non-uniform (not all black or white pixels) 8 × 8 blocks in the GT image.
  • Pseudo-F-Measure: It is introduced in Ref. [81], which makes use of pseudo-recall and pseudo-precision. The advantage of using pseudo-recall and pseudo-precision is that they use the weighted distance between output images as boundaries of characters in the extracted document and the boundaries of characters in the ground truth (GT) image. One other advantage of using the pseudo nature of recall is its consideration of local stroke width in output images, while pseudo-recall takes into consideration the local stroke width, and the pseudo nature of precision grows to the stroke width of the connected component in ground truth images.
    p F M = 2 × p   Recall   ×   Precision   p   Recall   +   Precision  
    where pRecall is defined as the percentage of the skeletonized ground truth image.
  • Peak signal-to-noise ratio (PSNR) [82]: The ratio between the maximum possible power used to represent a signal and the power of the corrupted noise that affects the fidelity of its representation. Since many signals have a wide dynamic range, PSNR is often expressed as a pair number using a decibel scale. PSNR is mainly used in image processing to quantify the reconstruction quality of images and videos affected by lossy compression.
PSNR is defined by the mean square error (MSE). Given a noise-free m × n monochrome image I and its noise approximation K, the MSE is defined as
M S E = 1 m n i = 0 m 1 j = 0 n 1 [ I i , j K i , j ] 2
PSNR is defined as
P S N R = 10 · log 10 ( M A X I 2 M S E )
where M A X I 2 is the maximum possible pixel value of the image.
  • Character error rate (CER) [83]: Character Error Rate is computed based on the Levenshtein distance. It is the minimum number of character-level operations required to transform the ground truth or reference text into the OCR output text. CER is formulated as follows:
    C E R = S + D + I N
    where S is the number of substitutions, D is the number of deletions, I is the number of insertions, and N is the number of characters quoted or ground truth text. The lower the CER value, the better the OCR performance model. CER can be normalized to ensure that it does not fall out of the range of 0–100 and is not affected by insertion errors. In normalizing CER, C is the number of correctly identified in the text sentence. The normalized CER is formulated as follows:
    C E R n o r m a l i z e d = S + D + I S + D + I + C
  • Word error rate (WER) [84]: In order to maintain consistency between the identified word sequences and the standard word sequences, certain words need to be replaced, deleted, or inserted, and the total number of these inserted, replaced, or deleted words, divided by the percentage of the total number of words in the standard word sequences, is the WER.
    W E R = S w + D w + I w N
    where S w is the number of substitutions, D w is the words deleted, I w is the words inserted, and N is the number of words.
  • Structural similarity (SSIM) [85]: The structural similarity index defines structural information from the perspective of image composition as an attribute that reflects the structure of objects in a scene independently of luminance and contrast and models distortion as a combination of three different factors: luminance, contrast, and structure. Given patch “x” from one image and corresponding patch “y” from another image. The formula is as follows:
    S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
    where μ x is the average of x, μ y is the average of y, σ x 2 is the variance of x, σ y 2 is the variance of y, and σ x y is the covariance of x and y. C 1 , C 2 stand for constant.

4.3. Experiment

We conducted experiments based on 10 DIBCO and H-DIBCO test datasets for the background texture and page smudge problems. The handwriting fading problem was investigated with experiments for the DIBCO2018 dataset. The results are shown in Table 2. The experimental results of the illumination non-uniformity problem are shown in Table 3. The experimental results of the deblurring problem are shown in Table 4.
Table 2. Experimental results of background texture, page smudging, and fading of lettering.
Table 2. Experimental results of background texture, page smudging, and fading of lettering.
DegradationDatasetMethodFMFPSPSNRDRD
background textureDIBCO
H-DIBCO
Otsu74.2276.9914.5430.36
Niblack41.1241.576.6791.23
Sauvola79.1282.9516.078.61
Bezmaternykh’s UNet89.2990.5321.323.29
FD-Net95.2596.6522.841.22
page smudgeVo’s DSN88.0490.8118.944.47
Bhowmik’s GiB83.1687.7216.728.82
Gallego’s SAE79.2281.1216.099.75
Zhao’s cGAN87.4588.8718.815.56
Peng’s woConvCRF86.0987.4018.994.83
handwriting fadingH-DIBCO
2018
Bhunia [35]59.2559.1811.809.56
Xiong [36]88.3490.3719.114.93
DP-LinkNet95.9996.8522.711.09
Suh [40]84.9591.5817.0416.86
DocEnTr90.5993.9719.463.35
For the background texture problem, it can be seen from the experimental results that the deep learning approach is more advantageous compared to the traditional thresholding approach; in particular, the FD-Net model effect is excellent in the comparison of these evaluation index data. As for the page smudging problem, it can be seen from the experimental results that the model effect using cGAN is better than the other four methods. For the experiments on handwriting fading, models specializing in antique documents, such as DP-LinkNet and DocEnTr, work better on the corresponding datasets. These two model structures are able to extract deeper semantic information. For the problem of uneven illumination, it is found by the results that the methods with better processing effects are those with more a priori conditions on the dataset, such as ST-CGAN and BEDSR-Net models, which use a dataset that requires three parts: shaded image, corresponding unshaded image, and shaded mask. The different datasets targeted by BEDSR-Net can also have better results. For the deblurring problem, most current algorithms are based on GAN structure changes to encode the blurred information accurately and achieve high-quality deblurring of images. In recent years, some Transformer-based models have emerged to be applied to the deblurring problem, which handles the balance between spatial texture details and high-level context relatively well, but still does not work well for aggregating information about local and non-local pixel interactions, allowing only part of the valuable information to be passed further through the network hierarchy, and not well for passing texture details between channels. The model performance still needs to be improved. Furthermore, there are still possibilities for future development discussions.

4.4. Problems and Development Direction

Document image enhancement tasks are far from being solved; some tasks have only been studied minimally. In Table 5, we discuss these issues and future work.

5. Conclusions

In recent years, many network models have been proposed to solve various degradation problems in document image enhancement. This article discusses six kinds of image enhancement classification problems, such as shadow removal under uneven illumination, BEDSR-Net, CANet, etc., that are proposed to solve this problem. For example, DeblurGAN, Blur2sharp, etc., are also offered to solve the image recovery problem for blurred text. These methods can handle specific types of degradation problems very well. However, document image enhancement tasks still have some issues. For example, because of the small number of open-source datasets, many methods use GAN models to generate data, which leads to the generation of data that need to be sufficiently considered. It is not enough to be used in a wide variety of factual scenarios, and even though the models are effective for such data, the generalization ability of the models is not sufficient. Many of the current methods are based on machine learning methods for processing, but machine learning methods require data to be trained for learning. Therefore, sufficient datasets are also needed. Sufficient datasets can better refine the data and optimize the results. In the future, models trained with suitable datasets will be better able to handle document image enhancement problems.

Author Contributions

Conceptualization, Y.Z. and R.Z.; methodology, Y.Z.; software, Z.Y.; validation, Y.Z., J.H. and J.S.; formal analysis, Y.Z.; investigation, Z.Y.; resources, J.H.; data curation, J.S.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z.; visualization, R.Z.; supervision, S.Z.; project administration, S.Z.; funding acquisition, S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural Science Foundation of Fujian Province grant number 2020J01295. And The APC was funded by Natural Science Foundation of Fujian Province.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Su, B.; Lu, S.; Tan, C.L. Robust document image binarization technique for degraded document images. IEEE Trans. Image Process. 2012, 22, 1408–1417. [Google Scholar]
  2. Sulaiman, A.; Omar, K.; Nasrudin, M.F. Degraded historical document binarization: A review on issues, challenges, techniques, and future directions. J. Imaging 2019, 5, 48. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Chen, X.; He, X.; Yang, J.; Wu, Q. An effective document image deblurring algorithm. In Proceedings of the CVPR 2011, Washington, DC, USA, 20–25 June 2011; pp. 369–376. [Google Scholar]
  4. Kligler, N.; Katz, S.; Tal, A. Document enhancement using visibility detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; pp. 2374–2382. [Google Scholar]
  5. Mesquita, R.G.; Mello, C.A.; Almeida, L. A new thresholding algorithm for document images based on the perception of objects by distance. Integr. Comput.-Aided Eng. 2014, 21, 133–146. [Google Scholar] [CrossRef]
  6. Hedjam, R.; Cheriet, M. Historical document image restoration using multispectral imaging system. Pattern Recognit. 2013, 46, 2297–2312. [Google Scholar] [CrossRef]
  7. Lu, D.; Huang, X.; Sui, L. Binarization of degraded document images based on contrast enhancement. Int. J. Doc. Anal. Recognit. 2018, 21, 123–135. [Google Scholar] [CrossRef]
  8. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  9. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  10. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  11. Anvari, Z.; Athitsos, V. A pipeline for automated face dataset creation from unlabeled images. In Proceedings of the 12th ACM International Conference on PErvasive Technologies Related to Assistive Environments, New York, NY, USA, 5–7 June 2019; pp. 227–235. [Google Scholar]
  12. Lin, W.-A.; Chen, J.-C.; Castillo, C.D.; Chellappa, R. Deep density clustering of unconstrained faces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; pp. 8128–8137. [Google Scholar]
  13. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  14. Gu, S.; Zuo, W.; Guo, S.; Chen, Y.; Chen, C.; Zhang, L. Learning dynamic guidance for depth image enhancement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3769–3778. [Google Scholar]
  15. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
  16. Singh, G.; Mittal, A. Various image enhancement techniques—A critical review. Int. J. Innov. Sci. Res. 2014, 10, 267–274. [Google Scholar]
  17. Lin, Y.-H.; Chen, W.-C.; Chuang, Y.-Y. Bedsr-net: A deep shadow removal network from a single document image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12905–12914. [Google Scholar]
  18. Hansen, P.C.; Nagy, J.G.; O’leary, D.P. Deblurring Images: Matrices, Spectra, and Filtering; SIAM: Philadelphia, PA, USA, 2006. [Google Scholar]
  19. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  20. Sauvola, J.; Pietikäinen, M. Adaptive document image binarization. Pattern Recognit. 2000, 33, 225–236. [Google Scholar] [CrossRef] [Green Version]
  21. Niblack, W. An Introduction to Digital Image Processing; Strandberg Publishing Company: København, Denmark, 1985. [Google Scholar]
  22. Westphal, F.; Grahn, H.; Lavesson, N. Efficient document image binarization using heterogeneous computing and parameter tuning. Int. J. Doc. Anal. Recognit. 2018, 21, 41–58. [Google Scholar] [CrossRef] [Green Version]
  23. Jana, P.; Ghosh, S.; Bera, S.K.; Sarkar, R. Handwritten document image binarization: An adaptive K-means based approach. In Proceedings of the 2017 IEEE Calcutta Conference (CALCON), Kolkata, India, 2–3 December 2017; pp. 226–230. [Google Scholar]
  24. Howe, N.R. Document binarization with automatic parameter tuning. Int. J. Doc. Anal. Recognit. (IJDAR) 2013, 16, 247–258. [Google Scholar] [CrossRef] [Green Version]
  25. Rani, U.; Kaur, A.; Josan, G. A New Contrast Based Degraded Document Image Binarization. In Cognitive Computing in Human Cognition; Springer: Berlin/Heidelberg, Germany, 2020; pp. 83–90. [Google Scholar]
  26. Bezmaternykh, P.V.; Ilin, D.A.; Nikolaev, D.P. U-Net-bin: Hacking the document image binarization contest. Кoмпьютерная Оптика 2019, 43, 825–832. [Google Scholar] [CrossRef]
  27. Xiong, W.; Yue, L.; Zhou, L.; Wei, L.; Li, M. FD-Net: A Fully Dilated Convolutional Network for Historical Document Image Binarization. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Beijing, China, 29 October–1 November 2021; pp. 518–529. [Google Scholar]
  28. Vo, Q.N.; Kim, S.H.; Yang, H.J.; Lee, G. Binarization of degraded document images based on hierarchical deep supervised network. Pattern Recognit. 2018, 74, 568–586. [Google Scholar] [CrossRef]
  29. Bhowmik, S.; Sarkar, R.; Das, B.; Doermann, D. GiB: A Game theory Inspired Binarization technique for degraded document images. IEEE Trans. Image Process. 2018, 28, 1443–1455. [Google Scholar] [CrossRef] [PubMed]
  30. Calvo-Zaragoza, J.; Gallego, A.-J. A selectional auto-encoder approach for document image binarization. Pattern Recognit. 2019, 86, 37–47. [Google Scholar] [CrossRef] [Green Version]
  31. Zhao, J.; Shi, C.; Jia, F.; Wang, Y.; Xiao, B. Document image binarization with cascaded generators of conditional generative adversarial networks. Pattern Recognit. 2019, 96, 106968. [Google Scholar] [CrossRef]
  32. Peng, X.; Wang, C.; Cao, H. Document binarization via multi-resolutional attention model with DRD loss. In Proceedings of the 2019 International Conference on Document Analysis and Recognition (ICDAR), Sydney, Australia, 20–25 September 2019; pp. 45–50. [Google Scholar]
  33. Souibgui, M.A.; Kessentini, Y. DE-GAN: A conditional generative adversarial network for document enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1180–1191. [Google Scholar] [CrossRef]
  34. Jia, F.; Shi, C.; He, K.; Wang, C.; Xiao, B. Degraded document image binarization using structural symmetry of strokes. Pattern Recognit. 2018, 74, 225–240. [Google Scholar] [CrossRef]
  35. Bhunia, A.K.; Bhunia, A.K.; Sain, A.; Roy, P.P. Improving document binarization via adversarial noise-texture augmentation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2721–2725. [Google Scholar]
  36. Xiong, W.; Zhou, L.; Yue, L.; Li, L.; Wang, S. An enhanced binarization framework for degraded historical document images. EURASIP J. Image Video Process. 2021, 2021, 13. [Google Scholar] [CrossRef]
  37. Xiong, W.; Jia, X.; Yang, D.; Ai, M.; Li, L.; Wang, S. DP-LinkNet: A convolutional network for historical document image binarization. KSII Trans. Internet Inf. Syst. (TIIS) 2021, 15, 1778–1797. [Google Scholar]
  38. Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  39. Zhou, L.; Zhang, C.; Wu, M. D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 182–186. [Google Scholar]
  40. Suh, S.; Kim, J.; Lukowicz, P.; Lee, Y.O. Two-stage generative adversarial networks for binarization of color document images. Pattern Recognit. 2022, 130, 108810. [Google Scholar] [CrossRef]
  41. Souibgui, M.A.; Biswas, S.; Jemni, S.K.; Kessentini, Y.; Fornés, A.; Lladós, J.; Pal, U. Docentr: An end-to-end document image enhancement transformer. arXiv 2022, arXiv:2201.10252. [Google Scholar]
  42. Bako, S.; Darabi, S.; Shechtman, E.; Wang, J.; Sunkavalli, K.; Sen, P. Removing shadows from images of documents. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; pp. 173–183. [Google Scholar]
  43. Wang, J.; Li, X.; Yang, J. Stacked conditional generative adversarial networks for jointly learning shadow detection and shadow removal. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; pp. 1788–1797. [Google Scholar]
  44. Jung, S.; Hasan, M.A.; Kim, C. Water-filling: An efficient algorithm for digitized document shadow removal. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; pp. 398–414. [Google Scholar]
  45. Chen, Z.; Long, C.; Zhang, L.; Xiao, C. Canet: A context-aware network for shadow removal. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4743–4752. [Google Scholar]
  46. Liu, Z.; Yin, H.; Mi, Y.; Pu, M.; Wang, S. Shadow removal by a lightness-guided network with training on unpaired data. IEEE Trans. Image Process. 2021, 30, 1853–1865. [Google Scholar] [CrossRef] [PubMed]
  47. Gangeh, M.J.; Tiyyagura, S.R.; Dasaratha, S.V.; Motahari, H.; Duffy, N.P. Document enhancement system using auto-encoders. In Proceedings of the Workshop on Document Intelligence at NeurIPS 2019, Vancouver, BC, Canada, 14 December 2019. [Google Scholar]
  48. Jemni, S.K.; Souibgui, M.A.; Kessentini, Y.; Fornés, A. Enhance to read better: A multi-task adversarial network for handwritten document image enhancement. Pattern Recognit. 2022, 123, 108370. [Google Scholar] [CrossRef]
  49. Liu, Y.; Guo, M.; Zhang, J.; Zhu, Y.; Xie, X. A novel two-stage separable deep learning framework for practical blind watermarking. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1509–1517. [Google Scholar]
  50. Jiang, P.; He, S.; Yu, H.; Zhang, Y. Two-stage visible watermark removal architecture based on deep learning. IET Image Process. 2020, 14, 3819–3828. [Google Scholar] [CrossRef]
  51. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  52. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Paul Smolley, S. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Honolulu, HI, USA, 21–26 July 2017; pp. 2794–2802. [Google Scholar]
  53. Liu, Y.; Zhu, Z.; Bai, X. Wdnet: Watermark-decomposition network for visible watermark removal. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3685–3693. [Google Scholar]
  54. Ge, S.; Xia, Z.; Fei, J.; Sun, X.; Weng, J. A Robust Document Image Watermarking Scheme using Deep Neural Network. arXiv 2022, arXiv:2202.13067. [Google Scholar] [CrossRef]
  55. Hradiš, M.; Kotera, J.; Zemcık, P.; Šroubek, F. Convolutional neural networks for direct text deblurring. In Proceedings of the BMVC, Swansea, UK, 7–10 September 2015. [Google Scholar]
  56. Pan, J.; Sun, D.; Pfister, H.; Yang, M.-H. Blind image deblurring using dark channel prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1628–1636. [Google Scholar]
  57. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  58. Nah, S.; Hyun Kim, T.; Mu Lee, K. Deep multi-scale convolutional neural network for dynamic scene deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3883–3891. [Google Scholar]
  59. Kupyn, O.; Budzan, V.; Mykhailych, M.; Mishkin, D.; Matas, J. Deblurgan: Blind motion deblurring using conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–23 June 2018; pp. 8183–8192. [Google Scholar]
  60. Takano, N.; Alaghband, G. Srgan: Training dataset matters. arXiv 2019, arXiv:1903.09922. [Google Scholar]
  61. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  62. Lee, H.; Jung, C.; Kim, C. Blind deblurring of text images using a text-specific hybrid dictionary. IEEE Trans. Image Process. 2019, 29, 710–723. [Google Scholar] [CrossRef]
  63. Lu, B.; Chen, J.-C.; Chellappa, R. Unsupervised domain-specific deblurring via disentangled representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10225–10234. [Google Scholar]
  64. Liu, J.; Tan, J.; He, L.; Ge, X.; Hu, D. Blind image deblurring via local maximum difference prior. IEEE Access 2020, 8, 219295–219307. [Google Scholar] [CrossRef]
  65. Neji, H.; Hamdani, T.; Halima, M.; Nogueras-Iso, J.; Alimi, A.M. Blur2sharp: A gan-based model for document image deblurring. Int. J. Comput. Intell. Syst. 2021, 14, 1315–1321. [Google Scholar] [CrossRef]
  66. Gonwirat, S.; Surinta, O. DeblurGAN-CNN: Effective Image Denoising and Recognition for Noisy Handwritten Characters. IEEE Access 2022, 10, 90133–90148. [Google Scholar] [CrossRef]
  67. Deng, F.; Wu, Z.; Lu, Z.; Brown, M.S. BinarizationShop: A user-assisted software suite for converting old documents to black-and-white. In Proceedings of the 10th Annual Joint Conference on Digital Libraries, Gold Coast, Australia, 21–25 June 2010; pp. 255–258. [Google Scholar]
  68. Hedjam, R.; Nafchi, H.Z.; Moghaddam, R.F.; Kalacska, M.; Cheriet, M. Icdar 2015 contest on multispectral text extraction (ms-tex 2015). In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR), Tunis, Tunisia, 23–26 August 2015; pp. 1181–1185. [Google Scholar]
  69. Gatos, B.; Ntirogiannis, K.; Pratikakis, I. ICDAR 2009 document image binarization contest (DIBCO 2009). In Proceedings of the 2009 10th International Conference on Document Analysis and Recognition, Barcelona, Spain, 26–29 July 2009; pp. 1375–1382. [Google Scholar]
  70. Pratikakis, I.; Gatos, B.; Ntirogiannis, K. ICDAR 2011 Document Image Binarization Contest (DIBCO 2011). In Proceedings of the International Conference on Document Analysis & Recognition, Beijing, China, 18–21 September 2011. [Google Scholar]
  71. Pratikakis, I.; Gatos, B.; Ntirogiannis, K. ICDAR 2013 document image binarization contest (DIBCO 2013). In Proceedings of the 2013 12th International Conference on Document Analysis and Recognition, Washington, DC, USA, 25–28 August 2013; pp. 1471–1476. [Google Scholar]
  72. Pratikakis, I.; Zagoris, K.; Barlas, G.; Gatos, B. ICDAR2017 competition on document image binarization (DIBCO 2017). In Proceedings of the 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), Kyoto, Japan, 9–15 November 2017; pp. 1395–1403. [Google Scholar]
  73. Pratikakis, I.; Zagoris, K.; Karagiannis, X.; Tsochatzidis, L.; Mondal, T.; Marthot-Santaniello, I. ICDAR 2019 Competition on Document Image Binarization (DIBCO 2019). In Proceedings of the 2019 International Conference on Document Analysis and Recognition (ICDAR), Sydney, Australia, 20–25 September 2019. [Google Scholar]
  74. Pratikakis, I.; Gatos, B.; Ntirogiannis, K. H-DIBCO 2010-handwritten document image binarization competition. In Proceedings of the 2010 12th International Conference on Frontiers in Handwriting Recognition, Washington, DC, USA, 16–18 November 2010; pp. 727–732. [Google Scholar]
  75. Pratikakis, I.; Gatos, B.; Ntirogiannis, K. ICFHR 2012 competition on handwritten document image binarization (H-DIBCO 2012). In Proceedings of the 2012 International Conference on Frontiers in Handwriting Recognition, Bari, Italy, 18–20 September 2012; pp. 817–822. [Google Scholar]
  76. Ntirogiannis, K.; Gatos, B.; Pratikakis, I. ICFHR2014 competition on handwritten document image binarization (H-DIBCO 2014). In Proceedings of the 2014 14th International Conference on Frontiers in Handwriting Recognition, Crete, Greece, 1–4 September 2014; pp. 809–813. [Google Scholar]
  77. Pratikakis, I.; Zagoris, K.; Barlas, G.; Gatos, B. ICFHR2016 handwritten document image binarization contest (H-DIBCO 2016). In Proceedings of the 2016 15th International Conference on Frontiers in Handwriting Recognition (ICFHR), Shenzhen, China, 23–26 October 2016; pp. 619–623. [Google Scholar]
  78. Pratikakis, I.; Zagori, K.; Kaddas, P.; Gatos, B. ICFHR 2018 Competition on Handwritten Document Image Binarization (H-DIBCO 2018). In Proceedings of the 2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR), Niagara Falls, NY, USA, 5–8 August 2018. [Google Scholar]
  79. IEEE/Iet Electronic Library. Available online: https://ieeexplore.ieee.org/Xplore/home.jsp (accessed on 13 February 2023).
  80. China National Knowledge Infrastructure. Available online: https://www.cnki.net/ (accessed on 13 February 2023).
  81. Ntirogiannis, K.; Gatos, B.; Pratikakis, I. Performance evaluation methodology for historical document image binarization. IEEE Trans. Image Process. 2012, 22, 595–609. [Google Scholar] [CrossRef] [PubMed]
  82. Lu, H.; Kot, A.C.; Shi, Y.Q. Distance-reciprocal distortion measure for binary document images. IEEE Signal Process. Lett. 2004, 11, 228–231. [Google Scholar] [CrossRef] [Green Version]
  83. Bazzi, I.; Schwartz, R.; Makhoul, J. An omnifont open-vocabulary OCR system for English and Arabic. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 495–504. [Google Scholar] [CrossRef]
  84. Klakow, D.; Peters, J. Testing the correlation of word error rate and perplexity. Speech Commun. 2002, 38, 19–28. [Google Scholar] [CrossRef]
  85. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Document image enhancement problems.
Figure 1. Document image enhancement problems.
Applsci 13 07855 g001
Figure 2. (a) Background texture; (b) page smudging; (c) handwriting fading; (d) poor lighting conditions. Reproduced with permission from Lin et al. [17], CVF Conference on Computer Vision and Pattern Recognition (CVPR); published by IEEE, 2020. (e) Watermark removal; (f) deblur.
Figure 2. (a) Background texture; (b) page smudging; (c) handwriting fading; (d) poor lighting conditions. Reproduced with permission from Lin et al. [17], CVF Conference on Computer Vision and Pattern Recognition (CVPR); published by IEEE, 2020. (e) Watermark removal; (f) deblur.
Applsci 13 07855 g002aApplsci 13 07855 g002bApplsci 13 07855 g002c
Figure 3. The structure of FD-Net, constructed by Xiong et al. [27]. Reproduced with permission from Xiong et al., Springer eBook; published by Springer Nature, 2021.
Figure 3. The structure of FD-Net, constructed by Xiong et al. [27]. Reproduced with permission from Xiong et al., Springer eBook; published by Springer Nature, 2021.
Applsci 13 07855 g003
Figure 4. Model structure constructed by Jorge and Gallego [30]. The output layer consists of the activation level assigned to each input feature (white denotes maximum activation). Published under a Creative Commons Attribution 4.0 license.
Figure 4. Model structure constructed by Jorge and Gallego [30]. The output layer consists of the activation level assigned to each input feature (white denotes maximum activation). Published under a Creative Commons Attribution 4.0 license.
Applsci 13 07855 g004
Figure 5. The structure of the model constructed by Zhao et al. [31]. The whole-image generator is composed of G1 and G2. The two sub-generators are cascaded, and G2 optimizes the results of G1. Reproduced with permission from Zhao et al., Pattern Recognition; published by Elsevier, 2019.
Figure 5. The structure of the model constructed by Zhao et al. [31]. The whole-image generator is composed of G1 and G2. The two sub-generators are cascaded, and G2 optimizes the results of G1. Reproduced with permission from Zhao et al., Pattern Recognition; published by Elsevier, 2019.
Applsci 13 07855 g005
Figure 8. BEDSR-Net model structure constructed by Lin et al. [17]. It consists of two sub-networks: BE-Net for estimating the global background color of the document and SR-Net for removing shadows. Given the input shadow image, BE-Net predicts the background color. As a side product, it generates an attention map, which depicts how likely it is that each pixel belongs to the shadow-free background. Along with the input shadow image, the estimated background color and the attention map are fed into the SR-Net for determining the shadow-free version of the input shadow image. Reproduced with permission from Lin et al., CVF Conference on Computer Vision and Pattern Recognition (CVPR); published by IEEE, 2020.
Figure 8. BEDSR-Net model structure constructed by Lin et al. [17]. It consists of two sub-networks: BE-Net for estimating the global background color of the document and SR-Net for removing shadows. Given the input shadow image, BE-Net predicts the background color. As a side product, it generates an attention map, which depicts how likely it is that each pixel belongs to the shadow-free background. Along with the input shadow image, the estimated background color and the attention map are fed into the SR-Net for determining the shadow-free version of the input shadow image. Reproduced with permission from Lin et al., CVF Conference on Computer Vision and Pattern Recognition (CVPR); published by IEEE, 2020.
Applsci 13 07855 g008
Figure 9. CANet model structure constructed by Chen et al. [45]. At Stage I, the contextual feature is firstly extracted via a pre-trained DenseNet; meanwhile, the designed contextual patch matching module (CPM) is used to acquire a set of contextual matching pairs; then, a contextual feature transfer mechanism is applied to transfer contextual information from non-shadow patches to shadow patches to recover the L and A/B channels of the shadow-removal image. At Stage II, the recovered L and A/B channel information is integrated with the input shadow image and fed into a DenseUNet to generate the final shadow-removal result. Published under a Creative Commons Attribution 4.0 license.
Figure 9. CANet model structure constructed by Chen et al. [45]. At Stage I, the contextual feature is firstly extracted via a pre-trained DenseNet; meanwhile, the designed contextual patch matching module (CPM) is used to acquire a set of contextual matching pairs; then, a contextual feature transfer mechanism is applied to transfer contextual information from non-shadow patches to shadow patches to recover the L and A/B channels of the shadow-removal image. At Stage II, the recovered L and A/B channel information is integrated with the input shadow image and fed into a DenseUNet to generate the final shadow-removal result. Published under a Creative Commons Attribution 4.0 license.
Applsci 13 07855 g009
Figure 10. The structure of the model constructed by Jemni et al. [48]. Published under a Creative Commons Attribution 4.0 license.
Figure 10. The structure of the model constructed by Jemni et al. [48]. Published under a Creative Commons Attribution 4.0 license.
Applsci 13 07855 g010
Figure 11. Two-stage visible watermark removal architecture constructed by Jiang et al. [50]. Reproduced with permission from Jiang et al., IET Image Processing; published by John Wiley and Sons, 2021.
Figure 11. Two-stage visible watermark removal architecture constructed by Jiang et al. [50]. Reproduced with permission from Jiang et al., IET Image Processing; published by John Wiley and Sons, 2021.
Applsci 13 07855 g011
Figure 12. Structure of the model proposed in [54], constructed by Ge et al. In the training stage, the watermark w is expanded, reshaped, and concatenated with the cover image I c . Then, a combination of the cover image I c and the watermark w is encoded to generate the watermarked image I w . Next, I w is input into the noise layer which simulates the attacks on input, generating the noised and watermarked image I w . Finally, I w is input into Decoder D to extract the watermark w which could be similar to or the same as w. After training, Encoder E is used to embed the watermark and Decoder D is used for watermark extraction. Reproduced with permission from Ge et al., Multimedia Tools and Applications; published by Springer Nature, 2023.
Figure 12. Structure of the model proposed in [54], constructed by Ge et al. In the training stage, the watermark w is expanded, reshaped, and concatenated with the cover image I c . Then, a combination of the cover image I c and the watermark w is encoded to generate the watermarked image I w . Next, I w is input into the noise layer which simulates the attacks on input, generating the noised and watermarked image I w . Finally, I w is input into Decoder D to extract the watermark w which could be similar to or the same as w. After training, Encoder E is used to embed the watermark and Decoder D is used for watermark extraction. Reproduced with permission from Ge et al., Multimedia Tools and Applications; published by Springer Nature, 2023.
Applsci 13 07855 g012
Figure 13. Training process of blur2sharp model structure constructed by Neji et al. [65]. Published under a Creative Commons Attribution 4.0 license.
Figure 13. Training process of blur2sharp model structure constructed by Neji et al. [65]. Published under a Creative Commons Attribution 4.0 license.
Applsci 13 07855 g013
Figure 14. DeblurGAN-CNN model structure constructed by Gonwirat and Sarayut [66]. Published under a Creative Commons Attribution 4.0 license.
Figure 14. DeblurGAN-CNN model structure constructed by Gonwirat and Sarayut [66]. Published under a Creative Commons Attribution 4.0 license.
Applsci 13 07855 g014
Table 1. Degradation problems targeted by the datasets.
Table 1. Degradation problems targeted by the datasets.
DatasetTaskNo. of ImageReal vs. Synthetic
Jung’s datasetpoor lighting conditions 159real
RDSRDpoor lighting conditions540real
BMVCdeblurring3M train/35K testsynthetic
Bickley diarymultiple7real
SMADImultiple240synthetic
DIBCO and H-DIBCOmultiple10, 10real
DocImgENwatermark removal10 k train/10 k validation/10 k testsynthetic
DocImgCNwatermark removal230 k train/10 k validation/10 k testsynthetic
Table 3. Experimental results of uneven illumination.
Table 3. Experimental results of uneven illumination.
DegradationMethodJung’s DatasetsRDSRD
PSNRSSIMPSNRSSIM
poor lighting conditionsBako [42]23.700.901528.240.8664
ST-CGAN23.710.904630.310.9016
Kligler [4]24.450.833222.530.7056
Jung [44]28.490.910814.450.7054
BEDSR-Net27.230.911533.480.9084
Table 4. Defuzzification experimental results.
Table 4. Defuzzification experimental results.
DegradationDatasetMethodPSNRSSIMCER
deblurringBMVC textHradis [55]30.60.987.2
Pan [56]21.840.9335.3
Zhu [57]19.570.8953.0
Nah [58]22.270.9250.6
Lu [63]22.560.9510.1
Table 5. Degradation status problems or future developments.
Table 5. Degradation status problems or future developments.
DegradationProblems or Future Developments
background textureThe challenge of the background texture problem is mostly for color document images, and much of the current research is still focused on the processing of grayscale text images with better results. The color document images, on the other hand, cause a loss of text content when processed by model algorithms due to the difference in color. Therefore, methods for color document images need to be investigated. The text images include handwritten images and printed images, and there is still less discussion on the comparison of these two types of datasets, which can be processed and studied.
page smudgeCharacter recognition is very challenging when the image document has ghosting problems (ink is visible on the other side of the text page, but the ink is not precisely from the other side, causing interference with the text on the target page) and when the RGB image has a variety of ink colors and intensities that vary from moment to moment. Current methods are less effective for low-contrast document image bleed-through problems. Therefore, we need to propose a model approach based on such problems to solve these problems.
handwriting fadingAlthough the fading of handwriting will affect the readability of the document and the OCR effect, there are severe and minor fading cases, and minor fading is still relatively easy to identify. This type of degradation problem deals more with handwritten text images. For severely faded documents, the datasets of many studies are not published, and there is still a need to produce datasets for such problems.
the poor lighting conditionsThe problem of overexposure in uneven illumination conditions, i.e., the result of adding too much flash to devices such as cameras during digitization, has been less studied [54,55]. There is no suitable dataset for this problem. Therefore, the exposure processing problem in image document enhancement still needs sufficient research in the future.
watermark removalThere is less research on watermark removal for document images, mainly due to the differences in language and text, resulting in limited watermark removal techniques for document images. For example, English text, a combination of 26 letters, and Chinese characters are three-dimensional homogeneous characters, resulting in less effective information redundancy in Chinese characters. Additionally, most of the methods are now only applicable to black text, the watermark color of a single document. However, watermark removal involves the integrity of the original content and intellectual property issues, so we need to handle it carefully.
deblurringModels mainly generate blurred text datasets, but this is not enough to be used in real scenarios, which are diverse and many situations are not considered. The current number of blurred text datasets is still insufficient, especially since Chinese datasets are even more scarce. The insufficient coverage of scenarios in the dataset also leads to algorithmic models that do not take advantage of them.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Zuo, S.; Yang, Z.; He, J.; Shi, J.; Zhang, R. A Review of Document Image Enhancement Based on Document Degradation Problem. Appl. Sci. 2023, 13, 7855. https://doi.org/10.3390/app13137855

AMA Style

Zhou Y, Zuo S, Yang Z, He J, Shi J, Zhang R. A Review of Document Image Enhancement Based on Document Degradation Problem. Applied Sciences. 2023; 13(13):7855. https://doi.org/10.3390/app13137855

Chicago/Turabian Style

Zhou, Yanxi, Shikai Zuo, Zhengxian Yang, Jinlong He, Jianwen Shi, and Rui Zhang. 2023. "A Review of Document Image Enhancement Based on Document Degradation Problem" Applied Sciences 13, no. 13: 7855. https://doi.org/10.3390/app13137855

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop