# ensnet_ensconce_text_in_the_wild__71108ac0.pdf The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) Ens Net: Ensconce Text in the Wild Shuaitao Zhang, Yuliang Liu, Lianwen Jin, Yaoxiong Huang, Songxuan Lai School of Electronic and Information Engineering South China University of Technology lianwen.jin@gmail.com A new method is proposed for removing text from natural images. The challenge is to first accurately localize text on the stroke-level and then replace it with a visually plausible background. Unlike previous methods that require image patches to erase scene text, our method, namely ensconce network (Ens Net), can operate end-to-end on a single image without any prior knowledge. The overall structure is an end-to-end trainable FCN-Res Net-18 network with a conditional generative adversarial network (c GAN). The feature of the former is first enhanced by a novel lateral connection structure and then refined by four carefully designed losses: multiscale regression loss and content loss, which capture the global discrepancy of different level features; texture loss and total variation loss, which primarily target filling the text region and preserving the reality of the background. The latter is a novel local-sensitive GAN, which attentively assesses the local consistency of the text erased regions. Both qualitative and quantitative sensitivity experiments on synthetic images and the ICDAR 2013 dataset demonstrate that each component of the Ens Net is essential to achieve a good performance. Moreover, our Ens Net can significantly outperform previous state-of-the-art methods in terms of all metrics. In addition, a qualitative experiment conducted on the SBMNet dataset further demonstrates that the proposed method can also preform well on general object (such as pedestrians) removal tasks. Ens Net is extremely fast, which can preform at 333 fps on an i5-8600 CPU device. 1. Introduction Scene text is ubiquitous in our daily life, and it conveys valuable information. However, various private information, such as ID numbers, telephone numbers, car numbers, and home addresses (Inai et al. 2014) may easily be exposed in natural scene images. Such important private information can be easily collected automatically by the machines engaged in fraud, marketing, or other illegal activities. Therefore, a method that can ensconce the text in the wild would be beneficial. In addition to preventing privacy disclosure, Scene text erasing can also facilitate many image processing and computer vision applications, such as information Authors contributed equally as first author Corresponding author Copyright c 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: Examples of scene text erasing. Given a scene text image, the goal is to ensconce the text and replace it with a visually plausible background while retaining the non-text regions. Left: input images. Right: our results. reconstruction and visual translation. Examples of scene text erasing are shown in Fig 1. The challenges of erasing the scene text are as follows: 1) scene text erasing methods should be able to perceive the strokelevel position of the text in advance, which is more complicated than bounding-box-level scene text detection methods that have been comprehensively researched (Liang, Doermann, and Li 2005); 2) after removing text, the original text region should be filled with a new visually plausible background; 3) the non-text regions should be retained in their original appearance. To the best of our knowledge, the method proposed in (Nakamura et al. 2017) is the first and the only method that addresses the scene text erasing issue. However, their method uses cropped image patches as training data, which limits the erasing performance due to loss of global context information. We herein propose a novel method named the ensconce network (Ens Net) to address this task, which includes the following characteristics: Ens Net can erase the scene text end-to-end on the whole image, which improves the erasing results and executes very fast. A novel lateral connection structure is designed to effectively capture the detailed context information. A refined loss function including multiscale regression loss, content loss, texture loss, and total variation loss (tv) loss is proposed in the optimization stage to ensure the background reconstruction and integrity of the non-text region. A local-aware discriminator is proposed to guide the network to replace the text region with a more plausible background. In addition, we also synthesized a dataset to benchmark the scene text erasing and facilitate future research. The experimental results on both the synthetic dataset and the ICDAR 2013 dataset (Karatzas et al. 2013) demonstrate that the Ens Net can significantly outperform previous stateof-the-art methods in terms of all metrics. 2. RELATED WORK In recent years, several methods have been proposed to remove graphic text from born-digital images such as captions, subtitles, and annotations (Khodadadi and Behrad 2012; Modha and Dave 2014; Wagh and Patil 2015). However, these methods required the text to be axis-aligned, clean, and well focused, therefore they are not applicable to scene text erasing because of its complexities, uneven illuminations, perspective distortions, etc. Recently, owing to the immense success of deep learning in computer-vision tasks, (Nakamura et al. 2017) were the first to successfully design a Scene text eraser to ensconce the text in the wild. Specifically, this method used a single-scaled sliding-window-based neural network to erase the scene text. However, Nakamura et al. s method is prone to dividing large text into multiple image patches, thus significantly reducing the consistency of the erasing results. In addition, using cropped image patches as training data will restrict the operating speed to some extent. (Isola et al. 2017) proposed an extremely fast Pix2Pix method that not only learns the mapping from the input image to output image, but also introduces a novel loss function to train the mapping. Although this state-of-the-art method is not proposed specifically to handle the scene text erasing task, it can be easily reproduced for comparison. 3. METHODOLOGY 3.1. Framework Overview Fig. 2 shows the overall architecture of the proposed Ens Net, which consists of two primary parts: a generator G and a discriminator D. Given a scene text image x and the ground truth z, Ens Net attempts to produce a non-text image y that is as real as possible by solving the following optimization problem: min G max D Ex pdata(x),z [log(1 D(x, G(x, z)))] + Ex pdata(x,y),z [log D(x, y)] . (1) Following the ideas of previous GAN-based methods (Mirza and Osindero 2014; Goodfellow et al. 2014), the proposed method alternatively updates G and D, and the entire procedure is end-to-end trainable. 3.2. Generator The construction of G contains three mutually promoted modules: a lightweight Fully-Convolution-Network (FCN) backbone, lateral connections, and refined loss. Fully convolution network The FCN consists of a convolution-pathway and a deconvolution-pathway. The former utilizes a lightweight Resnet18 network. Based on the last convolutional layer of the Resnet18, a 1 1 convolutional layer converted by the last fully-connected layer is applied to predict the text/non-text score map. The deconvolution pathway consists of five deconvolutional layers, with kernel size set 4, stride step set 2, and padding size set 1 for each layer. Lateral connections It is typically considered that lower level features exhibit stronger semantics, while higher level exhibit semantically weak features but more detailed information including pixel-level color, texture, the position information of objects, etc (Shen et al. 2017; Dong, Loy, and Tang 2016). Hence, we designed the lateral connections, as shown in Fig. 3, to integrate the higher-level semantics with details from the lower layers. The proposed lateral connection includes a transforming block and an up-sampling block: 1) The transforming block starts with a shrinking layer, that reduces the feature dimensions by using a 1 1 convolution. Subsequently, two same-size (3 3) convolutional layers are stacked to perform a nonlinear transformation, that can not only replace large-kernel convolutions (Dong et al. 2016) to achieve large receptive fields, but also improve the computation efficiency. Finally, we use an expanding layer to enlarge the feature map channels by a 1 1 convolution as reverse of the shrinking operation. The transformation block takes the Residual2b to the Residual5b of the Resnet18 as input. 2) For the up-sampling block, we used a deconvolutional layer to enlarge the feature map. The up-sampled features maps are then element-wise summed with the corresponding ones from the transforming block. Additionally, unlike previous CNN-based methods, we replaced all Re LU/Leaky Re LU layers with the ELU layers (Clevert, Unterthiner, and Hochreiter 2016; Yang, Lu, and et al 2017) after each layer except the layers in the convolution pathway. The ELU layers render the generator network training more stable compared with the Re LU/Leaky Re LU layers (Nair and Hinton 2010) as it can handle large negative responses during the training process. Refined loss Our loss functions target both per-pixel reconstructed accuracy as well as composition, i.e. how smoothly the text regions can harmonize with their surrounding context. As shown in Fig. 2, there are four functions designed for our generative network: multiscale regression loss, content loss, texture loss, and total variation loss. Figure 2: The overall structure of Ens Net. The network consists of three seamlessly-connected components: 1) the refined generator backbone (enhanced by the lateral connections); 2) refined-loss modules; 3) Ens Net discriminator. The network is end-to-end trainable with extremely high efficiency. Figure 3: The structure of lateral connection. 1) The multiscale regression loss. Given an input image with text Iin, initial binary mask M (0 for non-text regions and 1 for text regions), the generator prediction Iout, and the ground truth image Igt, we extracted features from different deconvolutional layers to form outputs of different sizes. By adopting this, more contextual information from different scales can be captured. The multiscale regression loss function is defined as: Lm (M, Iout, Igt) = i=1 λi( Mi (Iout(i) Igt(i)) 1+α (1 Mi) (Iout(i) Igt(i)) 1), where Iout(i) indicates the ith output extracted from the deconvolutional pathway, Mi, Igt(i) indicates the binary mask and ground truth that have the same scale as that of Iout(i) respectively, and α weights the importance between text and non-text regions. λi is the weight for the ith scale. Concretely, the output of the last, and 3rd and 5th last layers are used, whose sizes are 1 2, and 1 of the input size, respectively. Practically, we set α to 6 and λi to 0.6, 0.8, 1.0 to put more weight at the larger scale. L1 loss is adopted on the network output for the text regions and the non-text regions, respectively. 2) The content loss. Recently, a loss function measured on different high-level features has been demonstrated effective for feature reconstructing (Johnson, Alahi, and Li 2016). To further enhance the text erasing performance, we introduce content constraints on high-level features that we term as content loss. The content loss penalizes the discrepancy between the features of the output image and the corresponding ground truth image on certain layers in the CNN. We feed an output image and the corresponding ground truth to the CNN, and enforce the response of the output image to match that of the ground truth at the predetermined feature layers of the CNN, which will facilitate the network in detecting and erasing text regions. The content loss is defined as follows: n=1 (An(Iout) An(Igt) 1 n=1 (An(Icomp) An(Igt) 1. Where, Icomp is the output image Iout, with the non-text regions of Iout being set to the ground truth. An is the activation map of the n-th selected layer. In our method, we compute the feature loss at layers pool1, pool2, and pool3 of a pretrained VGG16 (Simonyan and Zisserman 2014). 3) The texture loss. As discussed earlier, the visual quality should also be considered into the optimization function. Hence, we introduce the texture loss that ensures that the restored text regions match with the non-text regions. The loss is motivated by the recent success of neural style transfers (Gatys, Ecker, and Bethge 2015). Texture loss performs an autocorrelation (Gram matrix) (Gatys, Ecker, and Bethge 2016) on each high-level feature map before applying the L1 loss, which can be defined as follows: n=1 1 Cn Hn Wn ((An(Iout))T (An(Iout)) ((An(Igt))T (An(Igt)) 1, n=1 1 Cn Hn Wn ((An(Icomp))T (An(Icomp)) ((An(Igt))T (An(Igt)) 1, where the (Hn Wn) Cn is the shape of the high-level activation map, An. Similar to the idea proposed in (Zhang, Sindagi, and Patel 2017) , we aim to penalize the discrepancy of the texture appearance of text regions and non-text regions, such that the network can capture the globally style features for more reasonably displacing the text regions. Again, we include loss terms for the raw output Iout and the output of text-erased regions Icomp. 4) The total variation loss. The last Ltv (Johnson, Alahi, and Li 2016) is targeted at global denoising, as defined below: i,j Ii,j out Ii+1,j out 1+ Ii,j out Ii,j+1 out 1, (6) where i, j is the pixel position. To exploit all merits, we combine these four novel losses together with the appropriate weights to form the refined loss function, which is defined as the following: Lrefined = LM + λe LC + λi LT + λt Ltv. (7) The hyper-parameters λe, λi, and λt in Eq. 7 control the balance between the four losses. For our model, not all the loss weighting schemes for the refined loss will generate satisfactory results. Fig. 4 shows the result of the model trained with small or large weights of content loss and texture loss. In our experiments, λe, λi, and λt are empirically set to 0.5, 50.0, and 25.0, respectively. It is noteworthy that such parameters setting works well when generalizing to other tasks, which is discussed in Section. 4.5. 3.3. Discriminator To differentiate fake images from real ones, the original GANs (Goodfellow et al. 2014) discriminate the results based on the whole image level. However, the non-text regions occupy a large proportion of the image and are typically real, which makes the discriminator difficult to focus on the text regions. Therefore, it is intuitive to design a localaware discriminator to attentively maintain the consistency of text-erased regions and their surrounding texture. The proposed local-aware discriminator only penalizes the erased text patches. Specifically, following the Patch GAN (Isola et al. 2017), we execute our discriminator across Figure 4: In the top row, images from left to right: input image; result with small content loss and texture loss weights; result with refined loss. In the bottom row, images from left to right: input image; result using large content loss and texture loss weights; result with refined loss. an image and obtain an S S (S=62) feature tensor. By designation, each entry of the tensor corresponds to a N N (N = 70) patch (i.e., receptive filed) of the image, and is assigned a label to indicate its realness. The key novelty of our discriminator is the labeling strategy. For the typical condition discriminative network (Isola et al. 2017), they will assign negative labels for any generated images, which is contrary to our purpose. Therefore, we assign a locality-dependent label for the each position of the 62*62 tensor according to the text mask. label = 0, if sum(M) > 0, 1, otherwise. (8) Where M is the binary mask (as mentioned in the above section). Given the discriminator prediction Pi, and the label Li for each N N patch, we define the loss as follows: N N (1 Li)(log(Pi)). (9) The scaling factor sum (Mi) applies the appropriate scaling to adjust for the varying amount of text-erased regions. Our underlying idea is to guide our discriminator to focus on regions indicated by the text mask. The structure of the discriminative architecture is shown in Fig. 2. 4. EXPERIMENT To quantitatively evaluate the performance of scene text erasing methods, a dataset should provide both text images with stroke-level ground truths and real background images without text. However, such text dataset does not exist currently; thus, we constructed a synthetic dataset for evaluating the performance. In addition, we evaluated the performance on the ICDAR 2013 (Karatzas et al. 2013) dataset following the same metric as reported by (Nakamura et al. 2017) for Figure 5: Illustration of improvements obtained using different components of our network. Images from left to right: (a): input images; (b): erased results by baseline; (c): erased results by lateral connections; (d): erased results of refined loss; (e): erased result by the complete architecture (Ens Net). a fair comparison; the metric utilizes an auxiliary detector to validate whether the text of images can be detected after processing by the erasing method. An ablation study is also conducted to evaluate each component of Ens Net. 4.1. Dataset and evaluation metrics Synthetic data We apply text synthesis technology (Gupta, Vedaldi, and Zisserman 2016) on scene images to generate samples, as shown in the first row of Fig. 6. Compared to the inpainting algorithm (Criminisi, P erez, and Toyama 2004), the label generated by the synthesis process is more accurate and reasonable. In our experiments, the training set consists of a total of 8000 images and the test set contains 800 images; all the training and test samples are resized to 512 512. The synthetic data are available at: https://github.com/HCIILAB/Scene-Text-Removal. Real-word dataset ICDAR 2013 (Karatzas et al. 2013) is a widely used dataset for scene text detection, containing 229 images for training, and 223 for testing. ICDAR 2017 MLT (Nayef et al. 2017) is a benchmark multi-lingual scene text dataset. In our experiment, we collected 1000 images from the ICDAR 2017 MLT subdataset which only contains English text to enlarge real data, and the background image (label) is generated by manually erasing the text. An example is shown in Fig. 6. For a fair comparison, all our method and counterpart methods strictly use the same training set and hyper-parameters. Evaluation metrics (Nakamura et al. 2017) proposed a new evaluation method that utilizes a baseline detector to adversely evaluate the erased results by computing how low the precision is, as well as the recall and f-score of the ICDAR 2013 test set (Karatzas et al. 2013). However, this method Figure 6: Examples of training samples for Ens Net learning. Top: synthetic sample. Bottom: real sample. does not consider the quality of the erased image, i.e, if the image contents are completely erased, the result would be the best, which is not reasonable. Therefore, to impartially compare the proposed method with other methods on the synthetic dataset, we adopted the previous image inpainting (Yang, Lu, and et al 2017; Liu et al. 2018) metrics to evaluate our method, which includes the following: 1) ℓ2 error, PSNR, which computes the peak signal to noise ratio (PSNR) for an image; 2) SSIM (Wang et al. 2004), which computes the mean structural similarity index between two images; 3) AGE, which calculates the average of the graylevel absolute difference between the ground truth and the computed background image; 4) p EPs, which calculate the percentage of error pixels; 5) p CEPS, which calculates the percentage of clustered error pixels (number of pixels whose four-connected neighbors are also error pixels). A higher SSIM and PSNR or lower AGE, p EPs, p CEPS and the ℓ1 error represent better results. For the real dataset, because we do not have ground truth background images, we follow the same procedure as reported by (Nakamura et al. 2017) to calculate the precision, recall, and f-score. Additionally, visual estimates are also used on real dataset to qualitatively compare the performance of various methods. 4.2. Ablation study In this section, we apply the sensitivity analysis to verify the contributions of different components of Ens Net: Baseline, Lateral Connections, Refined Loss, and Ens Net. All experiments use exactly the same settings (input size is set to 512 512). The quantitative results are shown in Tab. 1, and the qualitative results can be visualized in Fig. 5. Baseline We directly apply fully convolution network to train baseline model only with L1 loss. Lateral connections In Tab. 1, compared with the baseline model, the lateral connections obtain a lower ℓ2 error. This implies that the lateral connections can effectively learn more detailed information to restore the image detail. Meanwhile, from the Fig. 5, we found that lateral connections are conducive for our generative network to achieve clearer results. Refined Loss Tab. 1 shows that the refined loss can improve erasing result on any type of metric mentioned above. Some intuitive examples are shown in Fig. 5, where the fish scale artifacts or blocky checkerboard artifacts caused by c GAN are ameliorated by our refined loss, thereby demonstrating the effectiveness of using the proposed refined loss. Ens Net By seamlessly connecting with the proposed local-awareness discriminative network, the performance of Ens Net can be further improved, as shown in both Tab. 1 and Fig. 5. Ens Net alleviates the dilemma of the extreme imbalance between the text regions and the non-text regions; thus, it can attentively replace the text with a more plausible background. 4.3. Comparison with state-of-the-art methods We compared the performance of the proposed Ens Net with the relevant and recent state-of-the-art methods: Pix2Pix (Isola et al. 2017) and Scene text eraser (Nakamura et al. 2017). The results on the synthetic and real dataset are shown in Tab. 1, and demonstrate that the proposed Ens Net can significantly outperform previous state-of-the-art methods (Isola et al. 2017; Nakamura et al. 2017) in all metrics. Tab. 2 adopts two different detection methods (Zhou, Yao, and et al 2017) (Fu et al. 2017) and two protocols: the Det Eval and the ICDAR 2013 evaluation (Karatzas et al. 2013; Wolf and Jolion 2006) to further demonstrate the robustness of our method. The speeds of different methods are also listed in Tab 1. Our method can achieve 333 fps on a i58600 CPU device. Although Ens Net is slightly slower than Pix2Pix (Isola et al. 2017) , it is twice faster than the scene text eraser (Nakamura et al. 2017). Furthermore, we qualitatively compared the results generated by different methods, as shown in Fig. 7. Based on the observation, we found that Pix2Pix (Isola et al. 2017) can only remove parts of the text region, and the restored regions contain many fish scale artifacts. As for the scene text eraser (Nakamura et al. 2017), most of the text in the images can not be erased completely; meanwhile, some non-text backgrounds are also influenced. In contrast, the effectiveness of our method can be intuitively visualized in Fig. 7. The results demonstrate that the Ens Net can not only almost perfectly erase the text regions (even large-size text) but also maintain the reality of the backgrounds. Figure 7: Results of comparing a few different methods on sample images from a real dataset. Images from left to right: input image, Pix2Pix (Isola et al. 2017), Scene text eraser (Nakamura et al. 2017), and our method. 4.4. Generalization to general object removal In this section, an experiment on the general object removal task is conducted to test the generalization ability of Ens Net. We evaluate Ens Net on a well-known Scene Background Modeling (SBMnet) dataset (Jodoin et al. 2017), which is targeted at background estimation algorithms. However, SBMnet is a video-based dataset, implying that the methods can take advantage of the sequence frames to reconstruct all the background information. Therefore, it is not fair to quantitatively compare our method with other methods because Table 1: Quantitative evaluation results of different components of our method and other methods on synthetic and real datasets. B: Baseline; LC: lateral connections; RL: refined loss; TD: text discriminator; R: recall; P: precision; F: f-measure. RPF metric is calculated by ICDAR 2013 Evaluation, and DSSD (Fu et al. 2017) is adopted as an auxiliary detector. Note that although the results of the baseline contain lower values in RPF, it is significantly worse than other methods in terms of other metrics such PSNR and SSIM (%). This implies that the previously used RPF metric is not a perfect metric (visualization results and other metrics should also be taken into considered). Method Synthesis ICDAR 2013 FPS PSNR SSIM(%) ℓ2 error AGE p EPs p CEPS R P F R P F Original images - - - - - - 53.65 68.21 60.06 86.11 95.66 90.64 - B 20.30 62.18 0.4256 13.84 0.1918 0.1091 0.10 1.41 0.19 2.28 41.67 4.33 - B + LC 24.83 86.74 0.0627 5.48 0.0538 0.0385 26.72 40.89 32.32 51.91 79.20 62.71 - B + LC + RL 30.23 92.78 0.0234 2.51 0.0227 0.0121 8.35 26.59 12.71 14.64 72.21 24.34 - B + LC + RL + TD (Ens Net) 37.36 96.44 0.0021 1.73 0.0069 0.0020 1.14 23.45 2.17 12.58 70.00 21.33 333 Erased by Pix2Pix (2017) 25.60 89.86 0.2465 5.60 0.0571 0.0423 28.31 65.30 39.50 32.33 85.08 46.84 382 Erased by Scene Text Eraser (2017) 14.68 46.13 0.7148 13.29 0.1859 0.0936 22.35 30.12 25.66 34.48 60.57 43.95 166 Table 2: Comparison among previous methods and proposed method. R: Recall; P: Precision; F: F-measure. Image dataset Synthesis ICDAR 2013 Evaluation protocol ICDAR Eval Det Eval ICDAR Eval Det Eval Methods R P F R P F R P F R P F DSSD (2017) Original image 53.65 68.21 60.06 55.19 68.94 61.30 86.11 95.66 90.64 86.08 95.87 90.71 Erased by Pix2Pix (2017) 28.31 65.30 39.50 29.54 65.89 40.79 32.33 85.08 46.85 33.97 85.95 48.70 Erased by Scene text eraser (2017) 22.35 30.12 25.66 22.67 30.76 26.10 34.48 60.57 43.95 35.27 61.37 44.79 Erased by Our 1.14 23.45 2.17 1.21 23.45 2.29 12.58 70.00 21.33 13.46 70.62 22.61 EAST (2017) Original image 51.22 79.10 62.18 52.50 79.14 63.12 70.10 81.50 75.37 70.70 81.61 75.77 Erased by Pix2Pix (2017) 21.03 77.12 33.05 21.78 77.44 34.00 10.19 69.45 17.78 10.37 69.45 18.05 Erased by Scene text eraser (2017) 24.30 40.30 30.32 24.74 40.61 30.75 10.08 39.09 16.03 10.35 39.09 16.37 Erased by Our 0 0 0 0 0 0 5.66 73.42 10.51 5.75 73.42 10.67 our method uses only one static image as the input without any context information of other frames. It is noteworthy that this is a novelty of Ens Net that enables a plausible background to be reconstructed. The qualitative results are shown in Fig. 8. Without specific tuning, our method can still generalize excellently to the pedestrian removal task. Figure 8: Experimental results on pedestrian removal. Images from left to right: input images and removal results. 5. CONCLUSION This paper proposes Ens Net - a novel method that can effectively erase text from scene text images. All components of Ens Net are proved crucial to achieve good erasing results, including 1) lateral connection, which can effectively integrate high-level semantics and low-level detailed information; 2) a novel refined loss, which can remarkably reduce the artifacts; and 3) a simple but highly effective local-aware discriminator, which ensures the local consistency of the text-erased regions. We also constructed a new synthetic dataset for the scene text erasing task, which will be published for future research. To the best of our knowledge, Ens Net is the first method that can be trained end-to-end to directly erase the text on a whole image level, thus achieving outstanding erasing performance on a synthetic dataset and the ICDAR 2013 dataset. It also significantly outperformed previous state-ofthe-art methods in terms of all metrics. Furthermore, a qualitative experiment conducted on SBMNet proved that our method could perform well in general object removal tasks, and further investigations would be carried out in future research. Acknowledgement This research is supported in part by GD-NSF (no. 2017A030312006), the National Key Research and Development Program of China (No. 2016YFB1001405), NSFC (Grant No.: 61472144, 61673182, 61771199), and GDSTP (Grant No.: 2017A010101027), GZSTP(no. 201607010227). References Clevert, D.-A.; Unterthiner, T.; and Hochreiter, S. 2016. Fast and accurate deep network learning by exponential linear units (elus). ICLR. Criminisi, A.; P erez, P.; and Toyama, K. 2004. Region filling and object removal by exemplar-based image inpainting. IEEE Transactions on image processing 13(9):1200 1212. Dong, C.; Chen, C. L.; He, K.; and Tang, X. 2016. Image super-resolution using deep convolutional networks. IEEE Trans Pattern Anal Mach Intell 38(2):295 307. Dong, C.; Loy, C. C.; and Tang, X. 2016. Accelerating the super-resolution convolutional neural network. In European Conference on Computer Vision, 391 407. Springer. Fu, C.-Y.; Liu, W.; Ranga, A.; Tyagi, A.; and Berg, A. C. 2017. DSSD: Deconvolutional single shot detector. In ar Xiv preprint ar Xiv:1701.06659. Gatys, L. A.; Ecker, A. S.; and Bethge, M. 2015. A neural algorithm of artistic style. ar Xiv preprint ar Xiv:1508.06576. Gatys, L. A.; Ecker, A. S.; and Bethge, M. 2016. Image style transfer using convolutional neural networks. In CVPR, 2414 2423. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; and Bengio, Y. 2014. Generative adversarial nets. In Advances in neural information processing systems, 2672 2680. Gupta, A.; Vedaldi, A.; and Zisserman, A. 2016. Synthetic data for text localisation in natural images. In CVPR, 2315 2324. Inai, K.; Frinken, V.; Feng, Y.; and Uchida, S. 2014. Selective concealment of characters for privacy protection. In ICPR, 333 338. Isola, P.; Zhu, J. Y.; Zhou, T.; and Efros, A. A. 2017. Imageto-image translation with conditional adversarial networks. CVPR. Jodoin, P. M.; Maddalena, L.; Petrosino, A.; and Wang, Y. 2017. Extensive benchmark and survey of modeling methods for scene background initialization. IEEE Transactions on Image Processing PP(99):1 1. Johnson, J.; Alahi, A.; and Li, F. F. 2016. Perceptual losses for real-time style transfer and super-resolution. In ECCV, 694 711. Karatzas, D.; Shafait, F.; Uchida, S.; and et al. 2013. Icdar 2013 robust reading competition. In ICDAR, 1484 1493. Khodadadi, M., and Behrad, A. 2012. Text localization, extraction and inpainting in color images. In 2012 20th Iranian Conference on, Electrical Engineering (ICEE), 1035 1040. IEEE. Liang, J.; Doermann, D.; and Li, H. 2005. Camera-based analysis of text and documents: a survey. International journal on document analysis and recognition 7(2):84 104. Liu, G.; Reda, F. A.; Shih, K. J.; Wang, T.-C.; Tao, A.; and Catanzaro, B. 2018. Image inpainting for irregular holes using partial convolutions. In ECCV. Mirza, M., and Osindero, S. 2014. Conditional generative adversarial nets. ar Xiv preprint ar Xiv:1411.1784. Modha, U., and Dave, P. 2014. Image inpainting-automatic detection and removal of text from images. International Journal of Engineering Research and Applications (IJERA), ISSN. Nair, V., and Hinton, G. E. 2010. Rectified linear units improve restricted boltzmann machines. In International Conference on International Conference on Machine Learning, 807 814. Nakamura, T.; Zhu, A.; Yanai, K.; and Uchida, S. 2017. Scene text eraser. In ICDAR. Nayef, N.; Yin, F.; Bizid, I.; Choi, H.; Feng, Y.; and et al. 2017. Icdar2017 robust reading challenge on multi-lingual scene text detection and script identification-rrc-mlt. In ICDAR, 1454 1459. IEEE. Shen, X.; Chen, Y. C.; Tao, X.; and Jia, J. 2017. Convolutional neural pyramid for image processing. In CVPR. Simonyan, K., and Zisserman, A. 2014. Very deep convolutional networks for large-scale image recognition. ar Xiv preprint ar Xiv:1409.1556. Wagh, P. D., and Patil, D. 2015. Text detection and removal from image using inpainting with smoothing. In 2015 International Conference on, Pervasive Computing (ICPC), 1 4. IEEE. Wang, Z.; Bovik, A. C.; Sheikh, H. R.; and Simoncelli, E. P. 2004. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing 13(4):600 612. Wolf, C., and Jolion, J. M. 2006. Object count/area graphs for the evaluation of object detection and segmentation algorithms. International journal on document analysis and recognition 8(4):280 296. Yang, C.; Lu, X.; and et al. 2017. High-resolution image inpainting using multi-scale neural patch synthesis. In CVPR, volume 1, 3. Zhang, H.; Sindagi, V.; and Patel, V. M. 2017. Image deraining using a conditional generative adversarial network. ar Xiv preprint ar Xiv:1701.05957. Zhou, X.; Yao, C.; and et al. 2017. East: An efficient and accurate scene text detector (2017). In CVPR.