# deep_automatic_natural_image_matting__4afd662a.pdf Deep Automatic Natural Image Matting Jizhizi Li1 , Jing Zhang1 and Dacheng Tao2 1The University of Sydney, Australia 2JD Explore Academy, JD.com, China jili8515@uni.sydney.edu.au, jing.zhang1@sydney.edu.au, dacheng.tao@gmail.com Automatic image matting (AIM) refers to estimating the soft foreground from an arbitrary natural image without any auxiliary input like trimap, which is useful for image editing. Prior methods try to learn semantic features to aid the matting process while being limited to images with salient opaque foregrounds such as humans and animals. In this paper, we investigate the difficulties when extending them to natural images with salient transparent/meticulous foregrounds or non-salient foregrounds. To address the problem, a novel endto-end matting network is proposed, which can predict a generalized trimap for any image of the above types as a unified semantic representation. Simultaneously, the learned semantic features guide the matting network to focus on the transition areas via an attention mechanism. We also construct a test set AIM-500 that contains 500 diverse natural images covering all types along with manually labeled alpha mattes, making it feasible to benchmark the generalization ability of AIM models. Results of the experiments demonstrate that our network trained on available composite matting datasets outperforms existing methods both objectively and subjectively. The source code and dataset are available at https://github.com/Jizhizi Li/AIM. 1 Introduction Natural image matting refers to estimating a soft foreground from a natural image, which is a fundamental process for many applications, e.g., film post-production and image editing [Chen et al., 2013; Zhang and Tao, 2020]. Since image matting is a highly ill-posed problem, previous methods usually adopt auxiliary user input, e.g. trimap [Sun et al., 2004; Cai et al., 2019], scribble [Levin et al., 2007], or background image [Sengupta et al., 2020] as constraints. While traditional methods estimate the alpha value by sampling neighboring pixels [Wang and Cohen, 2007] or defining affinity metrics for alpha propagation [Levin et al., 2008], deep learning-based approaches solve it by learning discriminative representations from a large amount of labeled data and predicting alpha matte directly [Lu et al., 2019; Li and Lu, 2020]. Figure 1: Matting results of GFM and our method on three types of natural images, i.e., SO, STM, NS, from the top to the bottom. However, extra manual effort is required while generating such auxiliary inputs, which makes these methods impractical in automatic industrial applications. To address the limitations, automatic image matting (AIM) has attracted increasing attention recently [Zhang et al., 2019], which refers to automatically extracting the soft foreground from an arbitrary natural image. Prior AIM methods [Qiao et al., 2020; Li et al., 2020] solve this problem by learning semantic features from the image to aid the matting process but is limited to images with salient opaque foregrounds, e.g., human [Shen et al., 2016; Chen et al., 2018], animal [Li et al., 2020]. It is difficult to extend them to images with salient transparent/meticulous foregrounds or non-salient foregrounds due to their limited semantic representation ability. To address this issue, we dig into the problem of AIM by investigating the difficulties of extending it to all types of natural images. First, we propose to divide natural images into three types according to the characteristics of foreground alpha matte. As shown in Figure 1, the three types are: 1) SO Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) (Salient Opaque): images that have salient foregrounds with opaque interiors, e.g. human, animal; 2) STM (Salient Transparent/Meticulous): images that have salient foregrounds with transparent or meticulous interiors, e.g. glass, plastic bag; and 3) NS (Non-Salient): images with non-salient foregrounds, e.g. smoke, grid, raindrop. Some examples are shown in Figure 1. Then, we systematically analyze the ability of baseline matting models for each type of image in terms of understanding global semantics and local matting details. We find that existing methods usually learn implicit semantic features or use an explicit semantic representation that is defined for a specific type of image, i.e., SO. Consequently, they are inefficient to handle different types of images with various characteristics in foreground alpha mattes, e.g., salient opaque/transparent and non-salient foregrounds. In this paper, we make the first attempt to address the problem by devising a novel automatic end-to-end matting network for all types of natural images. First, we define a simple but effective unified semantic representation for the above three types by generalizing the traditional trimap according to the characteristics of foreground alpha matte. Then, we build our model upon the recently proposed effective GFM model [Li et al., 2020] with customized designs. Specifically, 1) we use the generalized trimap as the semantic representation in the semantic decoder to adapt it for all types of images; 2) we exploit the effective SE attention [Hu et al., 2018] in the semantic decoder to learn better semantic features to handle different characteristics of foreground alpha mattes; 3) we improve the interaction between the semantic decoder and matting decoder by devising a spatial attention module, which guides the matting decoder to focus on the details in the transition areas. These customized designs prove to be effective for AIM rather than trivial tinkering, as shown in Figure 1. Besides, there is not a test bed for evaluating AIM models on different types of natural images. Previous methods use composition images by pasting different foregrounds on background images from COCO dataset [Lin et al., 2014] for evaluation, which may introduce composition artifacts and are not representative for natural images as mentioned in [Li et al., 2020]. Some recent works collect natural test images with manually labeled alpha mattes, however, they are limited to specific types of images, such as portrait images [Ke et al., 2020] or animal images [Li et al., 2020], which are not suitable for comprehensive evaluation of matting models for AIM. To fill this gap, we establish a benchmark AIM-500 by collecting 500 diverse natural images covering all three types and many categories and manually label their alpha mattes. The main contributions of this paper are threefold: 1) we make the first attempt to investigate the difficulties of automatic natural image matting for all types of natural images; 2) we propose a new matting network with customized designs upon a reference architecture that are effective for AIM on different types of images; and 3) we establish the first natural images matting benchmark AIM-500 by collecting 500 natural images covering all three types and manually labeling their alpha mattes, which can serve as a test bed to facilitate future research on AIM. 2 Rethinking the Difficulties of AIM Matting without auxiliary inputs. Prevalent image matting methods [Sun et al., 2021; Cho et al., 2016] solve the problem by leveraging auxiliary user inputs, e.g. trimap [Tang et al., 2019], scribble [Levin et al., 2007], or background image [Sengupta et al., 2020], which provide strong constraints on the solution space. Specifically, given a trimap, the matting models only need to focus on the transition area and distinguish the details by leveraging the available foreground and background alpha matte information. However, there is usually little chance to obtain auxiliary information in real-world automatic application scenarios. Thereby, AIM is more challenging since the matting models need to understand the holistic semantic partition of foreground and background of an image, which may belong to different types as been described previously. Nevertheless, AIM is more appealing to automatic applications and worth more research efforts. Matting on natural images. Since it is difficult to label accurate alpha mattes for natural images at scale, there are no publicly available large-scale natural image matting datasets. Usually, foreground images are obtained by leveraging chroma keying from images captured with a green background screen [Xu et al., 2017]. Nevertheless, the amount of available foregrounds is only about 1,000. To build more training images, they are synthesized with different background images from public datasets like COCO [Lin et al., 2014]. However, synthetic images contain composite artifacts and semantic ambiguity [Li et al., 2020]. Matting models trained on them may have a tendency to find cheap features from these composite artifacts and thus overfit on the synthetic training images, resulting in a poor generalization performance on real-world natural images. To address this domain gap issue between synthetic training images and natural test images, some efforts have been made in [Li et al., 2020; Hou and Liu, 2019]. As for evaluation, previous methods [Zhang et al., 2019; Qiao et al., 2020] also adopt synthetic dataset [Xu et al., 2017] for evaluating AIM models, which is a bias evaluation setting. For example, a composite image may contain multiple objects including the original foreground object in the candidate background image, but the ground truth is only the single foreground object from the foreground image. Besides, the synthetic test images may also contain composite artifacts, making it less possible to reveal the overfitting issue. Matting on all types of images. AIM aims to extract the soft foreground from an arbitrary natural image, which may have foreground objects with either opaque interior, transparent/meticulous interior, or non-salient foregrounds like texture, fog, or water drops. However, existing AIM methods are limited to a specific type of images with opaque foregrounds, e.g., human [Shen et al., 2016; Chen et al., 2018; Zhang et al., 2019; Li et al., 2021; Ke et al., 2020] and animals [Li et al., 2020]. DAPM [Shen et al., 2016] first generates a coarse foreground shape mask and uses it as an auxiliary input for following matting process. Late Fusion [Zhang et al., 2019] predicts foreground and background separately and uses them in a subsequent fusion process. HAtt Mat- Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) ting [Qiao et al., 2020] predicts foreground profile and guides the matting process to refine the precise boundary. GFM [Li et al., 2020] predicts foreground, background, and transition area simultaneously and combines them with the matting result at the transition area as the final matte. Although they are effective for images with salient and opaque foregrounds, extending them to all types of images is not straightforward. In the context of AIM, the term semantic is more related to foreground and background rather than the semantic category of foreground objects. Since there are very different characteristics for the three types of images, it is hard to learn useful semantic features to recognize foreground and background, especially for images with transparent objects or non-salient foregrounds, without explicit supervisory signals. 3 Proposed Methods 3.1 Improved Backbone for Matting In this work, we choose Res Net-34 [He et al., 2016] as the backbone network due to its lightweight and strong representation ability. However, Res Net-34 is originally designed to solve high-level classification problem, while both highlevel semantic features and low-level detail features should be learned to solve AIM. To this end, we improve the vanilla Res Net-34 with a simple customized modification to make it better suitable for the AIM task. In the first convolutional layer conv1 in Res Net-34, they use a stride of 2 to reduce the spatial dimension of feature maps. Followed by a max-pooling layer, the output dimension is 1/4 of the original image after the first block. While these two layers can significantly reduce computations and increase the receptive field, it may also lose many details, which are not ideal for image matting. To customize the backbone for AIM, we modify the stride of conv1 from 2 to 1 to keep the spatial dimension of the feature map as the original image size to retain local detail features. To retain the receptive field, we add two max-pooling layers with a stride of 2. Besides, we change the stride of all the first convolutional layers in stage1-stage4 of Res Net-34 from 2 to 1 and add a max-pooling layer with a stride of 2 accordingly. It is noteworthy that the indices in the max-pooling layers are also kept and will be used in the corresponding max-unpooling layers in the local matting decoder to preserve local details, as shown in Figure 3. We retrain the customized Res Net-34 on Image Net and use it as our backbone network. Experiment results demonstrate it outperforms the vanilla one in terms of both objective metrics and subjective visual quality. 3.2 Unified Semantic Representation As discussed in Section 2, the characteristics of different types of images are very different. In order to provide explicit semantic supervisory signals for the semantic decoder to learn useful semantic features and partition the image into foreground, background, and transition areas, we propose a unified semantic representation. For an image belonging to SO type, there are always foreground, background, and transition areas. Thereby, we adopt the traditional trimap as the semantic representation, which can be generated by erosion and dilation from the ground Figure 2: (a) Three types of images. (b) The traditional trimap representation. (c) The unified semantic representations, i.e., trimap, duomap, and unimap, respectively. White: foreground, Black: background, Gray: transition area. truth alpha matte. For an image belonging to STM type, there are no explicit foreground areas. In other words, the foreground should be marked as a transition area and a soft foreground alpha matte should be estimated. Thereby, we use a duomap as its semantic representation, which is a 2-class map denoting the background and transition areas accordingly. For an image belonging to NS type, it is hard to mark all the explicit foreground and background areas, since the foreground is always entangled with the background. Thereby, we use a unimap as its semantic representation, which is a 1-class map denoting the whole image as the transition area. To use a unified semantic representation for all the three types of images, we derive the trimap, duomap, and unimap from the traditional trimap as follows: Ui = Ti, type = SO Ui = 1.5Ti T 2 i , type = STM Ui = 0.5, type = NS , (1) where T represents the traditional trimap obtained by erosion and dilation from the ground truth alpha matte. For every pixel i, Ti {0, 0.5, 1}, where the background area is 0, the foreground area is 1, and the transition area is 0.5, respectively. Ui is the unified semantic representation at pixel i. Note that for images belonging to type STM and NS, the traditional trimaps always contain trivial foreground and background pixels as shown in Figure 2(b), which are very difficult for the semantic decoder to predict. Instead, our unified representations are a pure duomap or unimap as shown in Figure 2(c), which represents the holistic foreground objects or denotes there are no salient foreground objects. In order to predict the unified semantic representation, we redesign the semantic decoder in GFM [Li et al., 2020] as shown in Figure 3. Specifically, we also use five blocks in the decoder. For each decoder block, there are three sequential 3 3 convolutional layers and an upsampling layer. To further increase the capability of the decoder for learning discriminative semantic features, we adopt the Squeeze-and Excitation (SE) attention module [Hu et al., 2018] after each Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) Figure 3: The structure of our matting network for AIM. decoder block to re-calibrate the features, thereby selecting the most informative features for predicting unified representations and filtering out the less useful ones. We also adopt pyramid pooling module (PPM) [Zhao et al., 2017] to enlarge the receptive field. The upsampled PPM features are concatenated with the output of SE module and used as the input for the next decoder block. We use the cross-entropy loss to supervise the training of the semantic decoder. 3.3 Guided Matting Process As shown in Figure 3, there are six blocks following a UNet structure [Ronneberger et al., 2015] in matting decoder.. Each block contains three sequential 3 3 convolutional layers. The encoder feature is concatenated with first decoder block output and fed into the following block. The output then being concatenated with the corresponding encoder output and passes a max unpooling layer with reversed indices to recover fine structural details, serves as input of next block. Motivated by previous study proving that attention mechanism can provide support on learning discriminating representations [Ma et al., 2020], we devise a spatial attention module to guide the matting process by leveraging the learned semantic features from the semantic decoder to focus on extracting details only within transition area. Specifically, the output feature from the last decoder block in the semantic decoder is used to generate spatial attention, since it is more related to semantics. Then, it goes through a max-pooling layer and average pooling layer along the channel axis, respectively. The pooled features are concatenated and go through a convolutional layer and a sigmoid layer to generate a spatial attention map. We use it to guide the matting decoder to attend to the transition area via an element-wise production operation and an element-wise sum operation. Given the predicted unified representation U and guided matting result M from the semantic decoder and matting decoder, we can derive the final alpha matte α as follows: α = (1 2 |U 0.5|) M + 2 |U 0.5| U. (2) We adopt the commonly used alpha loss [Xu et al., 2017] and Laplacian loss [Hou and Liu, 2019] on the predicted alpha matte M and the final alpha matte α to supervise the matting decoder. Besides, we also use the composition loss [Xu et al., 2017] on the final alpha matte α. 4 Experiment 4.1 Benchmark: Automatic Image Matting-500 As discussed in Section 2, previous work [Zhang et al., 2019; Qiao et al., 2020] evaluate their models either on synthetic test set such as Comp-1k [Xu et al., 2017] or in-house test set of natural images limited to specific types, e.g., human and animal images [Li et al., 2020; Chen et al., 2018]. In this paper, we establish the first natural image matting test set AIM500, which contains 500 high-resolution real-world natural images from all three types and many categories. We collect the images from free-license websites and manually label the alpha mattes with professional software. The shorter side of each image is at least 1080 pixels. In Table 1, we compare AIM-500 with other matting test sets including DAPM [Shen et al., 2016], Comp-1k [Xu et al., 2017], HAtt [Qiao et al., 2020], and AM-2k [Li et al., 2020] in terms of the volume, whether or not do they provide natural original images, the amount of three types images, and the object classes. We can see that AIM-500 is larger and more diverse than others, making it suitable for benchmarking AIM models. AIM-500 contains 100 portrait images, 200 animal images, 34 images Dataset Volume Natural SO STM NS Class DAPM 200 200 0 0 Portrait Comp-1k 50 28 17 5 Mixed HAtt 50 30 11 9 Mixed AM-2k 200 200 0 0 Animal AIM-500 500 424 43 33 Mixed Table 1: Comparison between AIM-500 with other matting test set. Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) Figure 4: Some examples from our AIM-500 benchmark. with transparent objects, 75 plant images, 45 furniture images, 36 toy images, and 10 fruit images. We present some examples and their alpha mattes in Figure 4. Note that due to the privacy concern, all the portrait images have no identifiable information and are ready to release. 4.2 Implementation Details We trained our model and other representative matting models on the combination of matting datasets Comp-1k [Xu et al., 2017], HAtt [Qiao et al., 2020] and AM-2k [Li et al., 2020]. To reduce the domain gap issue, we adopted the highresolution background dataset BG-20k and the composition route RSSN proposed in [Li et al., 2020] to generate the training data. We composited each foreground image from Composition-1k and HAtt with five background images and each foreground image from AM-2k with two background images. The total number of training images is 8,710. To achieve better performance, we also adopted a type-wise data augmentation and transfer learning strategy during training. Type-wise data augmentation. After inspecting realworld natural images, we observed that NS images usually have a bokeh effect on the background. To simulate this effect, for all NS images, we added the blur effect on the background as RSSN to make the foreground prominent. Transfer learning. Although data augmentation could increase the number of training data, the number of original foreground images and their classes are indeed small. To mitigate the issue, we leveraged the salient object detection dataset DUTS [Wang et al., 2017] for training since it contains more real-world images and classes. However, since images in DUTS have small resolutions (about 300 400) and little fine details, we only used it for pre-training and adopted a transfer learning strategy to finetune the pre-trained model further on the above synthetic matting dataset. We trained our model on a single NVIDIA Tesla V100 GPU with batch size as 16, Adam as optimizer. When pretraining on DUTS, we resized all images to 320 320, set learning rate as 1 10 4, trained for 100 epochs. During finetuning on the synthetic matting dataset, we randomly crop a patch with a size in {640 640, 960 960, 1280 1280} from each image and resized it to 320 320, set the learning rate as 1 10 6, and trained for 50 epochs. It took about 1.5 days to train the model. We adopted the hybrid test strategy [Li et al., 2020] with the scale factors 1/3 and 1/4, respectively. 4.3 Objective and Subjective Results We compare our model with several state-of-the-art matting methods including SHM [Chen et al., 2018], LF [Qiao et al., 2020], HAtt [Qiao et al., 2020], GFM [Li et al., 2020], DIM [Xu et al., 2017], and an salient object detection method U2NET [Qin et al., 2020] on AIM-500. For LF, U2NET, and GFM, we used the code provided by the author. For SHM, HAtt, and DIM, we re-implemented the code since they are not available. Since DIM required a trimap for training, we used the ground truth trimap as the auxiliary input and denoted it as DIM . We adopted the same transfer learning strategy while training all the models. We chose the commonly used sum of absolute differences (SAD), mean squared error (MSE), Mean Absolute Difference (MAD), Connectivity(Conn.), Gradient(Grad.) [Rhemann et al., 2009], as the main metrics, and also calculated the SAD within transition areas, SAD by type and category for comprehensive evaluation. The results of objective metrics are summarized in Table 2. Some visual results are presented in Figure 51. As shown in Table 2, our model achieves the best in all metrics among all AIM methods and outperforms DIM in most of the metrics. Although DIM performs better in some SO category, e.g., animal, human, furniture, it indeed uses ground truth trimaps as auxiliary inputs while our model does not have such a requirement. Nevertheless, our model still outperforms DIM in the transition areas as well as on the STM and NS types, implying that the semantic decoder predicts accurate semantic representation and the matting decoder has a better ability for extracting alpha details. Besides, DIM is very sensitive to the size of trimap and may produce a bad result when there is a large transition area, e.g. boundary of the bear in Figure 5. U2NET can estimate the rough foregrounds for SO images, but it fails to handle STM and NS images, e.g. the crystal stone and the net, implying that a single decoder is overwhelmed to learn both global semantic features and local detail features to deal with all types of images since they have different characteristics. SHM uses a two-stage network to predict trimap and the final alpha, which may accumulate the semantic error and mislead the subsequent matting process. Consequently, it obtains large SAD errors in the whole image as well as in the transition area, i.e., 170.44 and 69.41. Some failure results can be found in the bear and crystal stone in Figure 5. LF adopts a classification network to distinguish foreground and background, but it is difficult to adapt to STM and NS images as they do not have explicit foreground and background, thereby resulting in large average SAD errors. HATT tries to learn foreground profile to support boundary detail matting. However, there are no explicit supervisory signal for the semantic branch, which makes it difficult to learn explicit semantic representations. As a strong baseline model, GFM outperforms other AIM methods, but it is still worse than us, i.e., 52.66 to 43.92, especially for STM and NS images, as seen from the crystal stone and the net. The results demonstrate that the customized designs in the network matter a lot for dealing 1More results can be found in the supplementary material. Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) Whole Image Tran. SAD-Type SAD-Category SAD MSE MAD Conn. Grad. SAD SO STM NS Avg. Animal Human Transp. Plant Furni. Toy Fruit Avg. U2NET 83.46 0.0348 0.0493 82.14 51.02 43.37 69.69 120.59 211.98 134.09 67.67 89.50 210.34 75.72 87.20 54.64 52.24 91.04 SHM 170.44 0.0921 0.1012 170.67 115.29 69.41 154.56 204.67 329.9 229.71 174.65 141.49 333.24 157.24 166.81 126.04 97.31 170.97 LF 191.74 0.0667 0.1130 181.26 63.51 78.13 177.98 220.22 331.34 243.18 167.90 131.96 276.13 228.94 249.70 224.50 287.40 223.79 HATT 479.17 0.2700 0.2806 473.98 238.63 114.23 509.75 338.11 270.07 372.64 579.96 484.85 264.35 433.96 299.19 447.01 401.73 415.86 GFM 52.66 0.0213 0.0313 52.69 46.11 37.43 35.45 123.15 181.90 113.50 28.18 27.61 190.50 75.77 80.94 51.42 27.87 68.90 DIM 49.27 0.0147 0.0293 47.10 29.30 49.27 19.51 115.42 345.33 160.09 16.41 15.10 273.96 95.60 43.06 34.67 17.00 70.83 Ours 43.92 0.0161 0.0262 43.18 33.05 30.74 31.80 94.02 134.31 86.71 26.39 24.68 148.68 54.03 62.70 53.15 37.17 58.11 Table 2: Quantitative results on AIM-500. DIM denotes the DIM method using ground truth trimap as an extra input. Tran.: Transition Area, Transp.: Transparent, Furni.: Furniture. Figure 5: Some visual results of different methods on AIM-500. More results can be found in the supplementary material. UNI TL MP SE SA SAD MSE MAD 81.08 0.0363 0.0480 76.90 0.0296 0.0456 52.66 0.0213 0.0313 51.23 0.0205 0.0307 48.52 0.0195 0.0287 48.95 0.0207 0.0293 44.50 0.0158 0.0262 43.92 0.0161 0.0262 Table 3: Ablation study results. UNI: unified semantic representations; TL: transfer learning; MP: backbone with max pooling; SE: SE attention; SA: spatial attention. different types of images. Generally, our method achieves the best performance both objectively and subjectively. 4.4 Ablation Study We present the ablation study results in Table 3. There are several findings. Firstly, the proposed unified semantic representation is more effective than the traditional trimap, which helps to reduce the SAD from 81.08 to 76.90. Secondly, the transfer learning strategy is very effective by leveraging the large scale DUTS dataset for pre-training, reducing the SAD from 75.90 to 52.66. Thirdly, all the customized designs including max-pooling in the backbone, SE attention module in the semantic decoder, and spatial attention based on semantic features to guide the matting decoder, are useful and complementary to each other. For example, our method that uses all these designs reduces SAD from 52.66 to 43.92. It provides explicit and non-trivial supervisory signals for the semantic decoder, facilitating it to learn effective semantic features. It is noteworthy that although our model without MP has lower MSE, however, the visual results of the alpha matte obtained by our model with MP is better, e.g., with clearer details. Thereby, we choose it as our final model. 5 Conclusion In this paper, we investigate the difficulties in automatic image matting, including matting without auxiliary inputs, matting on natural images, and matting on all types of images. We propose a unified semantic representation for all types by introducing the new concepts of duomap and unimap, proved to be useful. We devise an automatic matting network with several customized new designs to improve its capability for AIM. Moreover, we establish the first natural image test set AIM-500 to benchmark AIM models, which can serve as a test bed to facilitate future research. Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) References [Cai et al., 2019] Shaofan Cai, Xiaoshuai Zhang, Haoqiang Fan, Haibin Huang, Jiangyu Liu, Jiaming Liu, Jiaying Liu, Jue Wang, and Jian Sun. Disentangled image matting. In ICCV, 2019. [Chen et al., 2013] Qifeng Chen, Dingzeyu Li, and Chi Keung Tang. Knn matting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(9):2175 2188, 2013. [Chen et al., 2018] Quan Chen, Tiezheng Ge, Yanyu Xu, Zhiqiang Zhang, Xinxin Yang, and Kun Gai. Semantic human matting. In Proceedings of the ACM International Conference on Multimedia, pages 618 626, 2018. [Cho et al., 2016] Donghyeon Cho, Yu-Wing Tai, and Inso Kweon. Natural image matting using deep convolutional neural networks. In ECCV, 2016. [He et al., 2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016. [Hou and Liu, 2019] Qiqi Hou and Feng Liu. Context-aware image matting for simultaneous foreground and alpha estimation. In ICCV, 2019. [Hu et al., 2018] Jie Hu, Li Shen, and Gang Sun. Squeezeand-excitation networks. In CVPR, 2018. [Ke et al., 2020] Zhanghan Ke, Kaican Li, Yurou Zhou, Qiuhua Wu, Xiangyu Mao, Qiong Yan, and Rynson W.H. Lau. Is a green screen really necessary for real-time portrait matting? Ar Xiv, abs/2011.11961, 2020. [Levin et al., 2007] Anat Levin, Dani Lischinski, and Yair Weiss. A closed-form solution to natural image matting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 30(2):228 242, 2007. [Levin et al., 2008] Anat Levin, Alex Rav-Acha, and Dani Lischinski. Spectral matting. IEEE transactions on pattern analysis and machine intelligence, 30(10):1699 1712, 2008. [Li and Lu, 2020] Yaoyi Li and Hongtao Lu. Natural image matting via guided contextual attention. In AAAI, 2020. [Li et al., 2020] Jizhizi Li, Jing Zhang, Stephen J Maybank, and Dacheng Tao. End-to-end animal image matting. ar Xiv preprint ar Xiv:2010.16188, 2020. [Li et al., 2021] Jizhizi Li, Sihan Ma, Jing Zhang, and Dacheng Tao. Privacy-preserving portrait matting. ar Xiv preprint ar Xiv:2104.14222, 2021. [Lin et al., 2014] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll ar, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, 2014. [Lu et al., 2019] Hao Lu, Yutong Dai, Chunhua Shen, and Songcen Xu. Indices matter: Learning to index for deep image matting. In ICCV, 2019. [Ma et al., 2020] Benteng Ma, Jing Zhang, Yong Xia, and Dacheng Tao. Auto learning attention. Advances in Neural Information Processing Systems, 33, 2020. [Qiao et al., 2020] Yu Qiao, Yuhao Liu, Xin Yang, Dongsheng Zhou, Mingliang Xu, Qiang Zhang, and Xiaopeng Wei. Attention-guided hierarchical structure aggregation for image matting. In CVPR, 2020. [Qin et al., 2020] Xuebin Qin, Zichen Zhang, Chenyang Huang, Masood Dehghan, Osmar R Zaiane, and Martin Jagersand. U2-net: Going deeper with nested ustructure for salient object detection. Pattern Recognition, 106:107404, 2020. [Rhemann et al., 2009] Christoph Rhemann, Carsten Rother, Jue Wang, Margrit Gelautz, Pushmeet Kohli, and Pamela Rott. A perceptually motivated online benchmark for image matting. In CVPR, 2009. [Ronneberger et al., 2015] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computerassisted intervention, pages 234 241. Springer, 2015. [Sengupta et al., 2020] Soumyadip Sengupta, Vivek Jayaram, Brian Curless, Steven M Seitz, and Ira Kemelmacher-Shlizerman. Background matting: The world is your green screen. In CVPR, 2020. [Shen et al., 2016] Xiaoyong Shen, Xin Tao, Hongyun Gao, Chao Zhou, and Jiaya Jia. Deep automatic portrait matting. In ECCV, 2016. [Sun et al., 2004] Jian Sun, Jiaya Jia, Chi-Keung Tang, and Heung-Yeung Shum. Poisson matting. ACM Transactions on Graphics, 23(3):315 321, 2004. [Sun et al., 2021] Yanan Sun, Chi-Keung Tang, and Yu Wing Tai. Semantic image matting. In CVPR, 2021. [Tang et al., 2019] Jingwei Tang, Yagiz Aksoy, Cengiz Oztireli, Markus Gross, and Tunc Ozan Aydin. Learningbased sampling for natural image matting. In CVPR, 2019. [Wang and Cohen, 2007] Jue Wang and Michael F Cohen. Optimized color sampling for robust matting. In CVPR, 2007. [Wang et al., 2017] Lijun Wang, Huchuan Lu, Yifan Wang, Mengyang Feng, Dong Wang, Baocai Yin, and Xiang Ruan. Learning to detect salient objects with image-level supervision. In CVPR, 2017. [Xu et al., 2017] Ning Xu, Brian Price, Scott Cohen, and Thomas Huang. Deep image matting. In CVPR, 2017. [Zhang and Tao, 2020] Jing Zhang and Dacheng Tao. Empowering things with intelligence: A survey of the progress, challenges, and opportunities in artificial intelligence of things. IEEE Internet of Things Journal, 2020. [Zhang et al., 2019] Yunke Zhang, Lixue Gong, Lubin Fan, Peiran Ren, Qixing Huang, Hujun Bao, and Weiwei Xu. A late fusion cnn for digital matting. In CVPR, 2019. [Zhao et al., 2017] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. Pyramid scene parsing network. In CVPR, 2017. Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21)