# onesample_guided_object_representation_disassembling__84b66de1.pdf One-sample Guided Object Representation Disassembling Zunlei Feng Zhejiang University zunleifeng@zju.edu.cn Yongming He Zhejiang University yongminghe@zju.edu.cn Xinchao Wang Stevens Institute of Technology xinchao.wang@stevens.edu Xin Gao Alibaba Group zimu.gx@alibaba-inc.com Jie Lei Zhejiang University of Technology jasonlei@zjut.edu.cn Cheng Jin Fudan University jc@fudan.edu.cn Mingli Song Zhejiang University brooksong@zju.edu.cn The ability to disassemble the features of objects and background is crucial for many machine learning tasks, including image classification, image editing, visual concepts learning, and so on. However, existing (semi-)supervised methods all need a large amount of annotated samples, while unsupervised methods can t handle realworld images with complicated backgrounds. In this paper, we introduce the Onesample Guided Object Representation Disassembling (One-GORD) method, which only requires one annotated sample for each object category to learn disassembled object representation from unannotated images. For the annotated one-sample, we first adopt some data augmentation strategies to generate some synthetic samples, which can guide the disassembling of the object features and background features. For the unannotated images, two self-supervised mechanisms: dual-swapping and fuzzy classification are introduced to disassemble object features from the background with the guidance of annotated one-sample. What s more, we devise two metrics to evaluate the disassembling performance from the perspective of representation and image, respectively. Experiments demonstrate that the One GORD achieves competitive dissembling performance and can handle natural scenes with complicated backgrounds. 1 Introduction Learning disassembled object representation is a vital step in many machine learning tasks, including image editing, image classification, few/zero-shot learning, and visual concepts learning. For example, many image editing [4, 21, 26] for objects typically rely on image segmentation techniques and human labor, which only handles object in image level. The existing classification works [15, 18] usually train classifiers with large amounts of annotated samples to extract specific object features and identify them, which also has a serious cost of labor, time, and memory. For the few/zero-shot learning problem, most of works [2, 22, 27, 30] adopt representations extracted by pre-trained deep models as the features of specific objects. However, the representations extracted by pre-trained models usually contain many irrelevant features, which will disturb the performance of models. So, an object representation learning method that can learn the pure and entire features of the specific object with a few annotated data is desperately needed. Corresponding author. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Until now, most object representation learning methods [6, 8, 12, 13, 23] are proposed to handle simple scenes with multiple objects in an unsupervised manner. However, those methods can t handle real-world images with complicated backgrounds, which limits their application in many machine learning tasks. On the other hand, the existing supervised object representation learning methods are rarer. Some (semi-)supervised disentangling methods [11, 25] can be transferred to learn disassembled object representation through annotating the object information as labels. However, it still requires many annotated samples. Another line of works [14, 28] is concerned with obtaining the segmentation of objects and does not learn structured object representations. In this paper, we propose the One-sample Guided Object Representation Disassembling (One-GORD) method, which only requires one annotated sample for each object category to learn disassembled object representation from a large number of unannotated images. The proposed One-GORD is composed of two modules: the augmented one-sample supervision module and the guided selfsupervision module. In the one-sample guided module, we first generate some synthetic sample pairs with data augmentation strategies. Then, following the encoding-swapping-decoding" architecture, we swap the parts of their representations to reconstruct the synthetic ground-truth pairs, which guides the features of objects and backgrounds to be decoded into different parts of the representations. In the guided self-supervision module, we introduce two self-supervised mechanisms: fuzzy classification and dual swapping. For the dual swapping, given a pair of samples that are composed of the annotated one-sample and an unannotated image, we swap the first halves of their representations to reconstruct the hybrid images. Then, the first halves of the hybrids representations are swap back to reconstruct the original pair samples, which formats the self-supervision loss for the disassembling of unannotated images with the guidance of annotated one-sample. Meanwhile, the fuzzy classification supervises the first and latter halves of representation to extract features of any object category and background, respectively. Furthermore, to verify the effectiveness of the proposed method, we devise two metrics to evaluate the modularity of representations and the integrity of images. The former measures the modularity and portability of the latent representations, while the latter evaluates the visual completeness of the reconstructed images. As will be demonstrated in our experiments, the proposed One-GORD achieves truly promising performance. Our contribution is the proposed One-GORD, which only requires one annotated sample for learning disassembled object representation. Two self-supervised mechanisms format self-supervised losses for the disassembling of unannotated image representations with the guidance of annotated one-sample. Meanwhile, We also introduce two disassembling metrics, upon which the proposed One-GORD achieves truly encouraging results. 2 Related Work Representation learning [5] has achieved several breakthroughs. This includes representation disentangling [7, 16, 17], which disentangle attribute features into different parts of representation. Part of (semi-)supervised disentangling methods [25, 11] can be transferred to learn disassembled object representation through annotating the object information as labels. However, it still requires a lot of annotated samples. Another line of works is concerned with obtaining the segmentation of objects without considering representation learning. Most current approaches [28, 14] require explicitly annotated segmentations in the dataset, which limits the generalization of these models. Furthermore, these methods typically only segment images and don t learn structured object representations. Works for learning disassembled object representation are relatively rarer. Burgess et al. [6] proposed the MONet, where a VAE is trained end-to-end together with a recurrent attention network to provide attention masks around regions of images. The MONet can decompose 3D scenes into objects and background elements. Greff et al. [12] developed an amortized iterative refinement based method, which can segment scenes and learn disentangled object features. Van Steenkiste et al. [23] proposed the R-NEM that learns to discover objects and model their physical interactions from raw visual images, which is the extension of N-EM [13]. Dittadi and Winther [8] proposed the probabilistic generative model for learning of structured, compositional, object-based representations of visual scenes. Engelcke et al. [10] proposed the GENESIS for rendered 3D scenes, which decomposes and generates scenes by capturing relationships between scene components. Lin et al. [20] proposed the SPACE, which factorizes object representations of foreground objects while also decomposing Supervision Module Self-supervision Module Ia I a fφ ro r o rb r b r o ro rb r b Ia I a b Ia b I a Iu Ia fφ ru o ro ru b rb ro ru o Iu Ia b Ia u b Iu a fφ Figure 1: The architecture of the proposed One-GORD. It comprises two modules: the augmented one-sample supervision module and the guided self-supervision module. The former one is employed for the augmented annotated one-samples (Ia, I a) while the latter is for unannotated image Iu and the annotated image Ia. fφ, fψ and fϕ denote the encoder, decoder, and classifier, respectively. l0, lk and ... lk denote the background label, the k-th object category label and the k-th unknown object category label. Ia, I a and Iu denotes the reconstructed images. b Ia, b I a, b Ia u and b Iu a denote the hybrid images. Iu and Ia denote the dual reconstructed images. background segments of complex morphology. All the above methods learn disassembled object representation in an unsupervised manner. However, those methods work on synthetic images only but not real-world ones. 3 The Proposed Method In this section, we give more details of our proposed One-GORD model (Fig. 1). We start by introducing the augmented one-sample supervision module, which disassembles the object features from the background with the supervision information of augmented annotated one-samples. Then, we describe the self-supervision module, which disassembles the object representation with two self-supervised mechanisms under the guidance of the annotated one-sample. Finally, we summarize the complete algorithm. 3.1 Augmented One-sample Supervision Module Annotated One-sample The natural scene images are usually composed of complicated things, which results in that unannotated object representation leaning methods can t work well. To distinguish the foreground objects from the background, there must be one annotated sample for each object category that is intended to be disassembled. In this paper, we choose one image for each object category and annotate the category label lk, k {1, 2, 3, ..., K} and mask of the object, where K is the number of object categories. To enhance the influence of one-samples supervision, some data augmentation strategies are adopted to generated augmented images. The augmentation strategies include mirroring, adding noise, and changing background, which are usually optional for different datasets. For the augmented images, we randomly choose two samples Ia and I a, then get the ground-truth images e Ia and e I a by swapping their objects with the masks. Supervision Disassembling The supervision information contains two parts: classification supervision and reconstruction supervision. With encoder fφ and decoder fψ, the input images Ia and I a are encoded into the representations Ra and R a, which are decoded into the images Ia and I a. The latent representations are constrained to contain all the features of input images with the following basic reconstruction loss Lrec: Lrec = ||Ia Ia||2 2 + ||I a I a||2 2. Then, the representations Ra and R a are divided into two parts: [ro, rb] and [r o, r b], respectively. For the classification supervision, the object label lk, k {1, 2, 3, ..., K} supervises ro and r o to extract object features while the background label l0 supervises rb and r b extracting background features with the classification loss Lcla: vec X lk log(p), where lk is a one-hot label vector, p is the predicted probability of classifier fϕ with one part of the representations {ro, rb, r o, r b} as the input, Pvec denotes the summation of n-dimension vector. For the reconstruction supervision, through swapping the object parts ro and r o, the hybrid representations [r o, rb] and [ro, r b] are decoded into the hybrid images b Ia and b I a, respectively. The ground-truth images e Ia and e I a supervise the first half and latter half to extract features of object and background with the following reconstruction supervision loss L rec: L rec = ||e Ia b Ia||2 2 + ||e I a b I a||2 2. 3.2 Guided Self-supervision Module For large amounts of unannotated images, we introduce two self-supervised mechanisms: dual swapping and fuzzy classification. Dual swapping swaps parts of the unannotated image representation back and forth to reconstruct the original image, which generates the self-supervised information. The fuzzy classification supervises the features of unknown objects and background to be encoded into different parts of the representations with the fuzzy classification loss. Dual Swapping For the unannotated image Iu, the same autoencoder reconstructs it as image Iu with the following unsupervised reconstruction loss Lu rec: Lu rec = ||Iu Iu||2 2. Similarly, the encoded representation Ru = fφ(Iu) is divided into two parts [ru o , ru b ]. To bring the guidance of annotated one-samples, we swap the representations first parts of annotated one-samples and unannotated images. Then, the hybrid representations [ru o , rb] and [ro, ru b ] are decoded into the hybrid images b Ia u and b Iu a. Following the encoding-swapping-decoding" process again, the hybrid images b Ia u and b Iu a are reconstructed into Iu and Ia by swapping the representations first parts back. If the representation is well disassembled, the dual reconstructed image Iu should reconstruct the original image Iu. So, the dual swapping loss Ld rec is defined as follows: Ld rec = ||Iu Iu||2 2. Meanwhile, to ensure that the object features are encoded into the first part of one-sample, the hybrid image b Ia u should have the same object with the annotated one-sample Ia. So, the object reconstruction loss Lo rec is defined as follows: Lo rec = Ma ||b Ia u Ia||2 2, where Ma is the object mask of the one-sample Ia. The interaction between the unannotated images and annotated images will enhance the guidance of the annotated one-sample. Fuzzy Classification For the unannotated images, the object labels are unknown. It s hard to supervise the fixed part to extract specific object features. Nevertheless, the object features of the unannotated images are still discriminative and the features of the background should be different from them. If the first half of the representation containing pure object features, it will be classified into the particular object category easily. So we devise the fuzzy classification loss, which can constrain that the features of unknown objects be classified into their original categories. What s more, the fixed label l0 is also adopted to differentiate the background features from object features. So, fuzzy classification loss Lz rec is defined as follows: Lz cla = log{[1 k=1 (lk lk q)]} τ vec X l0 log(q0), where lk is one-hot object label, q = fϕ(ru o ) is the predicted probability of classifier fϕ with the first half of the representation ru o as input, q0 = fϕ(ru b ) is the predicted probability of classifier fϕ with the latter half of the representation ru b as input, Pvec denotes summation of multi-dimension vector, and τ is the balance parameter. 3.3 Complete Algorithm In summary, the total loss L contains all the loss terms in the above two modules. In the supervision module, the loss terms disassemble the object representation with the classification supervision and reconstruction supervision; In the self-supervision module, the loss terms disassemble the object representation with two self-supervision mechanisms under the guidance of the annotated one-samples. The total loss L is given as follows: L = αLrec + βLcla + γL rec + ηLu rec + λLd rec + ρLo rec + δLz cla, where α, β, γ, η, λ, ρ, and δ are the balance parameters. It is noticeable that all the encoders, decoders, and classifiers share the same parameters, respectively. 4 Disassembling Metric It s essential to measure the disassembling performance of different methods. To the very best of our knowledge, there is no quantitative metric for evaluating the disassembling performance directly. To measure the disassembling performance fairly, we begin by defining the properties that we expect the disassembled object representation to have. It should consider the latent representation and the reconstructed image. If the object representation is disassembled perfectly, the extracted object representation should be equally for different images with the same object. On the other hand, for the images reconstructed with the same object representation, objects should keep the same. Therefore, we devise two disassembling metrics to measure the modularity on the latent representation and the integrity of the reconstructed image, respectively. For modularity, we run inference on images that contain the fixing object and different backgrounds. If the modularity property holds for the inferred representations, there will be less variance in the inferred latent representations that correspond to the fixed object. For the T D test images that are composed of T kinds of object and D image for each object category, the Modularity Score M(T, D) is calculated as follows: M(T, D) = 1 T D vec X |zt d 1 where Pvec denotes summation of n-dimension vector, zt d denotes the object part of the representation extracted from the d-th image It d of the t-th object category. For integrity, we reconstruct the image e It d through swapping the background part of the test image It d with other background parts. Giving the test images {It d, t {1, 2, 3, ..., T}, d {1, 2, 3, ..., D}}, the Integrity Score V (T, D), which measures the object integrity of reconstructed images, is defined as follows: V (T, D) = 1 T D W W X Mt d|e It d It d|, where W is the pixel number of the image, PW 1 denotes summation of image pixel difference value. Mt d is the object mask of the test image It d. 5 Experiments In the experiment, One-GORD is compared with the unsupervised method, the semi-supervised method, and the supervised method. Those methods are validated on five datasets qualitatively and quantitatively. What s more, experiments demonstrate the comparative performance in the practical application: classification and image editing. 5.1 Implementation Details Dataset To verify the effectiveness of the proposed One-GORD, we adopt five datasets: SVHN [29], CIFAR-10 [3], COCO [19], Mugshot [11], and mini-Image Net [24], which are composed of different objects and complex backgrounds. For the COCO dataset, we choose ten object categories (bird, bottle, cat, dog, laptop, truck, tv, tie, sink and book). For the rest of datasets, all the categories are adopted in the experiment. The training and testing sample numbers are (20000,1000), (20000,1000), (40000,1000), (30000,1000), (10000,1000) for SVHN, CIFAR-10, COCO, Mugshot, and mini Image Net, respectively. Network architectures The encoders and decoders have the same architecture as Res Net2 [9] . The classifier network is the two-layer MLP with 20 and N neurons for each layer, where N is determined by the category number for each dataset. The Adam algorithm is adopted. The learning rate is set to 0.0005. Parameters settings In the experiment, the balance parameters τ, α, γ, η, λ are set to 1, and β is set to 10, ρ is set to 1000, and δ is set to 5. Through large experiments, we find that the crucial parameter are β, ρ and δ. Tuning β, ρ and δ may lead to better performance under the condition that τ, α, γ, η, λ are set to 1. 5.2 Qualitative Evaluation In the qualitative evaluation experiments, our methods are compared with AE, S-AE, DSD [11], MONet [6] and IODINE [12], which are shown in Fig. 2. AE is the basic autoencoder architecture. S-AE is the AE with a classifier, which supervises different parts of latent representation to extract object-related features with object labels. For DSD, the annotated input pairs are generated with the augmented annotated one-samples. From Fig. 2, we can see that the object swapped images reconstructed by the AE have overlapping features of the two input images. It indicates that the latent representations extracted by AE have mix features of objects and background. In the object swapped images reconstructed by the DSD, the objects have object features of two input images, which demonstrates that the DSD fails in disassembling object features from the background with the same annotated one-samples. For the MONet [6] and IODINE [12], the splitted objects and backgrounds of SVHN dataset are wrong. What s more, they fail on the CIFA-10, COCO and Mugshot, which verifies that the existing unsupervised methods fail in handling real-world images with complicated background. It s noticeable that the corresponding objects are swapped successfully in the results of One-GORD for the above five datasets, which verifies that One-GORD can handle real-world images with complicated background effectively. Even the reconstructed images lost some details, the swapped object and background only contain their corresponding features. In the second row of Fig. 2, the images reconstructed by autoencoder also not achieve perfect reconstructed results. The image reconstruction quality is usually decided by the dataset and network architecture, which will be optimized in our future works. 5.3 Quantitative Evaluation To compare different methods quantitatively, we adopt the Modularity Score and Integrity Score (Section 4) to measure the disassembling performance of our methods with S-AE, DSD [11], MONet [6] and IODINE [12]. In the experiments, T and D are set to 10 and 100, respectively. We sample 5 kinds of representation length ({10, 20, 30, 40, 50}) and test all methods in those length setting. Table 1 gives the average modularity score (AMS) and average integrity score (AIS) on the SVHN dataset (the first three rows) and the CIFA-10 dataset (the last two rows). For the modularity score, One-GORD has the smallest score than other methods, which shows that the disassembled object representations extracted by One-GORD are more similar than other methods for the images with the same object and different backgrounds. O-G s ,O-G f and O-G o achieves the larger score than One-GORD on two datasets, which indicates the necessity of One-GORD s each component. What s more, S-AE, DSD, MONet and IODINE achieve larger scores than One GORD. It means that existing methods can t disassemble object representation effectively, which is in accordance with the reconstructed visual results in Section 5.2. For the integrity score, MONet and IODINE achieves the larger score than other methods, which verifies that MONet and IODINE fails in disassembling object representation for the real-world image with a complicated background. The average integrity score of S-AE is higher than the score of One-GORD, which shows that the supervision of the object label is not enough for disassembling 2https://github.com/cianeastwood/qedr CIFA-10 Mugshot COCO ob. bg. mask m-ob. mask m-ob. m-bg. comb. mini-Image Net Input Rec. AE S-AE DSD[11] One GORD O-G s O-G f O-G o MONet [6] IODINE [12] Figure 2: The qualitative results of different methods on five datasets. For each dataset, given two input images, we show the images reconstructed with the object parts swapped representations. For MONet [6] and IODINE [12], we show the splitted objects and backgrounds. Rec. denotes reconstructed results of original images by AE. O-G s ,O-G f and O-G o denote the One-GORD without supervision module, fuzzy classification, and object reconstruction loss. ob. and bg. denote object and background. m-ob. and m-bg. denote masked object and masked background. comb. denotes the combined result of m-ob. and m-bg. . Table 1: The Average Modularity Score (AMS) and Average Integrity Score (AIS) in five representation length setting on SVHN (the first three rows) and CIFA-10 (the last two rows) datasets. Metric S-AE DSD [11] MONet [6] IODINE [12] One-GORD O-G s O-G f O-G o AMS 13.69 12.38 11.52 15.78 5.91 11.58 17.01 13.30 AIS 6.51 3.02 10.31 14.31 2.05 4.05 5.67 2.45 AMS 15.82 14.97 16.83 19.04 8.43 12.43 18.21 15.31 AIS 8.21 6.94 11.96 15.34 5.21 7.92 9.37 7.94 object features from background integrally. One-GORD achieves the smallest score among all methods, which indicates that the reconstructed objects are more intact than other methods . 5.4 Ablation Study In the One-GORD, the total loss L is composed of loss terms from two modules: augmented onesample supervision module and guided self-supervision module. To verify the necessity of the augmented one-samples, we remove the loss terms in the supervision module, which is denoted as O-G s . Meanwhile, we also do the ablation study by removing the fuzzy classification loss Lz rec and object reconstruction loss Lo rec from the guided self-supervision module, which are denoted as O-G o and O-G f , respectively. Table 1 gives the average modularity and average integrity scores of One-GORD, O-G s , O-G f and O-G o in five representation lengths ({10, 20, 30, 40, 50}). It s noticeable that O-G s achieves the largest average modularity and integrity scores than O-G f , O-G o , One-GORD, which demonstrates the necessity of the annotated one-sample. O-G o achieves a relatively high modularity score and a relatively small visual integrity score. The reason is that the object reconstruction loss can effectively Table 2: The classification performance on SVHN and CIFA-10 (All scores in %). Dataset Metric S-AE DSD [11] MONet [6] IODINE [12] One-GORD O-G s O-G f O-G o C-P 56.91 45.66 54.56 43.86 60.94 58.27 57.76 57.46 C-R 57.87 46.60 54.86 42.78 59.18 58.11 58.04 57.40 O-P 57.27 45.20 54.67 43.89 61.47 57.60 57.93 57.17 O-R 57.37 45.31 54.97 44.81 60.47 57.69 57.33 57.26 C-P 44.93 41.27 43.59 39.43 46.23 39.61 43.17 44.21 C-R 45.81 41.26 42.71 38.81 47.38 38.83 43.82 43.96 O-P 45.93 41.27 43.83 37.49 46.31 40.82 41.62 42.81 O-R 45.93 41.26 41.86 38.26 46.73 42.96 44.21 41.97 promote the reconstruction quality of the object. However, it affects the modularity of latent representation negatively, which leads to a higher modularity score. Without fuzzy classification, O-G f achieves a higher score than One-GORD, which verifies the effectiveness of the fuzzy classification loss. Table 2 shows the classification performance of One-GORD, O-G s , O-G f and O-G o on SVHN and CIFA-10, respectively. We can see that One-GORD achieves the best classification performance than other methods, which demonstrates that fuzzy classification, object reconstruction loss, and supervision module can enhance the disassembling performance effectively. 5.5 Application As described above, our method can be applied to many machine learning tasks, including image classification, image editing, visual concepts learning, and so on. In this section, we test the performance on two basic applications: image editing and image classification. For image editing, given one image, objects in the other seven images are swapped into it. The object swapped results are shown in Fig. 3, where we can see that the corresponding objects are successfully swapped for the five datasets. However, there are still some details lost in the reconstructed images, which will be studied in our future work. The benefit of image editing in the latent representation space is that the obscured backgrounds by the objects can be reconstructed well. For image classification, we compare our methods with other methods on SVHN and CIFA-10 with 1000 test samples. The per-class and overall precision (C-P and O-P) and recall scores (C-R and O-R) are calculated for the above methods, where the average score is taken over all classes and all test samples, respectively. To compare fairly, after obtaining the disassembled representation for each method, we adopt the same linear SVM [1] to train and test the classification performance. The classification performance on SVHN and CIFA-10 are shown in Table 2, respectively . We can see that our method achieves a higher score than other methods, which demonstrates that the object features extracted by our method are more intact and independent. Meanwhile, the O-G f and O-G o achieve the lower scores than other methods, which verifies the effectiveness of the fuzzy classification and object reconstruction loss once again. It s noticeable that even with the supervised label, the S-AE still achieves lower accuracy and recall scores than One-GORD, which indicates that disassembled object representation can effectively improve the classification performance. 6 Conclusion In this paper, we propose the One-GORD, which only requires one annotated sample for each object category to learn disassembled object representation from unannotated images. One-GORD is composed of two modules: the augmented one-sample supervision module and the guided self-supervision module. In the supervision module, we generate some augmented one-samples with data augmentation strategies. Then, the annotated mask and object label supervise the disassembling between the features of the object and background. In the self-supervision module, two self-supervised mechanisms (fuzzy classification and dual swapping) are adopted to generate self-supervised information, which can disassemble object representation of unannotated images with the guidance with annotated one-samples. What s more, we devise two disassembling metrics to measure the modularity of SVHN CIFA-10 Mugshot COCO Input images (candidate swapped objects) Background Composited images (object+background) Figure 3: The image editing results on different datasets. representations and the integrity of images, respectively. A large number of experiments demonstrate that the proposed One-GORD achieve competitive dissembling performance and can handle natural scenes with complicated backgrounds. In future work, we will focus on disassembling objects into different parts and optimizing network architecture to solve the details lost. Acknowledgement This work is supported by National Natural Science Foundation of China (61976186, 62002318), Zhejiang Provincial Science and Technology Project for Public Welfare (LGF21F020020), Programs Supported by Ningbo Natural Science Foundation (202003N4318), the Major Scientific Research Project of Zhejinag Lab (No. 2019KD0AC01) and Alibaba-Zhejiang University Joint Research Institute of Frontier Technologies. Broader Impact This research belongs to the image representation learning area. Positive: the proposed method can be applied to many machine learning tasks, including image editing, image classification, few/zeroshot learning, and visual concepts learning. It supplies a universal tool for other downstream tasks. Negative: the research can be adopted to generate some fake image, which also can be used for malicious purposes. [1] Sklearn.svm. http://scikit-learn.sourceforge.net/stable/modules/generated/sklearn. svm.SVC.html. [2] Zeynep Akata, Scott E Reed, Daniel J Walter, Honglak Lee, and Bernt Schiele. Evaluation of output embeddings for fine-grained image classification. Computer Vision and Pattern Recognition, pages 2927 2936, 2015. [3] Krizhevsky Alex. Learning multiple layers of features from tiny images. In Handbook of Systemic Autoimmune Diseases, 2009. [4] William A Barrett and Alan S Cheney. Object-based image editing. In International Conference on Computer Graphics and Interactive Techniques, volume 21, pages 777 784, 2002. [5] Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new perspectives. In IEEE Transactions on Pattern Analysis and Machine Intelligence, volume 35, pages 1798 1828, 2013. [6] Christopher P Burgess, Loic Matthey, Nicholas Watters, Rishabh Kabra, Irina Higgins, Matthew Botvinick, and Alexander Lerchner. Monet: Unsupervised scene decomposition and representation. In ar Xiv: Computer Vision and Pattern Recognition, 2019. [7] Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. pages 2180 2188, 2016. [8] Andrea Dittadi and Ole Winther. Lavae: Disentangling location and appearance. In ar Xiv: Learning, 2019. [9] Cian Eastwood and Christopher K I Williams. A framework for the quantitative evaluation of disentangled representations. In International Conference on Learning Representations, 2018. [10] Martin Engelcke, Adam R Kosiorek, Oiwi Parker Jones, and Ingmar Posner. Genesis: Generative scene inference and sampling with object-centric latent representations. International Conference on Learning Representations, 2020. [11] Zunlei Feng, Xinchao Wang, Chenglong Ke, An-Xiang Zeng, Dacheng Tao, and Mingli Song. Dual swap disentangling. In Advances in Neural Information Processing Systems, pages 5894 5904, 2018. [12] Klaus Greff, Raphael Lopez Kaufman, Rishabh Kabra, Nicholas Watters, Christopher P Burgess, Daniel Zoran, Loic Matthey, Matthew Botvinick, and Alexander Lerchner. Multi-object representation learning with iterative variational inference. In International Conference on Machine Learning, 2019. [13] Klaus Greff, Sjoerd Van Steenkiste, and Jurgen Schmidhuber. Neural expectation maximization. In Neural Information Processing Systems, pages 6691 6701, 2017. [14] Kaiming He, Georgia Gkioxari, Piotr Dollar, and Ross Girshick. Mask r-cnn. In IEEE Transactions on Pattern Analysis and Machine Intelligence, pages 1 1, 2018. [15] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Computer Vision and Pattern Recognition, pages 770 778, 2016. [16] Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. Beta-vae: Learning basic visual concepts with a constrained variational framework. In International Conference on Learning Representations, 2017. [17] Hyunjik Kim and Andriy Mnih. Disentangling by factorising. In International Conference on Machine Learning, pages 2649 2658, 2018. [18] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Neural Information Processing Systems, pages 1097 1105, 2012. [19] Tsungyi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollar, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European Conference on Computer Vision, pages 740 755, 2014. [20] Zhixuan Lin, Yifu Wu, Skand Vishwanath Peri, Weihao Sun, Gautam Singh, Fei Deng, Jindong Jiang, and Sungjin Ahn. Space: Unsupervised object-oriented scene representation via spatial attention and decomposition. International Conference on Learning Representations, 2019. [21] Pedro O Pinheiro, Ronan Collobert, and Piotr Dollar. Learning to segment object candidates. In ar Xiv: Computer Vision and Pattern Recognition, 2015. [22] Jie Song, Chengchao Shen, Yezhou Yang, Yang Liu, and Mingli Song. Transductive unbiased embedding for zero-shot learning. Computer Vision and Pattern Recognition, pages 1024 1033, 2018. [23] Sjoerd Van Steenkiste, Michael Chang, Klaus Greff, and Jurgen Schmidhuber. Relational neural expectation maximization: Unsupervised discovery of objects and their interactions. In International Conference on Learning Representations, 2018. [24] Oriol Vinyals, Charles Blundell, Timothy Lillicrap, Koray Kavukcuoglu, and Daan Wierstra. Matching networks for one shot learning. The 30th International Conference on Neural Information Processing Systems, 2016. [25] Chaoyue Wang, Chaohui Wang, Chang Xu, and Dacheng Tao. Tag disentangled generative adversarial network for object image re-rendering. In Twenty-Sixth International Joint Conference on Artificial Intelligence, pages 2901 2907, 2017. [26] John Winn. Object-level image editing. 2015. [27] Yongqin Xian, Bernt Schiele, and Zeynep Akata. Zero-shot learning-the good, the bad and the ugly. In Computer Vision and Pattern Recognition, pages 3077 3086, 2017. [28] Liu Xiao, Mingli Song, Dacheng Tao, Jiajun Bu, and Chun Chen. Random shape prior forest for multi-class object segmentation. In IEEE Transactions on Image Processing, volume 24, page 3060, 2015. [29] Netzer Yuval, Wang Tao, Coates Adam, Bissacco Alessandro, Wu Bo, and Y. Ng Andrew. Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on Deep Learning and Unsupervised Feature Learning, 2011. [30] Li Zhang, Tao Xiang, and Shaogang Gong. Learning a deep embedding model for zero-shot learning. Computer Vision and Pattern Recognition, pages 3010 3019, 2017.