# schema_inference_for_interpretable_image_classification__f2c79a87.pdf Published as a conference paper at ICLR 2023 SCHEMA INFERENCE FOR INTERPRETABLE IMAGE CLASSIFICATION Haofei Zhang , Mengqi Xue , Xiaokang Liu, Kaixuan Chen & Jie Song Zhejiang University {haofeizhang,mqxue,yijiyeah,chenkx,sjie}@zju.edu.cn Mingli Song Shanghai Institute for Advanced Study, Zhejiang University brooksong@zju.edu.cn In this paper, we study a novel inference paradigm, termed as schema inference, that learns to deductively infer the explainable predictions by rebuilding the prior deep neural network (DNN) forwarding scheme, guided by the prevalent philosophical cognitive concept of schema. We strive to reformulate the conventional model inference pipeline into a graph matching policy that associates the extracted visual concepts of an image with the pre-computed scene impression, by analogy with human reasoning mechanism via impression matching. To this end, we devise an elaborated architecture, termed as Schema Net, as a dedicated instantiation of the proposed schema inference concept, that models both the visual semantics of input instances and the learned abstract imaginations of target categories as topological relational graphs. Meanwhile, to capture and leverage the compositional contributions of visual semantics in a global view, we also introduce a universal Feat2Graph scheme in Schema Net to establish the relational graphs that contain abundant interaction information. Both the theoretical analysis and the experimental results on several benchmarks demonstrate that the proposed schema inference achieves encouraging performance and meanwhile yields a clear picture of the deductive process leading to the predictions. Our code is available at https://github.com/zhfeing/Schema Net-Py Torch. 1 INTRODUCTION Now this representation of a general procedure of the imagination for providing a concept with its image is what I call the schema for this concept1. Immanuel Kant Deep neural networks (DNNs) have demonstrated the increasingly prevailing capabilities in visual representations as compared to conventional hand-crafted features. Take the visual recognition task as an example. The canonical deep learning (DL) scheme for image recognition is to yield an effective visual representation from a stack of non-linear layers along with a fully-connected (FC) classifier at the end (He et al., 2016; Dosovitskiy et al., 2021; Tolstikhin et al., 2021; Yang et al., 2022a), where specifically the inner-product similarities are computed with each category embedding as the prediction. Despite the great success of DL, existing deep networks are typically required to simultaneously perceive low-level patterns as well as high-level semantics to make predictions (Zeiler & Fergus, 2014; Krizhevsky et al., 2017). As such, both the procedure of computing visual representations and the learned category-specific embeddings are opaque to humans, leading to challenges in security-matter scenarios, such as autonomous driving and healthcare applications. Equal contribution. Corresponding author. 1In Critique of Pure Reason (A140/B180). Published as a conference paper at ICLR 2023 0.183 #3 0.08 0.118 0.211 0.096 0.232 0.224 0.200 Elephant: 18.60 (#198) 2.65 (#150) Ingredient Visualization Instance IR-Graph Category IR-Graph 3 46 107 121 133 150 198 Local structure of #198 Local structure of #150 #3 0.08 0.211 0.07 0.224 #150 Instance Category Figure 1: An example showing how an instance IR-Graph is matched to the class imagination. The vertices of IR-Graphs represent visual semantics, and the edges indicate the vertex interactions. The graph matcher captures the similarity between the local structures of joint vertices (e.g., vertex #198 and vertex #150) by aggregating information from their neighbors as class evidences. The final prediction is defined as the sum of all evidence. Unlike prior works that merely obtain the targets in the black-box manner, here we strive to devise an innovative and generalized DNN inference paradigm by reformulating the traditional one-shot forwarding scheme into an interpretable DNN reasoning framework, resembling what occurs in deductive human reasoning. Towards this end, inspired by the schema in Kant s philosophy that describes human cognition as the procedure of associating an image of abstract concepts with the specific sense impression, we propose to formulate DNN inference into an interactive matching procedure between the local visual semantics of an input instance and the abstract category imagination, which is termed as schema inference in this paper, leading to the accomplishment of interpretable deductive inference based on visual semantics interactions at the micro-level. To elaborate the achievement of the proposed concept schema inference, we take here image classification, the most basic task in computer vision, as an example to explain our technical details. At a high level, the devised schema inference scheme leverages a pre-trained DNN to extract feature ingredients which are, in fact, the semantics represented by a cluster of deep feature vectors from a specific local region in the image domain. Furthermore, the obtained feature ingredients are organized into an ingredient relation graph (IR-Graph) for the sake of modeling their interactions that are characterized by the similarity at the semantic-level as well as the adjacency relationship at the spatial-level. We then implement the category-specific imagination as an ingredient relation atlas (IR-Atlas) for all target categories induced from observed data samples. As a final step, the graph similarity between an instance-level IR-Graph and the category-level IR-Atlas is computed as the measurement for yielding the target predictions. As such, instead of relying on deep features, the desired outputs from schema inference contribute only from the relationship of visual words, as shown in Figure 1. More specifically, our dedicated schema-based architecture, termed as Schema Net, is based on vision Transformers (Vi Ts) (Dosovitskiy et al., 2021; Touvron et al., 2021), which are nowadays the most prevalent vision backbones. To effectively obtain the feature ingredients, we collect the intermediate features of the backbone from probe data samples clustered by k-means algorithm. IR-Graphs are established through a customized Feat2Graph module that transfers the discretized ingredients array to graph vertices, and meanwhile builds the connections, which indicates the ingredient interactions relying on the self-attention mechanism (Vaswani et al., 2017) and the spatial adjacency. Eventually, graph similarities are evaluated via a shallow graph convolutional network (GCN). Our work relates to several existing methods that mine semantic-rich visual words from DNN backbones for self-explanation (Brendel & Bethge, 2019; Chen et al., 2019; Nauta et al., 2021; Xue et al., 2022b; Yang et al., 2022b). Particularly, Bag Net (Brendel & Bethge, 2019) uses a DNN as visual Published as a conference paper at ICLR 2023 Instance IR-Graph IR-Atlas Prediction Back propagation 0.158 0.892 Figure 2: The overall pipeline of our proposed Schema Net. Firstly, intermediate feature X from the backbone is fed into the Feat2Graph module. The converted IR-Graph is then matched to a set of imaginations, i.e., category-level IR-Atlas induced from observed data, for making the prediction. word extractor that is analogous to traditional bag-of-visual-words (Bo VW) representation (Yang et al., 2007) with SIFT features (Lowe, 2004). Moreover, Chen et al. (2019) develop Proto PNet framework that evaluates the similarity scores between image parts and learned class-specific prototypes for inference. Despite their encouraging performance, all these existing methods suffer from the lack of considerations in the compositional contributions of visual semantics, which is, however, already proved critical for both human and DNN inference (Deng et al., 2022), and are not applicable to schema inference in consequence. We discuss more related literature in Appendix A. In sum, our contribution is an innovative schema inference paradigm that reformulates the existing black-box network forwarding into a deductive DNN reasoning procedure, allowing for a clear insight into how the class evidences are gathered and deductively lead to the target predictions. This is achieved by building the dedicated IR-Graph and IR-Atlas with the extracted feature ingredients and then performing graph matching between IR-Graph and IR-Atlas to derive the desired predictions. We also provide a theoretical analysis to explain the interpretability of our schema inference results. Experimental results on CIFAR-10/100, Caltech-101, and Image Net demonstrate that the proposed schema inference yields results superior to the state-of-the-art interpretable approaches. Further, we demonstrate that by transferring to unseen tasks without fine-tuning the matcher, the learned knowledge of each class is, in fact, stored in IR-Atlas rather than the matcher, thereby exhibiting the high interpretability of the proposed method. 2 SCHEMANET In this section, we introduce our proposed Schema Net in detail. The overall procedure is illustrated in Figure 2, including a Feat2Graph module that converts deep features to instance IR-Graphs, a learnable IR-Atlas, and a graph matcher for making predictions. The main idea is to model an input image as an instance-level graph in which nodes are local semantics captured by the DNN backbone and edges represent their interactions. Meanwhile, a category-level graph is maintained as the imagination for each class driven by training samples. Finally, by measuring the similarity between the instance and category graphs, we are able to interpret how predictions are made. 2.1 PRELIMINARY We first give a brief review of the Vi T architecture. The simplest Vi T implementation for image classification is proposed by Dosovitskiy et al. (2021), which treats an image as a sequence of independent 16 16 patches. The embeddings of all patches are then directly fed to a Transformerbased network (Vaswani et al., 2017), i.e., a stack of multiple encoder layers with the same structure: a multi-head self-attention mechanism (MHSA) followed by a multilayer perceptron (MLP) with residual connections. Additionally, they append a class token (CLS) to the input sequence for classification alone rather than the average pooled feature. Following this work, Touvron et al. (2021) propose Dei T appending another distillation token (DIST) to learn soft decision targets from a teacher Published as a conference paper at ICLR 2023 model. This paper will mainly focus on Dei T backbones due to the relatively simple yet efficient network architecture. Let X R(n+ζ) d denotes the input feature (a sequence of n visual tokens and ζ auxillary tokens, e.g., CLS and DIST, each of which is a d-dimensional embedding vector) to a MHSA module. The output can be simplified as MHSA(X) = PH h=1 Ψh XWh, where Ψh R(n+ζ) (n+ζ) denotes the self-attention matrix normalized by row-wise softmax of head h {1, . . . , H}, and Wh Rd d is a learnable projection matrix. As the MLP module processes each token individually, the visual token interactions take place only in the MHSA modules, facilitating the extraction of their relationships from the perspective of Vi Ts. For convenience, we define some notations related to the self-attention matrix. Let Ψ be the average attention matrix of all heads and Ψ be the symmetric attention matrix: Ψ = ( Ψ + Ψ )/2. We partition Ψ into four submatrices: where ΨA Rζ n is the attention to the auxiliary tokens, ΨV Rn n represents the relationships of visual tokens, and Ψ Rζ ζ. Finally, let ψCLS Rn represent the attention values to the CLS token extracted from ΨA. 2.2 FEAT2GRAPH As shown in Figure 2, given a pre-trained backbone and an input image, the intermediate feature X is converted to an IR-Graph G for inference by the Feat2Graph module which includes three steps: (1) discretizing X into a sequence of feature ingredients with specific semantics; (2) mapping the ingredients to weighted vertices of IR-Graph; (3) assigning a weighted edge to each vertex pair indicating their interaction. We start by defining IR-Graph and IR-Atlas mathematically. Definition 1 (IR-Graph). An IR-Graph is an undirected fully-connected graph G = (E, V ), in which vertiex set V is the set of feature ingredients for an instance or a specific category, and edge set E encodes the interactions of vertex pairs. In particular, to indicate the vertex importance (for instance, bird head should contribute to the prediction of bird more than sky in human cognition), a non-negative weight λ R+ is assigned to each vertex. Let Λ R|V | + = {λi}|V | i=1 be the collection of all vertex weights. Besides, the interaction between vertex i and j is quantified by a non-negative weight ei,j R+. With a slight abuse of notation, we define E R|V | |V | + as the weighted adjacency matrix in the following sections. The IR-Atlas is defined to represent the imagination of all categories: Definition 2 (IR-Atlas). An IR-Atlas ˆG of C classes is a set of category-level IR-Graphs, in which element ˆGc = ( ˆEc, ˆVc) for class c has learnable vertex weights ˆΛc and edge weights ˆEc. Discretization. The main purpose of feature discretization is to mine the most common patterns appearing in the dataset, each of which corresponds to a human understandable semantic, named visual word, such as car wheel or bird head . However, local regions with the same semantic may show differently because of scaling, rotation, or even distortion. In our approach, we utilize the DNN backbone to extract a relatively uniform feature instead of the traditional SIFT feature utilized in (Yang et al., 2007). To be specific, a visual vocabulary Ω= {ωi}M i=1 (ωi Rd) of size M is constructed by k-means clustering running on the collection X of visual tokens extracted from the probe dataset2. Further, a deep feature X (CLS and DIST are removed) is discretized to a sequence of feature ingredients by replacing each element x with the index of the closest visual word Ingredient(x) = arg min i {1,...,M} x ωi 2. (2) Strictly, the ingredient is referred to as the index of visual word ω. With the entire ingredient set M = {1, . . . , M}, the discretized sequence is denoted as X = ( x1, . . . , xn), where xi M is computed from Equation (2). The detailed settings of M are presented in Appendix E.1. 2For probe dataset with D instances, the collection X has n D visual tokens (ignoring CLS and DIST). Published as a conference paper at ICLR 2023 Feat2Vertex. After we have computed the discretized feature X, the vertices of this feature are the unique ingredients: V = Unique( X). The importance of each vertex is measured from two criterions: the contribution to the DNN s final prediction and the appearance statistically. Formally, for vertex v M, the importance is defined as λv = α1λCLS v + α2λbag v = α1 X i Ξ(v| X) ψCLS i + α2|Ξ(v| X)|, (3) where Ξ(v| X) is the set of all the appeared position of ingredient v in X, ψCLS i is the attention between the i-th visual token and the CLS token, and α1,2 0 are learnable weights balancing the two terms. When α1 = 0, Λ is equivalent to Bo VW representation. Feat2Edge. With the self-attention matrix ΨV extracted from Equation (1), the interactions between vertices can be defined and computed with efficiency. For any two different vertices u, v V , the edge weight is the comprehensive consideration of the similarity in the view of Vi T and spatial adjacency. The first term is the average attention between all repeated pairs Π[(u, v)| X]: eattn u,v = 1 Π[(u, v)| X] (i,j) Π[(u,v)| X] ΨV i,j, (4) where Π[(u, v)| X] is the Cartesian product of Ξ(u| X) and Ξ(v| X), and ΨC i,j is the attention between the visual tokens at i and j positions. Furthermore, we define the adjacency as eadj u,v = 1 Π[(u, v)| X] (i,j) Π[(u,v)| X] 1 ϵ + Pos(i) Pos(j) 2 , (5) where function Pos( ) returns the original 2D coordinates of the input visual token with regard to the patch array after the patch embedding module of Vi Ts. Eventually, the interaction between vertices u and v is the weighted sum of the two components with learnable β1,2: eu,v = β1eattn u,v + β2eadj u,v. (6) Equations (4) and (5) are both invariant when exchanging vertex u and v, so our IR-Graph is equivalent to undirected graph. It is worth noting that only semantics and their relationships are preserved in IR-Graphs for further inference rather than deep features. 2.3 MATCHER Figure 3: Illustration of the matcher. After converting an input image to IR-Graph, the matcher finds the most similar category-level graph in IR-Atlas. The overall procedure is shown in Figure 3, which is composed of a GCN module and a similarity computing module (Sim) that generates the final prediction. Detailed settings of the matcher are in Appendix E.3. For feeding IR-Graphs to the GCN module, we assign each ingredient m M (i.e., graph vertices) with a trainable embedding vector, each of which is initialized from an d G-dimensional random vector drawn independently from multivariate Gaussian distribution N(0, Id G). In the GCN module, we adopt Graph Conv (Morris et al., 2019) with slight modifications for weighted edges. Let F R|V | d G be the input feature of all vertices to a Graph Conv layer, the output is computed as Graph Conv(F) = Norm (σ ((Id G + E)FW)) , (7) where σ denotes a non-linear activation function, Norm denotes feature normalization, and W Rd G d G is a Published as a conference paper at ICLR 2023 learnable projection matrix. After passing through total LG layers of Graph Conv, a weighted average pooling (WAP) layer summarizes the vertex embeddings F (LG) weighted by Λ, yielding the graph representation z = ΛF (LG). Now we have z for G and ˆZ RC d G = (ˆz1, . . . , ˆz C) for all graphs in ˆG. By computing the inner-product similarities, the final prediction logits is defined as y = ˆZz . Prediction interpretation. Given an instance graph G and a category graph ˆG, we now analyze how the prediction is made based on their vertices and edges. We start with analyzing the similarity score s = ˆzz , where z = ΛF (LG) and ˆz = ˆΛ ˆF (LG) are weighted sum of the vertex embeddings respectively. For the sake of further discussion, let f (l) v denotes the embedding vector of vertex v output from the l-th Graph Conv layer, and λ(l) v denotes its weight. The original vertex embedding of v is defined as f (0) v , which is identical for the same ingredient in any two graphs. Besides, let Φ = V ˆV be the set of shared vertices in G and ˆG. The computation of s can be expanded as ˆλuλv ˆf (LG) u f (LG) v . (8) Since directly analyzing Equation (8) is hard, we show an approximation and proof it in Appendix B. Theorem 1. For a shallow GCN module, Equation (8) can be approximated by ϕ Φ ˆλϕλϕ ˆf (LG) ϕ f (LG) ϕ Particularly, if LG = 0 and α1 = 0, our method is equivalent to Bo VW with a linear classifier. Further, as delineated in Corollary 1, the interpretability of the graph matcher with a shallow GCN can be stated as: (1) the final prediction score for a category is the summation of all present class evidence (shared vertices ϕ in Φ) represented by the vertex weights; (2) the shared neighbors (local structure) connected to the shared vertex ϕ also contribute to the final prediction. 2.4 TRAINING SCHEMANET Except for training Schema Net by cross-entropy loss LCE between the predictions and ground truths, we further constrain the complexity of IR-Atlas. More formally, the complexity is defined for both edges and vertex weights of all graphs in IR-Atlas as c=1 H( ˆΛc), Le = 1 u ˆV H( ˆEc,u), (10) where function H(x) computes the entropy of the input vector x Rk + normalized by its sum of k components, and ˆEc,u denotes the weighted edges connected to vertex u in ˆGc of class c. The final optimization goal is L = LCE + γv Lv + γe Le, (11) where γv and γe are hyperparameters. The overall training procedure is shown in Appendix C. Sparsification. Each graph in IR-Atlas is initialized as a fully-connected graph with random vertex and edge weights. However, this requires O(C|V |2) memory space for storage and training the whole set of edges. To alleviate this issue, we initialize IR-Atlas by averaging the instance IR-Graphs for each class and remove the edges connected to vertices whose weights are below a given threshold δt = 0.01 from the category-specific graph. Such a procedure will not only dramatically decrease the learnable parameters but also boost the final performance, as shown in Table 1. 3 EXPERIMENTS 3.1 IMPLEMENTATION Datasets. We evaluate our method on CIFAR-10/100 (Krizhevsky et al., 2009), Caltech-101 (Li et al., 2022), and Image Net (Deng et al., 2009). Particularly, Caltech-101 has around 9k images, and we manually split it into a training set with 7.4k images and a test set with 1.3k images. Published as a conference paper at ICLR 2023 Table 1: Comparison results on CIFAR-10/100 and Caltech-101 with different backbones. We report the top-1 accuracy, number of learnable parameters, and FLOPs. The results in gray color are merely for listing the accuracy of Vi Ts rather than for making comparisons. Backbone Method CIFAR-10 CIFAR-100 Caltech-101 Acc #param. FLOPs Acc #param. FLOPs Acc #param. FLOPs - Bo VW-SIFT 20.20 - - - - - - - - Base 96.69 5.53M 1.27G 82.88 5.54M 1.27G 92.53 5.54M 1.27G Backbone-FC 94.58 1.93K 1.06G 76.74 19.3K 1.06G 76.71 19.5K 1.06G Bo VW-Deep 94.95 10.3K 1.06G 75.45 103K 1.06G 60.82 104K 1.06G Bag Net 70.27 5.73M 1.10G 42.90 5.84M 1.10G 72.54 5.85M 1.10G Schema Net 95.92 396K 1.19G 77.11 105M 2.40G 83.24 106M 2.40G Schema Net-Init 95.96 234K 1.06G 78.45 573K 1.06G 87.57 564K 1.06G Dei T-Small Base 97.77 21.7M 4.63G 87.65 21.7M 4.63G 94.96 21.7M 4.63G Backbone-FC 96.62 3.85K 3.87G 82.44 38.5K 3.87G 86.94 38.9K 3.87G Bo VW-Deep 95.61 10.3K 3.89G 77.39 103K 3.89G 65.54 104K 3.89G Bag Net 83.90 22.1M 3.95G 50.49 22.2M 3.95G 78.13 22.2M 3.95G Schema Net 97.42 396K 4.00G 82.21 105M 5.21G 84.66 106M 5.21G Schema Net-Init 97.35 235K 3.89G 82.46 591K 3.89G 90.09 589K 3.89G Base 98.41 85.8M 17.6G 89.17 85.9M 17.6G 95.83 85.9M 17.6G Backbone-FC 97.04 7.69K 14.7G 81.66 76.9K 14.7G 88.20 77.7K 14.7G Bo VW-Deep 95.50 10.3K 14.7G 72.23 103K 14.7G 66.17 104K 14.7G Bag Net 90.71 86.6M 14.9G 67.84 86.8M 14.9G 86.55 86.8M 14.9G Schema Net 97.26 396K 14.8G 79.26 105M 16.1G 81.12 106M 16.1G Schema Net-Init 97.07 235K 14.7G 79.36 606K 14.7G 90.72 593K 14.7G Selection of the hyperparameters. Several hyperparameters are involved in our method, including λv,e in Equation (11) for adjusting the sparsity of IR-Atlas. We set λv = 0.5 and λe = 0.75 as the default value for the following evaluation and the sensitive analyses are presented in Appendix H. The initial values of learnable α1,2 and β1,2 are set to 0.5, and we show their learning curves in Figure 9. 3.2 EXPERIMENTAL RESULTS We evaluate our method and comparison baselines with the following settings: Bo VW-SIFT: the traditional Bo VW approach that utilizing SIFT feature for constructing the visual vocabulary (Yang et al., 2007). Base: the base Vi T model directly trained on the benchmark datasets with initial weights obtained from the official repository in (Touvron et al., 2021), which is our backbone. Backbone-FC: the frozen backbone with an FC layer. The intermediate features extracted from the frozen backbone are then fed to a global average pooling layer followed by a linear classifier, similar to the standard CNN protocol. Bo VW-Deep: the Bo VW approach with our extracted visual vocabulary. Bag Net: the implementation of Bag Net (Brendel & Bethge, 2019) with Vi T backbone which is constructed following our Base setting. Schema Net: our proposed Schema Net without initialization. Schema Net-Init: our proposed Schema Net initialized by the average instance IR-Graphs. The comparison results are shown in Table 1, where we demonstrate the top-1 accuracy, the number of learnable parameters, as well as the FLOPs for each setting. It is noticeable that the proposed Schema Net consistently outperforms the baseline methods. Specifically, for the backbone of Dei TTiny, ours achieves significant improvement (i.e., about 25.7% and 35.3% absolute gain on CIFAR-10 and CIFAR-100, respectively) over Bag Net. With a larger backbone such as Dei T-Small and Dei TBase, though the performance gap slightly shrinks, our Schema Net with initializations still yields results superior to Bag Net, by an average absolute gain of 12.7%. The reason is that, unlike Bag Net that uses summation to obtain the similarity map as the Bo VW representation (without feature discretization), Bo VW-Deep is implemented with the discretized visual words. As such, both of Published as a conference paper at ICLR 2023 0 10 20 30 40 50 60 70 80 90 100 Drop Rate (%) Bo VW-Deep+ Bo VW-Deep Bag Net+ Bag Net Ours+ Ours (a) Accuracy curves on CIFAR-10. 0 10 20 30 40 50 60 70 80 90 100 Drop Rate (%) Bo VW-Deep+ Bo VW-Deep Bag Net+ Bag Net Ours+ Ours (b) Accuracy curves on CIFAR-100. Figure 4: Accuracy decay curves of perturbation tests. The curve name ending with + / represents positive/negative perturbation, accordingly. the approaches do not consider the visual word interactions, leading to the inferior performance especially when the size of the visual vocabulary expands on CIFAR-100 and Caltech-101. Moreover, compared with Backbone-FC that merely relies on intermediate deep features, the proposed Schema Net also demonstrates encouraging results, achieving about a 1.4% absolute gain. Furthermore, when initialization and redundant vertices are removed as indicated in Section 3.1, our Schema Net-Init , with an even less learnable number of parameters, still leads to a higher accuracy. The proposed method also delivers gratifying performance, on par with the prevalent deep Vi T especially on CIFAR-10, but gives a clear insight on the reasoning procedure. We present more experimental results in Appendix F, including the comparison results on Image Net (Appendix F.1), employing other attribution methods (Appendix F.2), attacking with adversarial images (Appendix F.3), the extendability analysis of our Schema Net (Appendix F.4), and the inference cost of all components (Appendix F.5). 3.3 EVALUATION OF THE INTERPRETABILITY To figure out how interpretability contributes to the model prediction, we adopt the positive and negative perturbation tests presented in (Chefer et al., 2021) to give both quantitative and qualitative evaluations on the interpretability. For a fair comparison, the attention values to the CLS token ϕCLS are extracted as the pixel relevance for Bo VW-Deep, Bag Net, and Schema Net-Init with the Dei T-Tiny backbone. During the testing stage, we gradually drop the pixels in relevance descending/ascending order (for positive and negative perturbation, respectively) and measure the top-1 accuracy of the models. We plot the accuracy curves in Figure 4, showing that: (1) in the positive perturbation test, our method behaves similarly to Bo VW while outperforming Bag Net; (2) in the negative perturbation test, our method achieves better performance by a large margin (the area-under-the-curve of ours is about 5.71% higher than Bag Net on CIFAR-10, and 11.23% higher than Bag Net on CIFAR-100). 3.4 VISUALIZATION In Figure 5, we show the examples including instance IR-Graphs and learned IR-Atlas with Dei T-Tiny backbone trained on Caltech-101 dataset (for visualization purposes, the number of ingredients is set to 256, which are shown entirely in Figure 11). Each row shows several instances of a specific class. The first column includes the category graphs and excerpts of appeared ingredients in Figure 11 for quicker reference. We further render the vertices, i.e., ingredients, in the category graph with different colors, and the appeared ingredients in the instance images and graphs are colored uniformally corresponding to the category graph. We interpret the visualization from three perspectives: the interpretability of the ingredients, the interpretability of the edges, and the consistency between instance and category graphs. (1) Thanks to the powerful representation learning capability of DNNs, the extracted ingredients are able to represent explicit semantics (such as fuselage , bird legs , etc.) with robustness. Besides, the Published as a conference paper at ICLR 2023 IR-Atlas Instance IR-Graphs 17 48 204 213 234 196 0.047 0.043 2 3 111 133 207 48 #46 0.52 #48 0.33 #48 0.48 #19 0.32 6 44 129 236 241 117 0.034 0.046 #66 0.19 #6 0.17 Figure 5: Examples of the learned IR-Atlas and instance IR-Graphs randomly sampled from three categories ( panda , flamingo , and airplanes ) on Caltech-101. Please zoom in for a better view. learned weight of each vertex is also consistent with human intuition as the object ingredients get higher weights than the background ones. (2) The learned edge weights tend to connect object parts to the adjacent background and other object parts while ignoring the background connections, which helps distinguish fine-grain categories with the perception of the surroundings. (3) As the category graphs are driven by the instances, they learn to capture and model the general character of instances and eventually form the abstract imaginations of all categories. More examples are in Appendix I. 4 CONCLUSION AND OUTLOOK In this paper, we propose a novel inference paradigm, named schema inference, guided by Kant s philosophy towards resembling human deductive reasoning of associating the abstract concept image with the specific sense impression. To this end, we reformulate the traditional DNN inference into a graph matching scheme by evaluating the similarity between instance-level IR-Graph and category-level imagination in a deductive manner. Specifically, the graph vertices are visual semantics represented by common feature vectors from DNN s intermediate layer. Besides, the edges indicate the vertex interactions characterized by semantic similarity and spatial adjacency, which facilitate capturing the compositional contributions to the predictions. Theoretical analysis and experimental results on several benchmarks demonstrate the superiority and interpretability of schema inference. In future work, we will implement schema inference to more complicated vision tasks, such as visual question answering, that enables linking the visual semantics to the phrases in human language, achieving a more powerful yet interpretable reasoning paradigm. Published as a conference paper at ICLR 2023 ACKNOWLEDGMENTS This work is supported by National Natural Science Foundation of China (61976186, U20B2066, 62106220), Ningbo Natural Science Foundation (2021J189), the Starry Night Science Fund of Zhejiang University Shanghai Institute for Advanced Study (Grant No. SN-ZJU-SIAS-001), Open Research Projects of Zhejiang Lab (NO. 2019KD0AD01/018) and the Fundamental Research Funds for the Central Universities (2021FZZX001-23). Marco Ancona, Cengiz Oztireli, and Markus Gross. Explaining deep neural networks with a polynomial time algorithm for shapley value approximation. In Kamalika Chaudhuri and Ruslan Salakhutdinov (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp. 272 281. PMLR, 09 15 Jun 2019. URL https://proceedings.mlr.press/v97/ancona19a.html. Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. ar Xiv preprint ar Xiv:1607.06450, 2016. Sebastian Bach, Alexander Binder, Gr egoire Montavon, Frederick Klauschen, Klaus-Robert M uller, and Wojciech Samek. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. Plo S one, 10(7):e0130140, 2015. Wieland Brendel and Matthias Bethge. Approximating CNNs with bag-of-local-features models works surprisingly well on imagenet. In International Conference on Learning Representations, 2019. Xiaojun Chang, Pengzhen Ren, Pengfei Xu, Zhihui Li, Xiaojiang Chen, and Alexander G Hauptmann. A comprehensive survey of scene graphs: Generation and application. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021. Hila Chefer, Shir Gur, and Lior Wolf. Transformer interpretability beyond attention visualization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 782 791, 2021. Chaofan Chen, Oscar Li, Daniel Tao, Alina Barnett, Cynthia Rudin, and Jonathan K Su. This looks like that: deep learning for interpretable image recognition. Advances in neural information processing systems, 32, 2019. Huiqi Deng, Qihan Ren, Hao Zhang, and Quanshi Zhang. Discovering and explaining the representation bottleneck of dnns. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=i RCUlgmdf HJ. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248 255. Ieee, 2009. Josip Djolonga, Jessica Yung, Michael Tschannen, Rob Romijnders, Lucas Beyer, Alexander Kolesnikov, Joan Puigcerver, Matthias Minderer, Alexander D Amour, Dan Moldovan, et al. On robustness and transferability of convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16458 16468, 2021. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021. Zunlei Feng, Jiacong Hu, Sai Wu, Xiaotian Yu, Jie Song, and Mingli Song. Model doctor: A simple gradient aggregation strategy for diagnosing and treating cnn classifiers. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 616 624, 2022. Published as a conference paper at ICLR 2023 Hiroshi Fukui, Tsubasa Hirakawa, Takayoshi Yamashita, and Hironobu Fujiyoshi. Attention branch network: Learning of attention mechanism for visual explanation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10705 10714, 2019. Yunhao Ge, Yao Xiao, Zhi Xu, Meng Zheng, Srikrishna Karanam, Terrence Chen, Laurent Itti, and Ziyan Wu. A peek into the reasoning of neural networks: Interpreting with structural visual concepts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2195 2204, 2021. Spyros Gidaris, Andrei Bursuc, Nikos Komodakis, Patrick P erez, and Matthieu Cord. Learning representations by predicting bags of visual words. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6928 6938, 2020. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Dan Hendrycks, Steven Basart, Norman Mu, Saurav Kadavath, Frank Wang, Evan Dorundo, Rahul Desai, Tyler Zhu, Samyak Parajuli, Mike Guo, et al. The many faces of robustness: A critical analysis of out-of-distribution generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8340 8349, 2021. Roei Herzig, Moshiko Raboh, Gal Chechik, Jonathan Berant, and Amir Globerson. Mapping images to scene graphs with permutation-invariant structured prediction. Advances in Neural Information Processing Systems, 31, 2018. Adrian Hoffmann, Claudio Fanconi, Rahul Rade, and Jonas Kohler. This looks like that... does it? shortcomings of latent space prototype interpretability in deep networks. Co RR, abs/2105.02968, 2021. URL https://arxiv.org/abs/2105.02968. Yongcheng Jing, Yiding Yang, Xinchao Wang, Mingli Song, and Dacheng Tao. Amalgamating knowledge from heterogeneous graph neural networks. In CVPR, 2021a. Yongcheng Jing, Yiding Yang, Xinchao Wang, Mingli Song, and Dacheng Tao. Meta-aggregator: Learning to aggregate for 1-bit graph neural networks. In ICCV, 2021b. Yongcheng Jing, Yining Mao, Yiding Yang, Yibing Zhan, Mingli Song, Xinchao Wang, and Dacheng Tao. Learning graph neural networks for image style transfer. In ECCV, 2022. Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. ar Xiv preprint ar Xiv:1609.02907, 2016. Pang Wei Koh, Thao Nguyen, Yew Siang Tang, Stephen Mussmann, Emma Pierson, Been Kim, and Percy Liang. Concept bottleneck models. In International Conference on Machine Learning, pp. 5338 5348. PMLR, 2020. Yael Konforti, Alon Shpigler, Boaz Lerner, and Aharon Bar-Hillel. Inference graphs for cnn interpretation. In European Conference on Computer Vision, pp. 69 84. Springer, 2020. Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6):84 90, 2017. Fei-Fei Li, Marco Andreeto, Marc Aurelio Ranzato, and Pietro Perona. Caltech 101, 2022. URL https://data.caltech.edu/records/20086. Qimai Li, Zhichao Han, and Xiao-Ming Wu. Deeper insights into graph convolutional networks for semi-supervised learning. In Sheila A. Mc Ilraith and Kilian Q. Weinberger (eds.), Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, pp. 3538 3545, 2018. Songhua Liu, Kai Wang, Xingyi Yang, Jingwen Ye, and Xinchao Wang. Dataset distillation via factorization. In Neur IPS, 2022. Published as a conference paper at ICLR 2023 Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. Open Review.net, 2019. URL https://openreview.net/forum?id=Bkg6Ri Cq Y7. David G Lowe. Distinctive image features from scale-invariant keypoints. International journal of computer vision, 60(2):91 110, 2004. Wenjie Luo, Yujia Li, Raquel Urtasun, and Richard Zemel. Understanding the effective receptive field in deep convolutional neural networks. Advances in neural information processing systems, 29, 2016. Gr egoire Montavon, Sebastian Lapuschkin, Alexander Binder, Wojciech Samek, and Klaus-Robert M uller. Explaining nonlinear classification decisions with deep taylor decomposition. Pattern Recognition, 65:211 222, 2017. ISSN 0031-3203. doi: https://doi.org/10.1016/j.patcog. 2016.11.008. URL https://www.sciencedirect.com/science/article/pii/ S0031320316303582. Christopher Morris, Martin Ritzert, Matthias Fey, William L Hamilton, Jan Eric Lenssen, Gaurav Rattan, and Martin Grohe. Weisfeiler and leman go neural: Higher-order graph neural networks. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 4602 4609, 2019. Meike Nauta, Ron van Bree, and Christin Seifert. Neural prototype trees for interpretable fine-grained image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 14933 14943, June 2021. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019. MDJ Peters. Extending explainability of generative classifiers with prototypical parts. Master s thesis, University of Twente, 2022. Dawid Rymarczyk, Aneta Kaczy nska, Jarosław Kraus, Adam Pardyl, and Bartosz Zieli nski. Protomil: Multiple instance learning with prototypical parts for fine-grained interpretability. ar Xiv preprint ar Xiv:2108.10612, 2021. Ramprasaath R Selvaraju, Michael Cogswell, Abhishek Das, Ramakrishna Vedantam, Devi Parikh, and Dhruv Batra. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision, pp. 618 626, 2017. Suprosanna Shit, Rajat Koner, Bastian Wittmann, Johannes Paetzold, Ivan Ezhov, Hongwei Li, Jiazhen Pan, Sahand Sharifzadeh, Georgios Kaissis, Volker Tresp, et al. Relationformer: A unified framework for image-to-graph generation. ar Xiv preprint ar Xiv:2203.10202, 2022. Avanti Shrikumar, Peyton Greenside, and Anshul Kundaje. Learning important features through propagating activation differences. In International conference on machine learning, pp. 3145 3153. PMLR, 2017. Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. In Workshop at International Conference on Learning Representations, 2014. Erik Strumbelj and Igor Kononenko. An efficient explanation of individual classifications using game theory. The Journal of Machine Learning Research, 11:1 18, 2010. Mukund Sundararajan, Ankur Taly, and Qiqi Yan. Axiomatic attribution for deep networks. In International conference on machine learning, pp. 3319 3328. PMLR, 2017. Ilya O Tolstikhin, Neil Houlsby, Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Thomas Unterthiner, Jessica Yung, Andreas Steiner, Daniel Keysers, Jakob Uszkoreit, et al. Mlp-mixer: An all-mlp architecture for vision. Advances in Neural Information Processing Systems, 34: 24261 24272, 2021. Published as a conference paper at ICLR 2023 Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herv e J egou. Training data-efficient image transformers & distillation through attention. In International Conference on Machine Learning, pp. 10347 10357. PMLR, 2021. Suvidha Tripathi, Satish Kumar Singh, and Hwee Kuan Lee. Bag of visual words (bovw) with deep features - patch classification model for limited dataset of breast tumours, 2022. URL https://europepmc.org/article/PPR/PPR507364. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. Oriol Vinyals, Charles Blundell, Timothy Lillicrap, Daan Wierstra, et al. Matching networks for one shot learning. Advances in neural information processing systems, 29, 2016. Lauren J Wong and Sean Mc Pherson. Explainable neural network-based modulation classification via concept bottleneck models. In 2021 IEEE 11th Annual Computing and Communication Workshop and Conference (CCWC), pp. 0191 0196. IEEE, 2021. Mengqi Xue, Qihan Huang, Haofei Zhang, Lechao Cheng, Jie Song, Minghui Wu, and Mingli Song. Protopformer: Concentrating on prototypical parts in vision transformers for interpretable image recognition. ar Xiv preprint ar Xiv:2208.10431, 2022a. Mengqi Xue, Haofei Zhang, Qihan Huang, Jie Song, and Mingli Song. Learn decision trees with deep visual primitives. Journal of Visual Communication and Image Representation, 89: 103682, 2022b. ISSN 1047-3203. doi: https://doi.org/10.1016/j.jvcir.2022.103682. URL https: //www.sciencedirect.com/science/article/pii/S1047320322002024. Jun Yang, Yu-Gang Jiang, Alexander G Hauptmann, and Chong-Wah Ngo. Evaluating bag-of-visualwords representations in scene classification. In Proceedings of the international workshop on Workshop on multimedia information retrieval, pp. 197 206, 2007. Xingyi Yang, Zhou Daquan, Songhua Liu, Jingwen Ye, and Xinchao Wang. Deep model reassembly. In Neur IPS, 2022a. Xingyi Yang, Jingwen Ye, and Xinchao Wang. Factorizing knowledge in neural networks. In ECCV, 2022b. Yiding Yang, Xinchao Wang, Mingli Song, Junsong Yuan, and Dacheng Tao. Spagan: Shortest path graph attention network. In IJCAI, 2019. Yiding Yang, Zunlei Feng, Mingli Song, and Xinchao Wang. Factorizable graph convolutional networks. Neur IPS, 2020a. Yiding Yang, Jiayan Qiu, Mingli Song, Dacheng Tao, and Xinchao Wang. Distilling knowledge from graph convolutional networks. In CVPR, 2020b. Ruonan Yu, Songhua Liu, and Xinchao Wang. Dataset distillation: A comprehensive review. ar Xiv preprint ar Xiv:2301.07014, 2023. Alireza Zareian, Zhecan Wang, Haoxuan You, and Shih-Fu Chang. Learning visual commonsense for robust scene graph generation. In European Conference on Computer Vision, pp. 642 657. Springer, 2020. Mateo Espinosa Zarlenga, Pietro Barbiero, Gabriele Ciravegna, Giuseppe Marra, Francesco Giannini, Michelangelo Diligenti, Zohreh Shams, Frederic Precioso, Stefano Melacci, Adrian Weller, et al. Concept embedding models. ar Xiv preprint ar Xiv:2209.09056, 2022. Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In European conference on computer vision, pp. 818 833. Springer, 2014. Quanshi Zhang, Xin Wang, Ruiming Cao, Ying Nian Wu, Feng Shi, and Song-Chun Zhu. Extraction of an explanatory graph to interpret a cnn. IEEE transactions on pattern analysis and machine intelligence, 43(11):3863 3877, 2020a. Published as a conference paper at ICLR 2023 Quanshi Zhang, Xin Wang, Ying Nian Wu, Huilin Zhou, and Song-Chun Zhu. Interpretable cnns for object classification. IEEE transactions on pattern analysis and machine intelligence, 43(10): 3416 3431, 2020b. Zaixi Zhang, Qi Liu, Hao Wang, Chengqiang Lu, and Cheekong Lee. Protgnn: Towards selfexplaining graph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 9127 9135, 2022. Lingxiao Zhao and Leman Akoglu. Pairnorm: Tackling oversmoothing in gnns. In International Conference on Learning Representations, ICLR, 2020. Yiwu Zhong, Jing Shi, Jianwei Yang, Chenliang Xu, and Yin Li. Learning to generate scene graph from natural language supervision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1823 1834, 2021. Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. Learning deep features for discriminative localization. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2921 2929, 2016. Luisa M Zintgraf, Taco S Cohen, Tameem Adel, and Max Welling. Visualizing deep neural network decisions: Prediction difference analysis. In International Conference on Learning Representations, 2017. URL https://openreview.net/forum?id=BJ5Ue U9xx. Published as a conference paper at ICLR 2023 A ADDITIONAL RELATED WORK In this section, we introduce literature that is related to our method. A.1 VISUAL CONCEPT ANALYSIS Visual concept analysis aims to interpret the behavior of pre-trained deep models by extracting and tracking intuitive visual concepts during the DNN s inference procedure. Previous works use masks (Zhang et al., 2020b), explanatory graphs (Zhang et al., 2020a) and probabilistic models (Konforti et al., 2020) to interpret the internal layers of convolutional neural networks (CNNs) (Liu et al., 2022; Yu et al., 2023). Further, Deng et al. (2022) propose to capture and analyze the interaction of visual concepts contributing cooperatively to the prediction results for CNNs. However, they use Shapley values to compute the interaction, which is computationally expensive. Consequently, we adopt vision Transformers (Vi Ts) (Dosovitskiy et al., 2021; Touvron et al., 2021) as the DNN backbone of our method instead of conventional CNNs thanks to the multi-head self-attention (MHSA) mechanism (Vaswani et al., 2017) that explicitly encodes the interactions of visual tokens in the similarity level. More recently, VRX (Ge et al., 2021) is proposed to use a graphical model to interpret the prediction of a pre-trained DNN, in which edges are constructed to represent the spatial relation of visual semantics. They further utilize a GCN module to make a prediction with the input structural concept graphs by training to mimic DNN s prediction. Nevertheless, their approach requires training a graph for every input sample, which is extremely expensive for applying to large-scale datasets. Besides, VRX does not construct the class imagination that carries the category-specific knowledge, meaning that the GCN predictor must learn to memorize and distinguish all the class-specific graphs, which impairs the overall interpretability. Our proposed schema inference, however, explicitly creates an interactive matching procedure between the instances and category imagination. Moreover, our GNN model only captures the local structure for gathering class evidence, which is easier to accomplish. A.2 VISUAL CONCEPT LEARNING Visual concept learning refers to a range of methods that mine semantic-rich visual words from a DNN, which are further utilized to make interpretable predictions towards a self-explanatory model in reality. Concept bottleneck models (CBMs) (Koh et al., 2020; Zarlenga et al., 2022; Deng et al., 2022; Wong & Mc Pherson, 2021) link the neurons with human-interpretable semantics explicitly, encouraging the trustworthiness of DNNs. Human interventions of learned bottleneck layers can fix the misclassified concepts to improve the model performance. Another line of visual concept learning, part-prototype-based methods, collectively makes predictions on target tasks with DNN and semantic-rich prototypes, which can be divided into two schools according to their prototypes. (1) Bag-of-Visual-Word (Bo VW) approaches (Brendel & Bethge, 2019; Gidaris et al., 2020; Tripathi et al., 2022) obtain prototypes with semantics through passively clustering the hand-crafted features (Yang et al., 2007) or deep features into a set of discrete embeddings (visual words) that relate to specific visual semantics. The prediction procedure of these approaches is analog to the Bo W model in natural language processing (NLP), in that an interpretable representation is constructed based on the statistics of the occurrence of the visual words, and then fed to an interpretable classifier, such as the linear classifier or decision trees. (2) In contrast, part-prototypical networks (Proto PNets) (Chen et al., 2019) and the following works (Nauta et al., 2021; Xue et al., 2022a; Zhang et al., 2022; Peters, 2022; Rymarczyk et al., 2021) jointly train DNN (as the feature extractor) and parameterized prototypes. Then decisions are made based on the linear combination of similarity scores between prototypes and feature vectors at all the spatial locations. Despite the relatively high performance, as discussed in (Brendel & Bethge, 2019; Hoffmann et al., 2021), the similarity of the learned prototypes and deep features in the embedding space may be significantly different in the input space. As such, we only compare those approaches in which prototypes are generated based on clustering for the same level of interpretability. To the best of our knowledge, none of the existing works have explored a deductive inference paradigm based on the interaction of visual semantics. Published as a conference paper at ICLR 2023 A.3 SCENE GRAPH GENERATION A scene graph can be generated from an input image to excavate the collection of objects, attributes, and relationships in a whole scene. Typical graph-based representations and learning algorithms (Herzig et al., 2018; Chang et al., 2021; Shit et al., 2022; Zhong et al., 2021; Zareian et al., 2020) adopt graph neural networks (GNNs) (Kipf & Welling, 2016; Jing et al., 2021a; Yang et al., 2020b; Jing et al., 2021b; Yang et al., 2020a; Jing et al., 2022; Yang et al., 2019) to model the relationships between objects in a scene. Similar to scene graphs, our proposed Schema Net also utilizes GNNs to represent the relationships between ingredients related to local semantics. However, the nodes in GNNs of Schema Net represent more fine-grained representations, i.e., local semantics of objects, compared to those relating to the whole objects in scene graphs. Moreover, in Schema Net, the trained GNN is responsible for estimating the similarity between an instance IR-Graph and category IR-Graphs and making classification based on the similarity scores. A.4 FEATURE ATTRIBUTION FOR EXPLAINABLE AI In the Feat2Vertex module, we extract the attention to the CLS token as one important component of the ingredient importance. Currently, plenty of approaches have been proposed for indicating local relevance (forming saliency maps) to the DNN s prediction, termed feature attribution methods. Most existing works can be roughly divided into three classes: perturbation-based, gradient-based, and decomposition-based approaches. Perturbation-based approaches (Strumbelj & Kononenko, 2010; Zeiler & Fergus, 2014; Ancona et al., 2019; Zintgraf et al., 2017) compute attribution by evaluating output differences via removing or altering input features. Gradient-based approaches (Simonyan et al., 2014; Shrikumar et al., 2017; Sundararajan et al., 2017; Selvaraju et al., 2017; Feng et al., 2022) compute gradients w.r.t. the input feature through backpropagation. Decomposition-based approaches (Bach et al., 2015; Montavon et al., 2017; Chefer et al., 2021) propagate the final prediction to the input following the Deep Taylor Decomposition (Montavon et al., 2017). Besides, CAM (Zhou et al., 2016) and ABN (Fukui et al., 2019) provide interpretable predictions with a learnable attribution module. As most previous methods focus on CNNs, Chefer et al. (2021) propose Vi T-LRP tailored for vision Transformers. However, most of the methods mentioned above are designed to generate a saliency map for a particular class, making them inefficient in our implementation due to the O(C) computational complexity or dependency on gradients. B PROOF OF THEOREM 1 We first restate the theorem: Theorem (Theorem 1 restated). For a shallow GCN module, Equation (8) can be approximated by ϕ Φ ˆλϕλϕ ˆf (LG) ϕ f (LG) ϕ Particularly, if LG = 0 and α1 = 0, our method is equivalent to Bo VW with a linear classifier. For further discussion, we first prove the following lemma: Lemma 1. For random vectors f, g Rd G drawn independently from multivariate Gaussian distribution N(0, I), we have Ef,g f W Wg = 0 Ef f W Wf = W 2 F , (13) where F is the matrix Frobenius norm, and W Rd G d G is a projection matrix. Proof. To begin with, we expand term f W Wg as t=1 fsgtwr,swr,t. (14) Therefore, the expectation Ef,g f W Wg = t=1 wr,swr,t Efs [fs] Egt [gt] = 0, (15) Published as a conference paper at ICLR 2023 0 32 64 96 128 160 192 224 (a) Raw vertex embeddings. 0 32 64 96 128 160 192 224 (b) Last layer embeddings. Figure 6: Vertex embedding similarities between a vertex v in the instance IR-Graph and a vertex u in the category IR-Atlas, which are extracted from the original vertex embeddings and output from the last Graph Conv layer. We show the average similarity over 1000 samples on the Caltech-101 dataset with a visual vocabulary size of 256. Ef f W Wf = t=1 wr,swr,tfsft s=1 w2 r,sf 2 s s=1 w2 r,s = W 2 F . Particularly, if W is an identity matrix Ef f W Wf = d G. (17) Further, if the components of matrix W are drawn i.i.d. from Gaussian distribution N(0, 1) and are independent with f, Ef,W f W Wf = d2 G. (18) Now we proof Theorem 1. Proof. We start with a simple case that the depth of GCN module is zero. As G and ˆG share the same original vertex embeddings, we have ˆ λuλv ˆf (0) u f (0) v = X ˆ λuλvf (0) u f (0) v . (19) According to Lemma 1, the expectation of the same vertex term E ff = d G is more significant than different vertices terms ˆ λuλvf (0) u f (0) v for u = v when G and ˆG are constrained with sparsity. In Figure 6(a), we visualize the vertex similarities f (0) u f (0) v , showing that the similarities between different vertices are significantly lower than that of the same vertex. Published as a conference paper at ICLR 2023 Consequently, Equation (8) in the case of LG = 0 can be simplified as ϕ Φ ˆλϕλϕ f (0) ϕ 2 2, (20) where Φ = V ˆV is the set of shared vertices of G and ˆG. With an extra assumption that α1 = 0, λϕ therefore represents the count of visual word ϕ. Combing the learnable term ˆλϕ f (0) ϕ 2 2 as wϕ, the prediction of the matcher is y = ΛWFC, (21) where WFC R|V | C is the learnable projection matrix of a linear classifier. Now, we prove the whole theorem. For different vertices u ˆV and v V , the summation term su,v = ˆλuλv ˆf (LG) u f (LG) v can be rewritten as the aggregation form with feature output from the (LG 1)-th layer: su,v = ˆλuλv X (i,j) N ˆ G(u) NG(v) ˆeu,iev,j ˆf (LG 1) i W (LG)W (LG) f (LG 1) j where N ˆ G(u) represent the neighbors of vertex u in Graph ˆG and with u itself, and eu,u = 1 as the self loop. With the assumption that ˆf (LG 1) i and f (LG 1) j are whitened random vectors (we say a random vector x Rd is whitened if all the components are zero mean and independent from each other, which can be achieved by normalization), the summation term ˆf (LG 1) i W (LG)W (LG) f (LG 1) j with vertex i = j will be significantly larger than those with different vertices. We further illustrate the vertex similarities output from the final layer in Figure 6(b), showing that ˆf (LG) u f (LG) v is conspicuous for vertices u = v. Corollary 1. Consequently, su,v is significant if N ˆ G(u) and NG(v) have shared vertices, relative large edge weight product ˆeu,iev,j, and vertex weight product ˆλuλv, which means u and v have similar local structure, particularly in the case that u and v are the same vertex. In conclusion, Equation (8) can be simplified as the summation of the joint vertices Φ = ˆV V of the instance and category graph G and ˆG. C TRAINING ALGORITHM The training algorithm of our proposed Schema Net is shown in Algorithm 1 with initializing IR-Atlas and sparsification. D EFFECTIVE RECEPTIVE FIELD OF VITS In this section, we discuss the effective receptive field (ERF) of Vi Ts, which is crucial as the visual token in the intermediate layers of Vi Ts may relate to other positions due to MHSA, affecting the interpretability of the ingredients (visual word semantics). Although Luo et al. (2016) propose to measure the ERF for CNNs, it cannot be directly implemented to Transformer-base models. In this section, we propose a relatively simple yet effective approach to measuring the ERF of Vi Ts. Definition 3 (Transformer ERF). Supposed that visual sequence X Rn d is fed into a Vi T backbone with N layers, x is a randomly chosen anchor token in X, and ε Rd is the random vector drawn from Gaussian distribution N(0, I) normalizing to the unit vector. The Transformer ERF is the average Euclidean distance between the closest token y to x in the 2D token array and x so that when y is disturbed to ˆy = y + ϵ the output change of token x is less than a predefined threshold δr > 0. Published as a conference paper at ICLR 2023 Algorithm 1 Schema Net optimizer with initialization and sparsification. Input: D = {(xi, yi)}D i=1: the training dataset with D samples; ˆG = { ˆGc}C c=1: initial IR-Atlas; Backbone( ): the Vi T backbone; Matcher( , ): the graph matcher; Ω: the visual vocabulary; Θ: the set of all trainable parameters; δt: the sparsification threshold. 1: procedure INITIALIZATION(D, Ω) 2: Sample a subset D D as the probe dataset. 3: S1, . . . , SC , ˆG 4: for (x, y) D do 5: G FEAT2GRAPH(Backbone(x), Ω) 6: Sy Sy G 7: end for 8: for c = 1, . . . , C do 9: ˆG AVERAGE(Sc) compute average graph for each category 10: end for 11: return ˆG 12: end procedure 13: procedure TRAINING( ˆG, D, Ω) 14: for (x, y) D do 15: for ˆei,j ˆE do Removing redundant edges 16: if ˆλi < δt or ˆλj < δt then 17: ˆei,j NIL 18: end if 19: end for 20: G FEAT2GRAPH(Backbone(x), Ω) 21: ˆy Matcher(G, ˆG) 22: Compute the final loss and gradient Θ w.r.t. parameters Θ. 23: Update parameters Θ with Adam W optimizer. 24: end for 25: end procedure 26: ˆG INITIALIZATION(D, Ω) 27: TRAINING( ˆG, D, Ω) (a) Dei T-Tiny (b) Dei T-Small (c) Dei T-Base Figure 7: The heat maps of the receptive field of an anchor token over its neighborhood region. Averaged visualization results of 64 random images of three Vi Ts (Dei T-Ti, Dei T-S, and Dei T-B) from Caltech-101 dataset are demonstrated. The full image size is 224 224, and we crop a small region with a size of 96 96 with the anchor token as the center for better visualization. Figure 7 shows the visualization heatmap w.r.t. the change of the output anchor with different Vi T backbones adopted in our method. As we can observe, for an input image size of 224 224, the output tokens with relatively significant changes are mainly distributed in the circle with a radius of 25. Therefore, in our method, all the ingredients are in charge of 50 50 patches corresponding to the input image. Published as a conference paper at ICLR 2023 E EXPERIMENTAL DETAILS E.1 VISUAL VOCABULARY Table 2: Visual vocabulary size. Dataset Size M CIFAR-10 128 CIFAR-100 1024 Caltech-101 1024 Thanks to the relatively uniform representation of visual semantics generated by DNNs, our visual vocabulary size M can be set to around 10 C with competitive performance. Table 2 shows the default choice of M on different datasets. Results with different vocabulary sizes are presented in Appendix G. Furthermore, all the intermediate features are extracted from the 9-th layer of the backbone Vi Ts, including Dei T-Tiny, Dei TSmall, and Dei T-Base. E.2 IR-GRAPH Before an instance-level graph G and a category-level graph ˆG are fed to the graph matcher, their vertex weights and edge weights are normalized as follows. The vertex weights in Λ are divided by the sum of Λ Λnorm = Λ P|V | i=1 λi . (23) The adjacency matrix E is divided by the row-wise summation: E1,: P|V | i=1 e1,i ... E|V |,: P|V | i=1 e|V |,i Finally, the symmetric adjacency matrix Esym is defined as 2(Enorm + E norm) (25) E.3 GCN SETTINGS Now we describe the detailed settings of the GCN module in the graph matcher. We adopt Layer Norm (Ba et al., 2016) as Norm( ) function and rectified linear unit (Re LU) as the activation in Equation (7). Besides, the embedding dimension d G is set to 256 for all the experiments. E.4 TRAINING DETAILS The matcher and IR-Atlas in our method are optimized by Adam W (Loshchilov & Hutter, 2019) with a learning rate of 10 3, weight decay of 5 10 4, and cosine annealing as the learning rate decay schedule. We implement our method with Pytorch (Paszke et al., 2019) and train all the settings for 50 epochs with the batch size of 64 on one NVIDIA Tesla A100 GPU. All input images are resized to 224 224 pixels before feeding to our Schema Net. We adopt Res Net-style data augmentation strategies: random-sized cropping and random horizontal flipping. F ADDITIONAL RESULTS This section contains additional experimental results, highlighting the efficiency, robustness, and extendability of our proposed schema inference. F.1 RESULTS ON IMAGENET We further implement Schema Net on mini-Image Net (Vinyals et al., 2016) and Image Net-1k (Deng et al., 2009). We adopt Dei T-Small as the backbone, and the visual vocabulary size is set to 1024 for mini-Image Net and 8000 for Image Net-1k (due to the memory constraint). Particularly, as implementing 1000 fully-connected category-level IR-Graphs is expensive, we keep at most 500 Published as a conference paper at ICLR 2023 valid vertices for each class while all other vertices are pruned during the training process based on the initialized vertex weights. All other settings are identical to the experiments on Caltech-101. The results are presented in Table 3, drawing consistent conclusions as the main results in Table 1. Table 3: Comparison results (top-1 accuracy) on Image Net-1k and mini-Image Net. Dataset Base Backbone-FC Bo VW-Deep Bag Net Schema Net Schema Net-Init mini-Image Net 95.35 90.04 90.24 85.64 89.40 91.02 Image Net-1k 79.90 69.23 58.95 68.92 - 74.05 F.2 COMPARISON RESULTS OF USING ATTRIBUTION METHODS Instead of using raw attention extracted from the backbone in Equation (3), we further employ feature attribution methods for computing the ingredient importance. Specifically, we compare the results using raw attention, CAM (Zhou et al., 2016), and Transformer-LRP (Chefer et al., 2021) in Table 4. We can observe that using raw attention is superior to the attribution methods in terms of accuracy and running time. Such results can be explained that: (1) the attribution methods compute a saliency map for each class (particularly, Vi T-LRP conduct C-times backpropagation, consuming enormous time), which is significantly slower than our implementation; (2) further, as the backbone predicts soft probability distributions other than one-hot targets, the computed saliency maps for similar categories will be highlighted to some extent, impairing the graph matcher. Table 4: Comparison results (top-1 accuracy and running time) of using different attribution methods on CIFAR-10 and CIFAR-100 datasets. Dataset Raw Attention CAM Vi T-LRP Acc Time Acc Time Acc Time CIFAR-10 95.96 0.8h 93.58 3.6h 91.28 53.8h CIFAR-100 78.45 3.7h 77.55 15.3h - - F.3 ADVERSARIAL ATTACKS To analyze the robustness of our proposed Schema Net, we evaluate the pre-trained Schema Net (from Image Net-1k described in Appendix F.1) on two popular adversarial benchmarks: Image Net-A (Djolonga et al., 2021) and Image Net-R (Hendrycks et al., 2021). The comparison results are shown in Table 5, revealing that our method is more robust than the baselines. Table 5: Adversarial attack results (top-1 accuracy) on Image Net-A and Image Net-R datasets. Dataset Backbone-FC Bo VW-Deep Bag Net Schema Net-Init Image Net-A 5.41 7.36 5.53 13.05 Image Net-R 31.21 24.17 30.93 33.46 F.4 EXTENDABILITY We evaluate the extendability of our method by extending a trained Schema Net to unseen tasks, while keeping the original framework frozen. As such, only the category-level IR-Graphs for the new tasks are optimized and inserted into the original IR-Atlas. Specifically, the tasks are disjointly drawn from Caltech-101 dataset. The Base task has 21 classes, and task 1 to 4 has 20 classes. The backbone model, i.e., Dei T-Tiny, is trained on the Base task and then is kept unchanged for the new tasks. The experimental results are presented in Table 6, revealing limited performance degradation for incoming new tasks. Such results show that the graph matcher is only responsible for evaluating the graph similarity, while the category knowledge is stored in the imagination, i.e., IR-Atlas. Published as a conference paper at ICLR 2023 Table 6: Top-1 accuracy results when extending Schema Net to unseen tasks. The accuracies are evaluated on the corresponding task, and the average accuracy over all the tasks is given as well. Average Base Base 97.57 97.57 Task 1 +Task 1 94.68 95.95 93.69 Task 2 +Task 2 93.36 95.14 93.06 92.11 Task 3 +Task 3 93.63 95.55 93.06 92.47 93.75 Task 4 +Task 4 92.37 95.14 92.74 91.76 90.42 91.49 64 128 256 512 1024 2048 M (a) Schema Net accuracy with different visual vocabulary size M. 1 2 3 4 5 LG (b) Schema Net accuracy when using GCN with various depth LG. 0.25 0.50 0.75 1.00 (c) Sensitivity analysis of hyperparameters involved in Equation (11). Figure 8: Ablation study and sensitivity analysis of hyperparameters of our proposed Schema Net on CIFAR-100 backed with Dei T-Tiny. F.5 FEAT2GRAPH EVALUATION We here present a performance bottleneck of our method, i.e., the Feat2Graph module, in which a feature ingredient array is converted to IR-Graph without benefiting from GPU parallel acceleration. Specifically, as duplicated visual words may exist after the feature discretization, which happens from time to time, we implement this module in C++ to achieve better random access performance. The experiments are conducted on one NVIDIA Tesla A100 GPU platform with AMD EPYC 7742 64-Core Processor. In Table 7, we show the average running time of each component with an input batch size of 64. We can see that without acceleration in parallel, the running time for Feat2Edge is significantly longer than others (about 1.75 ms per image). However, the inference time is still acceptable for real-time applications. Table 7: Time costing (ms) of the components in Schema Net with input batch size of 64. Dataset Backbone Feat2Vertex Feat2Edge Matcher CIFAR-10 4.89 7.00 19.8 1.95 CIFAR-100 4.31 9.50 101.7 1.30 Caltech-101 4.12 8.12 111.7 1.42 G ABLATION STUDY G.1 VISUAL VOCABULARY SIZE Figure 8(a) shows Schema Net accuracy when using different sizes of visual vocabulary. We can observe that even with a relatively small size, e.g., M = 256 on CIFAR-100, the performance is still competitive (about 7% absolute degradation). Besides, for the case of M = 2048, however, the performance suffers from low generalizability of the vocabulary. Published as a conference paper at ICLR 2023 0 5 10 15 20 25 30 35 40 45 50 Epochs Figure 9: Learning curves of α1,2 and β1,2 with Dei T-Tiny backbone on Caltech-101 dataset. G.2 NUMBER OF GRAPHCONV LAYERS The effect of using different depths of GCN is illustrated in Figure 8(b). As LG increases to 5, the accuracy on CIFAR-100 decreases rapidly, yielding that deep GCN matcher impairs the graph matching performance, which has been explored due to the over-smoothing caused by many convolutional layers (Zhao & Akoglu, 2020; Li et al., 2018). In our method, however, the graph matcher with only two layers of Graph Conv is proven to be adequate for high performance and interpretability. G.3 ABLATION OF WEIGHT COMPONENTS Table 8: Ablation study of the components defining the vertex and edge weights with Dei T-Tiny as the backbone DNN. Settings Accuracy CIFAR-10 CIFAR-100 Learnable α1,2, β1,2 95.96 78.20 Fixed α1 = 0, α2 = 1 95.88 75.91 Fixed α1 = 1, α2 = 0 95.68 72.17 Fixed β1 = 0, β2 = 1 95.85 77.31 Fixed β1 = 1, β2 = 0 95.89 77.84 Fixed α1,2 = β1,2 = 0.5 95.92 75.62 The effect of λCLS and λbag defined in Equation (3), and eattn and eadj defined in Equation (6) are shown in Table 8. By setting the corresponding weight (α1,2 and β1,2) to zero once at a time, we are able to analyze the components individually. In general, removing any term will lead to varying degrees of performance degradation. More significantly, when removing the term λbag, the accuracy decreases by 6.03% on CIFAR-100, revealing that the statistics of the ingredients help filter noisy vertices. The learning curves of the learnable weights are shown in Figure 9, from which we can observe that the model tend to adopt a relatively larger α2 (for visual word count) while keep λCLS v term for filtering noisy and background ingredients. H SENSITIVITY ANALYSIS OF HYPERPARAMETERS H.1 SPARSIFICATION THRESHOLD Table 9 presents the Schema Net-Init top-1 accuracy trained with different sparsification threshold δt. We can observe that an appropriate value of δt, e.g., 0.01, will boost the model performance. Table 9: Sensitivity analysis of sparsification threshold δt. δt 0 0.001 0.01 0.02 0.05 0.1 Acc 77.27 77.96 78.45 78.43 77.51 76.92 Published as a conference paper at ICLR 2023 103 46 240 184 137 2 9.3 (#103) + 9.5 (#240) = 18.8 Rooster Ibis Instance IR-Graph f 0.19 0.27 Ingredients 0.22 0.15 0.09 Category: Ibis 1.6 (#2) + 14.3 (#103) + 2.0 (#137) = 17.9 188 46 240 234 204 17 3.4 (#17) + 6.6 (#46) + 14.9 (#240) = 24.9 Llama Panda #46 0.47 0.51 0.23 f 0.42 f0.38 Category: Panda 2.5 (#17) + 15.8 (#204) + 3.0 (#234) = 21.3 Instance IR-Graph Ingredients Figure 10: Illustration of misclassified samples. Better view in color. H.2 LOSS WEIGHTS Figure 8(c) shows the sensitivity analysis of λv and λe in Equation (11) that constrains the complexity of IR-Atlas. We evaluate the top-1 accuracy performance on the CIFAR-100 dataset with Dei T-Tiny as the backbone. The curve of λv shows that neither dense nor extreme sparse IR-Atlas impairs the performance, while our method is more robust when changing λe. I MORE VISUALIZATIONS I.1 VISUALIZATION AND ANALYSIS OF MISCLASSIFIED EXAMPLES Figure 10 shows two misclassified examples along with the classification evidence. In Figure 10(a), the object is misclassified to a fine-grained category rooster because of a noise ingredient #240, which should be #103. Unfortunately, ingredient #240 is a crucial vertex in the rooster s IR-Graph, contributing about 9.5 absolute gains in the logit, leading to misclassification. Figure 10(b) shows a more complicated example. We can observe that instead of discretizing the appeared human face to the face ingredients, the backbone provides features closer to #17, which is more similar to the animal s face. Moreover, some part of the object s body is assigned to #240 rather than the panda s body (#204). Thus, it creates a remarkable pattern, i.e., the interaction between #17 and #240, which Published as a conference paper at ICLR 2023 is the crucial local structure in the llama s graph. As a result, #240 and its local structure contribute about 14.9 gains in the logit. Although our schema inference framework is capable of revealing why an image is misclassified by highlighting the key points, in future work, we must explore a more compatible feature extractor that can generate more robust local features. I.2 VISUALIZATION OF INGREDIENTS In Figure 11, we visualize the whole set of ingredients on Caltech-101 for visualizations in Figure 5 and Figures 12 to 14. We extract 256 ingredients on Caltech-101 for a better view. For each cluster center generated from k-means clustering, we select the top-40 tokens that are closest to it and show the corresponding image patch (50 50 in pixel, delineated in Appendix D). We can observe that image patches of the same cluster share the same semantics. I.3 VISUALIZATION OF IR-ATLAS AND INSTANCE IR-GRAPHS We provide more visualization examples on Caltech-101 dataset, shown in Figures 12 to 14. Published as a conference paper at ICLR 2023 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 Ingredient ID 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 Ingredient ID Figure 11: Ingredient visualization on Caltech-101 dataset. Please zoom in for a better view. Published as a conference paper at ICLR 2023 IR-Atlas Instance IR-Graphs 3 72 121 133 196 84 0.7 #48 0.58 77 82 164 223 244 97 #82 0.22 #85 0.21 17 46 144 179 196 48 #48 0.42 #46 53 72 111 162 178 107 0.038 0.038 72 98 162 193 214 134 0.043 0.083 0.27 #80 0.18 Figure 12: Examples of the learned IR-Atlas and instance IR-Graphs randomly sampled from five categories ( bass , bonsai , emu , euphonium , and grand piano ) on Caltech-101. Published as a conference paper at ICLR 2023 IR-Atlas Instance IR-Graphs 0 28 72 78 196 35 0.052 0.043 0.107 0.37 #72 0.25 0.033 0.096 0.038 2 17 103 137 207 36 0.28 #2 0.14 0.69 #48 0.52 17 48 150 179 188 144 #17 0.21 #150 0.34 #46 0.32 3 9 100 133 253 29 0.57 #29 0.22 #29 0.56 #9 0.53 17 46 150 179 240 48 0.26 #17 0.21 Figure 13: Examples of the learned IR-Atlas and instance IR-Graphs randomly sampled from five categories ( hawksbill , ibis , kangaroo , leopards , and llama ) on Caltech-101. Published as a conference paper at ICLR 2023 IR-Atlas Instance IR-Graphs 17 46 144 196 240 108 #46 0.26 #144 0.55 #46 0.55 #17 0.29 #108 17 103 140 196 240 111 0.052 0.048 #45 0.33 #17 0.11 0.044 0.044 15 63 97 133 146 75 #15 0.42 #75 0.22 0.066 0.031 3 15 75 133 229 38 0.38 #15 0.29 29 108 196 234 253 133 #3 0.47 #29 0.38 #97 0.27 #121 Figure 14: Examples of the learned IR-Atlas and instance IR-Graphs randomly sampled from five categories ( okapi , rooster , sunflower , water lilly , and wild cat ) on Caltech-101.