# crosstransformers_spatiallyaware_fewshot_transfer__6720eae8.pdf Cross Transformers: spatially-aware few-shot transfer Carl Doersch Ankush Gupta Andrew Zisserman Deep Mind, London VGG, Department of Engineering Science, University of Oxford Given new tasks with very little data such as new classes in a classification problem or a domain shift in the input performance of modern vision systems degrades remarkably quickly. In this work, we illustrate how the neural network representations which underpin modern vision systems are subject to supervision collapse, whereby they lose any information that is not necessary for performing the training task, including information that may be necessary for transfer to new tasks or domains. We then propose two methods to mitigate this problem. First, we employ self-supervised learning to encourage general-purpose features that transfer better. Second, we propose a novel Transformer based neural network architecture called Cross Transformers, which can take a small number of labeled images and an unlabeled query, find coarse spatial correspondence between the query and the labeled images, and then infer class membership by computing distances between spatially-corresponding features. The result is a classifier that is more robust to task and domain shift, which we demonstrate via state-of-the-art performance on Meta-Dataset, a recent dataset for evaluating transfer from Image Net to many other vision datasets. Code available at: https://github.com/google-research/ meta-dataset. 1 Introduction General-purpose vision systems must be adaptable. Home robots must be able to operate in new, unseen homes; photo-organizing software must recognize unseen objects (e.g., to find examples of my sixth-grade son s abstract art project ); industrial quality-assurance systems must spot defects in new products. Deep neural network representations can bring some visual knowledge from datasets like Image Net [68] to bear on different tasks beyond Image Net [15, 32, 62], but empirically, this requires a non-trivial amount of labeled data in the new task. With too little labeled data, or for a large change in distribution, such systems empirically perform poorly. Research on meta-learning directly benchmarks adaptability. At training time, the algorithm receives a large amount of data and accompanying supervision (e.g., labels). At test time, however, the algorithm receives a series of episodes, each of which consists of a small number of datapoints from a different distribution than the training set (e.g., a different domain or different classes). Only a subset of this data has the accompanying supervision (called the support set); the algorithm must make predictions about the rest (the query set). Meta-Dataset [86] is particularly relevant for vision, as the challenge is few-shot fine-grained image classification. The training data is a subset of Image Net classes. At test time, each episode either contains images from the other Image Net classes, or from one of nine other visually distinct fine-grained recognition datasets. The algorithm must rapidly adapt its representations to the new classes and domains. Simple centroid-based algorithms like Prototypical Nets [17, 76] are near state-of-the-art on Meta Dataset, achieving around 50% accuracy on the held-out Image Net classes in Meta-Dataset s validation set (chance is roughly 1 in 20). An equivalent classifier trained on those validation classes can achieve roughly 84% accuracy on the same challenge. What accounts for the enormous discrepancy 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Query Nearest Neighbors gila monster gila monster hognose snake gila monster gila monster gila monster night snake hognose snake gila monster knot letter opener letter opener screwdriver letter opener ballpoint fountain pen buckle letter opener letter opener bassoon hammer hammer microphone hammer screw hammer syringe shovel hook screw buckeye buckeye buckeye hip pomegranate sunglasses buckeye buckeye buckeye screw : in test split Figure 1: Illustration of supervision collapse with nearest neighbors. In each row, the leftmost image is a query taken from the Meta-Dataset Image Net test classes, and the rest are the top 9 nearest neighbors from both training and test support set classes, using the embedding learned by a Prototypical Net (training details in the extended version [1]). Images belonging to the test split are indicated by a near the bottom left corner; rest are from the training split. For a simple classifier to work well on these test classes, semantically similar images should have similar representations, and so we hope the nearest neighbors would come from the same or semantically similar classes. Instead, we observe that only 5% of matches for test-set queries are from the same class as the query. Furthermore, many matches are all from the same incorrect training class (highlighted in red). We see a knot is matched with several gila monsters (and other reptiles); a bassoon with letter openers (and pens); a screw with hammers; another screw with buckeyes. The errors within that wrong class often have widely different appearance: for example, the bottom-most screw is matched with single buckeyes and also a pile of buckeyes. One interpretation is that the network picks up on image patterns during training that allow images of each class to be tightly grouped in the feature space, minimizing other ways that the image might be similar to other classes in preparation for a confident classification. For out-of-domain samples, the network can then overemphasize a spurious image pattern that suggests membership in one training-set class. This is the consequence of supervision collapse, where image patterns that might help make the correct associations are lost. between performance on within-distribution samples and out-of-distribution samples? We hypothesize that because the neural network backbone of Prototypical Nets is designed for classification, they do just this: represent only an image s (training-set) class, and discard information that might help with out-of-distribution classes. Doing so minimizes the losses for many meta-learning algorithms, including Prototypical Nets. We call this problem supervision collapse, and illustrate it in Figure 1. Our first contribution is to explore using self-supervision to overcome supervision collapse. We employ Sim CLR [16], which learns embeddings that discriminate between every image in the dataset while maintaining invariance to transformations (e.g., cropping and color shifts), thus capturing more than just classes. However, rather than treat Sim CLR as an auxiliary loss, we reformulate Sim CLR as episodes that can be classified in the same manner as a training episode. Our second contribution is a novel architecture called Cross Transformers, which extends Transformers [88] to few-shot fine-grained classification. Our key insight is that objects and scenes are generally composed of smaller parts, with local appearance that may be similar to what has been seen at training time. The classical example of this is the centaur that appeared in several early papers on visual representation [12, 41, 89], where the parts from the human and horse composed the centaur. Cross Transformers operationalize this insight of (i) local part-based comparisons, and (ii) accounting for spatial alignment, resulting in a procedure for comparing images which is more agnostic to the underlying classes. In more detail, first a coarse alignment between geometric or functional parts in the queryand support-set images is established using attention as in Transformers. Then, given this alignment, distances between corresponding local features are computed to inform classification. We demonstrate this improves generalization to unseen classes and domains. In summary, our contributions in this paper are: (i) We improve the robustness of our local features with a self-supervised technique, modifying the state-of-the-art Sim CLR [16] algorithm. (ii) We propose the Cross Transformer, a network architecture that is spatially aware and performs few-shot classification using more local features, which improves transfer. Finally, (iii) we evaluate and ablate how the choices in these algorithms impact Meta-Dataset [86] performance, and demonstrate state-of-the-art results on nearly every dataset within it, often by large margins. 2 Related Work Few-shot image classification. Few-shot learning [26, 36, 47, 54] has recently been primarily addressed in the meta-learning framework [58, 72, 81], where a model learns an update rule for the parameters of a base-learner model [6, 7, 73] through a sequence of training episodes [80, 90]. The meta-learner either learns to produce new parameters directly from the new data [9, 34, 57, 63, 65, 73, 74], or learns to produce an update rule to iteratively optimize the base learner to fit the new data [2, 6, 8, 39, 64, 99]. [28, 52, 60] do not use any explicit meta-learner model, but instead unroll the base-learner gradient updates and optimize for model initializations which generalize well on novel tasks. Matching-based methods [29, 76, 79, 91] instead learn representations for similarity functions [11, 17, 18, 45, 83], in the hope that the similarities will generalize to new data. Cross Transformers fall in this category, and share much of their architecture with Prototypical Nets [76]. Attention for few-shot learning. Cross Transformers attend [4] individually over each class s support set to establish local correspondences, whereas Matching Networks [91] attend over the whole support set to point to matching instances. [55] extend this idea to larger contexts using temporally dilated convolutions [87]. In the limit, attention over long-term experiences accumulated in memories [44, 57, 70, 77] can augment more traditional learning. Correspondences for visual recognition. Cross Transformers perform classification by matching more local parts. Discriminative parts [5, 10, 23, 35, 43, 84] and visual words [75, 102] have a rich history, and have found applications in deformable-parts models [27], classification [71, 102], and retrieval [13, 85]. Part-based correspondences for recognition [105] have been particularly successful in fine-grained face retrieval and recognition [49, 97]. Cross Transformers establish soft correspondences between pixels in the queryand support-set images; such dense pairwise interactions [94] have recently proved useful for generative networks [101], semantic matching [67] and tracking [92]. [50] learns spatially dense classifiers for few-shot classification, but pools the spatial dimensions of the prototypes, and hence does not have a notion of correspondence. Self-supervised learning for few-shot. Our work on Sim CLR episodes inherits from a line of self-supervised learning research, which typically deal with transfer from pretext tasks to semantic ones and must therefore represent more than their training data [3, 14, 16, 22, 25, 31, 37, 48, 61, 103, 104]. Some recent works [30, 78] demonstrate that this can improve few-shot learning, although these use self-supervised auxiliary losses rather than integrating self-supervised instance discrimination [16, 25, 56, 82, 96] into episodic training. Also particularly relevant are methods that use self-supervision for correspondence [42, 51, 92], which may in future work improve the correspondences that Cross Transformers use. 3 Stopping Collapse: Sim CLR Episodes and Cross Transformers We take a two-pronged approach to dealing with the supervision collapse problem. Modern approaches to few-shot learning typically involve learning an embedding for each image, followed by a classifier that aggregates information across an episode s support set in order to classify the episode s queries. Our first step aims to use self-supervised learning to improve the embedding so it expresses information beyond the classes, in a way that is as algorithm-agnostic as possible. Once we have these embeddings, we build a classifier called a Cross Transformer. Cross Transformers use Prototypical Nets [76] as a blueprint, chosen due to their simplicity and strong performance; the main modification is to aggregate information in a spatially-aware way. We begin by reviewing Prototypical Nets, and then describe the two approaches. Prototypical Nets are episodic learners, which means training is performed on the same kind of episodes that will be presented at test time: a query set Q of images, and a support set S which can be partitioned into classes c {1, 2, . . . , C}: each Sc = {xc i}N i=1 is composed of N example images xc i. Prototypical Nets learn a distance function between the query and each subset Sc. Both the queryand support-set images are first encoded into a D-dimensional representation Φ(x), using a shared Conv Net Φ : RH W 3 7 RD, where H, W are the height and width respectively. Then a prototype tc RD for the class c is obtained by averaging the representations of the support set Sc, tc = 1 |Sc| P x Sc Φ(x). Finally, a distribution of classes is obtained using softmax over the distances between the query image and class prototypes: p(y = c|xq) = exp( d(Φ(xq),tc)) PC c =1 exp( d(Φ(xq),tc )). In practice, the distance function d is fixed to be the squared Euclidean distance d(xq, Sc) = ||Φ(xq) tc||2 2. The learning objective is to train the embedding network Φ to maximize the probability of the correct class for each query. 3.1 Self-supervised training with Sim CLR Our first challenge is to improve the neural network embedding Φ: after all, if these features have collapsed to represent little information beyond the classes, then a subsequent classifier cannot can recover this information. But how can we train features to represent things beyond labels when our only supervision is the labels? Our solution is self-supervised learning, which invents pretext tasks that train representations without labels [22, 25], and better yet, has a reputation for representations that transfer beyond this pretext task. Specifically we use Sim CLR [16], which uses instance discrimination as a pretext task. It works by applying random image transformations (e.g., cropping or color shifts) twice to the same image, thus generating two views of that image. Then it trains the network so that representations of the two views of the same image are more similar to each other than they are to those of different images. Empirically, networks trained in this way become sensitive to semantic information, but also learn to discriminate between different images within a single class, which is useful for combating supervision collapse. While we could treat Sim CLR as an auxiliary loss on the embedding, we instead reformulate Sim CLR as episodic learning, so that the technique can be applied to all episodic learners with minimal hyper-parameters. To do this, we randomly convert 50% of the training episodes into what we call Sim CLR episodes, by treating every image as its own class. For clarity, we will call the original episodes that have not been converted Sim CLR episodes MD-categorization episodes, to emphasize that they use the original categories from Meta-Dataset. Specifically, let ρ( ) be Sim CLR s (random) image transformation function, and let S = {xi}|S| i=1 be a training support set. We generate a Sim CLR episode by sampling a new support set, transforming each image in the original support set S = {ρ(xi)}|S| i=1, and then generating query images by sampling other transformations from the same support set: Q = {ρ(random_sample(S))}|Q| i=1, where random_sample just takes a random image from the set.1 The original query set Q is discarded. The label for an image in the Sim CLR episode is its index in the original support set, resulting in an |S|-way classification for each query. Note that for a Sim CLR episode, the prototypes in Prototypical Nets average over just a single image, and therefore the Prototypical Net loss can be written as exp( d(Φ(ρ(xq)),Φ(ρ(xq)))) Pn i=1 exp( d(Φ(ρ(xq)),Φ(ρ(xi))). If we define d as the cosine distance rather than Euclidean, this loss is identical to the one used in Sim CLR. 3.2 Cross Transformers Given a query image xq and a support set Sc = {xc i}N i=1 for the class c, Cross Transformers aim to build a representation which enables local part-based comparisons between them. Cross Transformers start by making the image representation a spatial tensor, and then assemble query-aligned class prototypes by putting the support-set images Sc in correspondence with the query image. The distance between the query image and the query-aligned prototype for each class is then computed and used in a similar way to Prototypical Nets. In practice, we establish soft correspondences using attention [4] based Transformers [88]. In contrast, Prototypical Nets use flat vector representations which lose the location of image features, and have a fixed class prototype which is independent of the query image. 1We enforce that the sampled queries have the same class distribution as Q, and have no repeats. Dot-product query features at location p against all support set (Sc) spatial features in class Softmax across all spatial features in class Key Heads Value Heads Query-aligned prototype (tc) weighted sum Softmax normalized attention weights Support-set (Sc) image features for the category c Query image features a2 c a1 c ~a2 c ~a1 c Keys (kc) Values (vc) Queries (q) Figure 2: Cross Transformers. Construction of query-aligned class prototype vector tc p for the class c and the query image xq, focusing on the spatial location p in xq. The query vector qp is compared against keys kc from all spatial locations in the support set Sc to obtain attention scores ac, which are softmax normalized before being used to aggregate the values vc for the aligned prototype vector tc p. Concretely, Cross Transformers remove the final spatial pooling in Prototypical Nets embedding network Φ( ), such that the spatial dimensions H , W are preserved: Φ(x) RH W D. Following Transformers, key-value pairs are then generated for each image in the support set using two independent linear maps: the key-head Γ : RD 7 Rdk, and the value-head Λ : RD 7 Rdv respectively. Similarly, the query image features Φ(xq) are embedded using the query-head Ω: RD 7 Rdk. Dot-product attention scores are then obtained between keys and queries, followed by softmax normalization across all the images and locations in Sc. This attention serves as our coarse correspondence (see example attention visualizations in Figure 3 and the extended version [1]), and is used to aggregate the support-set features into alignment with the query. This process is visualized in Figure 2. Mathematically, let kc jm = Γ Φ(xc j)m be the key for the jth image in the support set for class c at spatial position m (index over the two dimensions H , W ), and similarly let qp = Ω Φ(xq)p be the query vector at spatial position p in the query image xq. The attention ac jmp R between the two is then obtained as: ac jmp = exp(ac jmp/τ) P i,n exp(ac inp/τ), where ac jmp = kc jm qp, and τ = p Next, the aligned prototype vector tc p corresponding to spatial location p in the query is obtained by aggregating the support-set values vc jm = Λ Φ(xc j)m using the attention weights above: jm ac jmpvc jm (2) Finally, squared Euclidean distances between aligned local features from the above prototype and corresponding query image values wp = Λ Φ(xq)p are aggregated as below. This scalar distance acts as a negative logit for a distribution over classes as in Prototypical Nets. d(xq, Sc) = 1 H W X p ||tc p wp||2 2 (3) Note we apply the same value-head Λ to both the query and support-set images. This ensures that the Cross Transformer behaves somewhat like a distance. That is, imagine a trivial case where, for one class, all images in Sc are identical to xq. We would want d(xq, Sc) to approach 0 even if the network is untrained, or if these images are highly dissimilar from those used for training. Sharing Λ between the support and query sets helps accomplish this: in fact, if ac jmp is 1 where p = m and 0 elsewhere for all j, then d(xq, Sc) will be identically 0 under this architecture, no matter the network Query Correspondence in support set Query Correspondence in support set Figure 3: Visualization of the attention a. We show four query images, along with three support-set images for each. Within each query image, we choose three spatial locations (red, green, and blue squares), and plot the Cross Transformer attention weights for each one in the corresponding color (brighter colors mean higher weight). The four examples are from Aircraft, CU-Birds, VGG Flowers, and Im Net test sets respectively (clockwise, starting from top-left). No matter which dataset, the attention masks are semantically relevant, even when the correspondence is not one-to-one. More visualizations are given in the extended version [1]. weights. To encourage this behavior for the attention a, we also set Γ = Ω, i.e., the key and query heads are the same. This way, in our trivial case, the attention is likely to be maximal for spatial locations that correspond, because kc jm and qp will be the same for p = m. For one experiment, we also augment the Cross Transformer with a global feature, which can help for some datasets like DTD (Describable Textures Dataset) with less spatial structure. 4 Experiments We evaluate on Meta-Dataset [86], specifically the setting where the training is performed on the Image Net train split only, which is 712 classes (plus 158 classes for validation, which are not used for training but only to perform early stopping). We then test on the remaining 130 held-out classes from Image Net, as well as 9 other image datasets. Note that this is in contrast to another popular (and easier) setting, where the training also uses a subset of categories from more of these datasets: usually all datasets except Traffic Signs and COCO. For clarity, we ll use Meta-Dataset Train-on-ILSVRC to denote training on Image Net only, and Meta-Dataset Train-on-all to denote when training occurs on more datasets. Test time consists of a series of episodes, each of which contains: (1) a support set between 50 and 500 labeled images which come from between 5 and 50 classes; and (2) an unlabeled query set with 10 images per class. Meta-Dataset aims for fine-grained recognition, so the classes in each episode are mutually similar: one episode may contain only musical instruments, another may contain only birds, etc. Meta-Dataset is useful for studying transfer because different test datasets encapsulate different kinds of transfer challenges. For test datasets like CU-Birds [93], there are numerous similar classes in Image Net train (20 bird classes in Image Net train, versus 100 in the CU-birds test dataset). In contrast, for test datasets like Aircraft [53], there is just a single corresponding class in Image Net train; therefore, algorithms which don t represent the intra-class variability for this class will be penalized. The Image Net test set has images in a similar domain to the Image Net train set but with different classes, while test datasets like COCO contain many similar classes to Image Net, but with domain shift (in COCO, instances are generally not the subject of their photographs, and may be low-resolution or occluded). Finally, test datasets like Omni Glot combine these challenges, i.e., different classes in a substantially different domain. 4.1 Implementation details To ensure comparability, we followed the public implementation of Prototypical Nets for Meta Dataset [86] wherever possible. This includes using the same hyperparameters, unless otherwise noted. For the hyperparameters that were chosen with a sweep on the validation set (learning rate schedule and weight decay), we simply used the best values discovered for Prototypical Nets for all the experiments in this paper. See the extended version [1] for details of the Cross Transformer architecture. We use no pretraining for Cross Transformers, although to be consistent with prior work [86] we use it for the experiments involving Prototypical Nets. We incorporate two improvements from Meta-Baseline [17], which at test time is similar to Prototypical Nets (though it isn t trained as an episodic learner). The first is to keep exponential moving averages for Batch Norm statistics during training, and use those for Batch Norm at test time. Second, we note that Meta-Baseline does not train on fine-grained episodes sampled from the Image Net hierarchy, as Prototypical Nets does, but rather on batches with uniformly-sampled classes. Empirically, Prototypical Nets trained only on fine-grained episodes struggle to do coarse-grained recognition, as required for datasets like COCO. Therefore, we only use the Image Net hierarchy to make 50% of episodes fine-grained; the rest have categories sampled uniformly. Choice of network. Prior implementations of networks like Prototypical Nets use relatively small networks (e.g., Res Net-18) with small input images (e.g. 126 126 pixels), and report that measures to increase capacity (e.g., Wide Res Nets [100]) provide minimal benefits. This is surprising given that standard networks show improvements for increasing capacity (e.g., Res Net-34 outperforms Res Net18 by 3% on Image Net [38]). Making our networks spatially-aware requires higher-resolution, and also higher-capacity networks are especially important in self-supervised learning [24, 46]. Therefore, our experiments increase resolution to the standard 224 224 and use Res Net-34, and we also use normalized stochastic gradient descent [19, 59], which we found improved stability when fine-tuning more complex networks. Table 1 compares the Prototypical Nets performance of this network to that of using a Res Net-18. Increased capacity leads to only slight performance improvements, which are more pronounced for datasets that are similar to Image Net; it harms, e.g., Omni Glot. Further details in the extended version [1]. For experiments with Cross Transformers, we also increased the resolution of the convolutional feature map by setting the stride of final block of the Res Net to 1, and using dilated convolutions to preserve the feature computation [33, 40]. This turns the usual 7 7 feature map for a 224 224 image into a 14 14 feature map. We ablate this choice in the extended version [1]. Augmenting CTX with a global feature. Recent works have also shown benefits for applying logistic regression (LR) at test time [83]. In practice, it is too expensive to apply LR to our queryaligned prototypes (as this would involve a separate classifier for every query). Therefore, we instead apply logistic regression to a globally-pooled feature and average the logits with those produced by the Cross Transformer. See the extended version [1] for details. Augmentation. While most experiments use no augmentation (apart from Sim CLR episodes) to be consistent with prior work [86], more recent work [17, 69, 83] showed that stronger data augmentation is effective. Therefore, for two experiments, we employ augmentation using the settings discovered in BOHB [69] (via Auto-Augment [20] on the validation set), with an extra stage that randomly downsamples and then upsamples images, which we find helpful as our network operates at higher resolution than many of the test datasets. This BOHB augmentation is only applied to the MDcategorization episodes, and not to the Sim CLR episodes. Note this BOHB augmentation is different from Sim CLR-style augmentation, which is used in Sim CLR Episodes as well as in the ablation (SC-Aug) in Table 1. See the extended version [1] for details. 4.2 Results for self-supervised learning with Sim CLR on Prototypical Nets We first analyze the impact of Sim CLR Episodes and other architectural choices in Table 1. For baseline Prototypical Nets, Sim CLR Episodes generally improve performance, but this depends on architectural choices. Improvements are largest for datasets that are more distant from Image Net, e.g., Omni Glot and Quickdraw, and datasets which require distinguishing between sub-categories Image Net categories, e.g., Aircraft and Traffic Signs. In Image Net, all commercial airplanes fall in a single Image Net class; therefore, the success of Sim CLR Episodes here suggests they recover features which are lost due to supervision collapse. Strangely, however, Sim CLR Episodes interact with Batch Norm: we find more robust improvements when computing Batch Norm statistics from the test-time support set, but not when using exponential moving averages (EMA) as suggested by [17]. One possible interpretation is that the network has learned to use Batch Norm to communicate information across the batch: e.g., to distinguish between Sim CLR Episodes and MD-categorization episodes. Using EMA at test time may prevent this, which may confuse the network. Interestingly, Table 1: Effects of architecture and Sim CLR Episodes on Prototypical Nets, for Meta-Dataset Train-on-ILSVRC. We ablate architectural choices: use of Exponential Moving Averages (EMA) at test time for Batch Norm (versus computing Batch Norm statistics on the support set at test time), image resolution (224, versus the baseline s 126), Res Net-34 (R34) replacing Res Net-18, Sim CLRstyle augmentation (SC-Aug), and the addition of 50% Sim CLR Episodes (SC-Eps). The test datasets from Meta-Dataset are Im Net: Meta-Dataset s Image Net Test classes; Omni: Omni Glot drawn characters; Acraft: Aircraft; Bird: CU-Birds; DTD: Textures; QDraw: Quick Draw drawings; Fungi: FGVCx fungi challenge; Flower: VGG Flowers; COCO: Microsoft COCO cropped objects. The best number in each column is bolded, along with others that are within a confidence interval [86]. Rank is the average rank for each method. Using Sim CLR Episodes provides improvements on almost all datasets, and provides especially large boosts for datasets which are dissimilar from Image Net, such as Omni Glot. However, simply using Sim CLR transformations without instance discrimination (SC-Aug) harms results on almost all datasets. Increased capacity provides small benefits on some datasets, especially the more realistic and Image Net-like datasets (e.g., birds), but actually harm others like Omni Glot. Note that in this table, Quick Draw uses the split from the original paper [86] rather than the (somewhat easier) split published for that paper s public benchmark. For all other tables, we use the split from the published benchmark. 224 R34 SC-Aug SC-Eps EMA Im Net Omni Acraft Bird DTD QDraw Fungi Flower Sign COCO Rank 49.10 59.27 49.31 68.43 66.70 45.83 38.48 85.34 49.49 42.88 5.55 49.77 55.70 52.06 68.58 67.27 49.86 37.68 84.32 50.27 41.92 5.20 51.66 57.22 51.63 71.73 69.72 47.31 42.07 87.29 47.45 44.38 4.35 52.51 49.87 56.47 72.81 68.45 51.41 42.16 87.92 54.40 40.60 3.30 47.58 55.73 46.93 57.75 54.88 42.91 37.42 83.82 46.88 43.36 7.55 47.94 51.79 54.58 62.84 58.64 46.36 36.06 76.88 48.35 38.77 7.45 49.67 65.21 54.46 60.94 63.96 50.64 37.84 88.70 51.61 42.97 4.35 53.69 68.50 58.04 74.07 68.76 53.30 40.73 86.96 58.11 41.70 1.90 Proto Nets [86] 50.50 59.98 53.10 68.79 66.56 48.96 39.71 85.27 47.12 41.00 5.35 we will show later that Sim CLR Episodes don t harm Cross Transformers as they harm Prototypical Nets when using EMA at test time, suggesting the two architectures solve the problem differently. Recall that converting an MD-categorization episode into a Sim CLR episode makes two changes to the episode: it 1) applies data augmentation, and 2) converts the classification problem to instance discrimination, by selecting images from the support set as a new query set, and requiring the network to predict the selected indices. To ensure that we are not simply seeing the effect of data augmentation, we also implemented a baseline (SC-Aug) that does 1 but not 2 to the input MDcategorization episodes, and does this augmentation for all episodes (rather than 50%, which is the fraction of MD-categorization episodes that are converted to Sim CLR episodes for SC-Eps experiments). Indeed, we see no improvements for this change, and in fact non-trivial performance loss from this, mirroring the result for supervised learning in the original paper [16]. This reinforces that Sim CLR was designed for self-supervised learning, and so the transformations are more severe than is usually optimal for supervised learning. Finally, we see small improvements from using larger networks and higher resolution for the baseline model. While our baseline is overall better than the baseline Prototypical Nets implementation [86], it is still below the state-of-the-art for centroid-based methods which rely more heavily on pretraining, and use no episodic training [17]. 4.3 Cross Transformers results Given these performant features, we next turn to Cross Transformers. Table 2 compares Cross Transformers (CTX) with and without Sim CLR episodes to several state-of-the-art methods, including the Prototypical Nets on which they are based. We see that Cross Transformers provide strong performance on their own, including having a better average rank than all baselines. With Sim CLR episodes providing more versatile features, we see further improvements, with performance on-par or better than the best methods on almost every dataset. We note particularly large improvements on Omni Glot, which has a large domain gap relative to the training data. We also see strong improvements on Street Signs, Aircraft, and Flowers, where multiple test-time categories map to few training-time categories, and often exhibit well-defined spatial correspondence. Table 2: Cross Transformers (CTX) comparison to state-of-the-art. We compare four versions of Cross Transformers to several state-of-the-art methods, which are the best performers among those evaluated for Meta-Dataset Train-on-ILSVRC. We see that CTX alone has better average rank than any baseline. Adding Sim CLR episodes (+Sim CLR Eps) and data augmentation inspired by BOHB [69] (+Aug) further improves results. Our full model is on-par or above prior methods on all but one dataset, sometimes with large gaps over the best baseline (e.g., +5% on Omni Glot, +13% on Aircraft, +5% on Signs), and furthermore, each prior method has some datasets where we outperform by a larger margin (the next best average rank [83], performs 19% worse on Aircraft and 17% worse on Omni Glot). Finally, adding a test-time Logistic Regression classifier inspired by Tian et al. [83] improves performance on the one dataset DTD textures that was otherwise lacking. Note that most of these methods [17, 69, 83] are unpublished concurrent work. Im Net Omni Acraft Bird DTD QDraw Fungi Flower Sign COCO Rank Finetuning [86] 45.78 60.85 68.69 57.31 69.05 42.60 38.20 85.51 66.79 34.86 12.20 Proto Nets [86] 50.50 59.98 53.10 68.79 66.56 48.96 39.71 85.27 47.12 41.00 12.65 Proto Nets+MAML [86] 49.53 63.37 55.95 68.66 66.49 51.52 39.96 87.15 48.83 43.74 11.55 CNAPS [66] 50.60 45.20 36.00 60.70 67.50 42.30 30.10 70.70 53.30 45.20 13.55 BOHB-L [69] 50.60 64.09 57.36 67.68 70.38 46.26 33.82 85.51 55.17 41.58 11.50 BOHB-NC [69] 51.92 67.57 54.12 70.69 68.34 50.33 41.38 87.34 51.80 48.03 10.15 BOHB-NC Ensemble [69] 55.39 77.45 60.85 73.56 72.86 61.16 44.54 90.62 57.53 51.86 7.45 Dhillon et al. [21] - - 68.69 74.26 77.35 - - 88.14 55.98 40.62 - Meta-Baseline [17] 59.20 69.10 54.10 77.30 76.00 57.30 45.40 89.60 66.20 55.70 7.20 Tian et al. LR [83] 60.14 64.92 63.12 77.69 78.59 62.48 47.12 91.60 77.51 57.00 5.50 Tian et al. LR-distill [83] 61.58 64.31 62.32 79.47 79.28 60.83 48.53 91.00 76.33 59.28 4.60 Proto Nets (Our implementation) 51.66 57.22 51.63 71.73 69.72 53.81 42.07 87.29 47.45 44.38 11.10 CTX 61.94 76.52 79.65 84.06 76.26 65.67 52.53 94.11 70.47 53.51 3.85 CTX+Sim CLR Eps 63.79 80.83 82.05 82.01 75.76 68.84 52.01 94.62 75.01 52.76 3.05 CTX+Sim CLR Eps+Aug 62.76 82.21 79.49 80.63 75.57 72.68 51.58 95.34 82.65 59.90 2.25 CTX+Sim CLR Eps+Aug+LR 62.25 82.03 77.41 76.66 80.29 72.24 49.39 93.05 75.25 60.35 3.40 DTD, however, is more challenging for basic CTX, which is unsurprising since textures have little of the kind of spatial correspondence that CTX attempts to find. COCO is also challenging, likely due to its extremely large intra-class variation (e.g., occlusion) and the fact that many categories overlap with Image Net-train categories, meaning that simply memorizing categories from the training set may be more useful than using test-time appearance. To explore this trade-off, we applied logistic regression at test time to a globally pooled feature (see the extended version [1]), which provides additional logits that are averaged with the CTX logits. We see non-trivial improvements on DTD by using this, although we sacrifice some performance on other datasets, such as Signs and Aircraft. This implies that there s a fundamental tension between learning categories based on global features, and decomposing the task into local features. More work is needed to better combine these two ideas. Finally, Figure 3 depicts the correspondence inferred by the Cross Transformer. The attention is often semantically meaningful: object parts are well matched, including heads, bodies, feet, engines, and strings. The attention is often not one-to-one either: for the flower, the single query flower is matched to multiple flowers in some of the support images. Furthermore, the matching works even when the fine-grained classes are not the same, such as the different species of birds, suggesting that the attention is indeed a coarse-grained matching that has not overfit to the training-set classes. 5 Conclusion Within a single domain, deep networks have a remarkable ability to compose and reuse features in order to achieve statistical efficiency. However, this work shows the hidden problem with such systems: the networks compose features in a way that conflates images which have different appearance but the same label, i.e., it loses information about intra-class variation that may be necessary to understand novel classes. We propose two techniques that help resolve this problem: self-supervised learning, which prevents features from losing that intra-class variation, and Cross Transformers, which help neural networks classify images using local features that are more likely to generalize. However, this problem is far from resolved. In particular, our algorithm provides less benefit when less spatial structure is available, when knowledge of train-time categories can be useful (as in, e.g., COCO), or when higher-level reasoning is required (e.g., finding conjunctions of multiple objects). Allowing this algorithm to use spatial structure only where relevant remains an open problem. Broader Impact The algorithm presented in this paper most directly applies to few-shot recognition, which has numerous uses in industry, including vision systems for robotics that must adapt to new objects, and photo-organizing software which must infer the presence of new classes of objects on-the-fly. Unfamiliar objects are ubiquitous in many real-world vision applications due to the so-called long tail [95] of objects that occur in real scenes, and therefore we expect our algorithm to improve the robustness of visual recognition systems. While our current work only addresses classification, many other tasks in computer vision, such as object detection and segmentation, use neural network representations that can likewise be made more robust using the kind of architectures presented here. Our algorithm attempts to build representations which factorize the object recognition problem into sub-problems (feature correspondence and feature comparison) that will each transfer correctly to new datasets. We hope that further research in this direction may help address dataset biases, including biases regarding race, gender, or other attributes [98], by helping to disentangle the truly meaningful traits from the spurious correlations. Finally, while this algorithm presents an advance to state-of-the-art in understanding rare objects, the general performance of such systems is still far below human performance. For safety-critical applications (e.g., surgery or self-driving cars), relying on the ability of vision systems to correctly interpret unusual situations is risky with current systems, even with the advances presented here. Funding Disclosure This work was funded by Deep Mind. Acknowledgments The authors would like to thank Pascal Lamblin for help with Meta-Dataset, Olivier Hénaff for help with Sim CLR, Yonglong Tian for help in reproducing baselines, and Relja Arandjelovi c for invaluable advice on the paper. They are also grateful to Jean-Baptiste Alayrac, Joao Carreira, Mateusz Malinowski, Viorica P atr aucean, Adria Recasens, and Lucas Smaira for helpful discussions, support, and feedback on the project. [1] Supplementary material (appendix) for the paper. http://arxiv.org/abs/2007.11498. [2] M. Andrychowicz, M. Denil, S. Gomez, M. W. Hoffman, D. Pfau, T. Schaul, B. Shillingford, and N. De Freitas. Learning to learn by gradient descent by gradient descent. In Neur IPS, 2016. [3] P. Bachman, R. D. Hjelm, and W. Buchwalter. Learning representations by maximizing mutual information across views. In Neur IPS, 2019. [4] D. Bahdanau, K. Cho, and Y. Bengio. Neural machine translation by jointly learning to align and translate. In Proc. ICLR, 2015. [5] E. Bart and S. Ullman. Cross-generalization: Learning novel classes from a single example by feature replacement. In Proc. CVPR, 2005. [6] S. Bengio, Y. Bengio, J. Cloutier, and J. Gecsei. On the optimization of a synaptic learning rule. In Preprints Conf. Optimality in Artificial and Biological Neural Networks, volume 2. Univ. of Texas, 1992. [7] Y. Bengio, S. Bengio, and J. Cloutier. Learning a synaptic learning rule. Citeseer, 1990. [8] L. Bertinetto, J. F. Henriques, P. H. Torr, and A. Vedaldi. Meta-learning with differentiable closed-form solvers. In Proc. ICLR, 2019. [9] L. Bertinetto, J. F. Henriques, J. Valmadre, P. Torr, and A. Vedaldi. Learning feed-forward one-shot learners. In Neur IPS, 2016. [10] L. Bourdev and J. Malik. Poselets: Body part detectors trained using 3d human pose annotations. In Proc. ICCV, 2009. [11] J. Bromley, I. Guyon, Y. Le Cun, E. Säckinger, and R. Shah. Signature verification using a" siamese" time delay neural network. In Neur IPS, 1994. [12] A. M. Bronstein, M. M. Bronstein, A. M. Bruckstein, and R. Kimmel. Partial similarity of objects, or how to compare a centaur to a horse. Proc. ICCV, 2009. [13] B. Cao, A. Araujo, and J. Sim. Unifying deep local and global features for image search. In Proc. ECCV, 2020. [14] M. Caron, P. Bojanowski, A. Joulin, and M. Douze. Deep clustering for unsupervised learning of visual features. In Proc. ECCV, 2018. [15] K. Chatfield, K. Simonyan, and A. Zisserman. Efficient on-the-fly category retrieval using convnets and gpus. In Asian Conference on Computer Vision, 2014. [16] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton. A simple framework for contrastive learning of visual representations. In Proc. ICML, 2020. [17] Y. Chen, X. Wang, Z. Liu, H. Xu, and T. Darrell. A new meta-baseline for few-shot learning. ar Xiv preprint ar Xiv:2003.04390, 2020. [18] S. Chopra, R. Hadsell, and Y. Le Cun. Learning a similarity metric discriminatively, with application to face verification. In Proc. CVPR, 2005. [19] J. Cortés. Finite-time convergent gradient flows with applications to network consensus. Automatica, 42(11), 2006. [20] E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q. V. Le. Autoaugment: Learning augmentation strategies from data. In Proc. CVPR, 2019. [21] G. S. Dhillon et al. A baseline for few-shot image classification. Proc. ICLR, 2020. [22] C. Doersch, A. Gupta, and A. A. Efros. Unsupervised visual representation learning by context prediction. In Proc. ICCV, 2015. [23] C. Doersch, S. Singh, A. Gupta, J. Sivic, and A. A. Efros. What makes paris look like paris? Proc. ACM SIGGRAPH, 31(4), 2012. [24] C. Doersch and A. Zisserman. Multi-task self-supervised visual learning. In Proceedings of the IEEE International Conference on Computer Vision, pages 2051 2060, 2017. [25] A. Dosovitskiy, J. T. Springenberg, M. Riedmiller, and T. Brox. Discriminative unsupervised feature learning with convolutional neural networks. In Neur IPS. 2014. [26] L. Fei-Fei, R. Fergus, and P. Perona. One-shot learning of object categories. IEEE PAMI, 2006. [27] P. F. Felzenszwalb, R. B. Girshick, D. Mc Allester, and D. Ramanan. Object detection with discriminatively trained part-based models. IEEE PAMI, 32(9), 2009. [28] C. Finn, P. Abbeel, and S. Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In Proc. ICML, 2017. [29] V. Garcia and J. Bruna. Few-shot learning with graph neural networks. In Proc. ICLR, 2018. [30] S. Gidaris, A. Bursuc, N. Komodakis, P. Perez, and M. Cord. Boosting few-shot visual learning with self-supervision. In The IEEE International Conference on Computer Vision (ICCV), October 2019. [31] S. Gidaris, P. Singh, and N. Komodakis. Unsupervised representation learning by predicting image rotations. In Proc. ICLR, 2018. [32] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proc. CVPR, 2014. [33] A. Giusti, D. C. Cire san, J. Masci, L. M. Gambardella, and J. Schmidhuber. Fast image scanning with deep max-pooling convolutional neural networks. In Intl. Conf. Image Proc., 2013. [34] D. Ha, A. Dai, and Q. V. Le. Hypernetworks. ar Xiv preprint ar Xiv:1609.09106, 2016. [35] X. Han, T. Leung, Y. Jia, R. Sukthankar, and A. C. Berg. Matchnet: Unifying feature and metric learning for patch-based matching. In Proc. CVPR, 2015. [36] B. Hariharan and R. Girshick. Low-shot visual recognition by shrinking and hallucinating features. In Proc. CVPR, 2017. [37] K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick. Momentum contrast for unsupervised visual representation learning. In Proc. CVPR, 2020. [38] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proc. CVPR, 2016. [39] S. Hochreiter, A. S. Younger, and P. R. Conwell. Learning to learn using gradient descent. In International Conference on Artificial Neural Networks. Springer, 2001. [40] M. Holschneider, R. Kronland-Martinet, J. Morlet, and P. Tchamitchian. A real-time algorithm for signal analysis with the help of the wavelet transform. In Wavelets, pages 286 297. 1990. [41] D. Jacobs, D. Weinshall, and Y. Gdalyahu. Class representation and image retrieval with non-metric distances. IEEE PAMI, 22(6):583 600, 2000. [42] J. Y. Jason, A. W. Harley, and K. G. Derpanis. Back to basics: Unsupervised learning of optical flow via brightness constancy and motion smoothness. In Proc. CVPR, 2016. [43] M. Juneja, A. Vedaldi, C. Jawahar, and A. Zisserman. Blocks that shout: Distinctive parts for scene classification. In Proc. CVPR, 2013. [44] Ł. Kaiser, O. Nachum, A. Roy, and S. Bengio. Learning to remember rare events. In Proc. ICLR, 2017. [45] G. Koch, R. Zemel, and R. Salakhutdinov. Siamese neural networks for one-shot image recognition. In ICML deep learning workshop, volume 2. Lille, 2015. [46] A. Kolesnikov, X. Zhai, and L. Beyer. Revisiting self-supervised visual representation learning. In Proc. CVPR, 2019. [47] B. M. Lake, R. Salakhutdinov, and J. B. Tenenbaum. Human-level concept learning through probabilistic program induction. Science, 350, 2015. [48] G. Larsson, M. Maire, and G. Shakhnarovich. Learning representations for automatic colorization. In Proc. ECCV, 2016. [49] H. Li, G. Hua, Z. Lin, J. Brandt, and J. Yang. Probabilistic elastic matching for pose variant face verification. In Proc. CVPR, 2013. [50] Y. Lifchitz, Y. Avrithis, S. Picard, and A. Bursuc. Dense classification and implanting for few-shot learning. In Proc. CVPR, 2019. [51] P. Liu, M. Lyu, I. King, and J. Xu. Selflow: Self-supervised learning of optical flow. In Proc. CVPR, 2019. [52] D. Maclaurin, D. Duvenaud, and R. Adams. Gradient-based hyperparameter optimization through reversible learning. In International Conference on Machine Learning, 2015. [53] S. Maji, E. Rahtu, J. Kannala, M. Blaschko, and A. Vedaldi. Fine-grained visual classification of aircraft. ar Xiv preprint ar Xiv:1306.5151, 2013. [54] E. G. Miller, N. E. Matsakis, and P. A. Viola. Learning from one example through shared densities on transforms. In Proc. CVPR, 2000. [55] N. Mishra, M. Rohaninejad, X. Chen, and P. Abbeel. A simple neural attentive meta-learner. In Proc. ICLR, 2017. [56] I. Misra and L. v. d. Maaten. Self-supervised learning of pretext-invariant representations. In Proc. CVPR, 2020. [57] T. Munkhdalai and H. Yu. Meta networks. In Proc. ICML, 2017. [58] D. K. Naik and R. J. Mammone. Meta-neural networks that learn by learning. In [Proceedings 1992] IJCNN International Joint Conference on Neural Networks, volume 1. IEEE, 1992. [59] Y. E. Nesterov. Minimization methods for nonsmooth convex and quasiconvex functions. Matekon, 29, 1984. [60] J. Nichol, Alex any Andrychowicz ed Achiam and J. Schulman. On first-order meta-learning algorithms. ar Xiv preprint ar Xiv:1803.02999, 2018. [61] M. Noroozi and P. Favaro. Unsupervised learning of visual representations by solving jigsaw puzzles. In Proc. ECCV, 2016. [62] M. Oquab, L. Bottou, I. Laptev, and J. Sivic. Learning and transferring mid-level image representations using convolutional neural networks. In Proc. CVPR, 2014. [63] E. Perez, F. Strub, H. De Vries, V. Dumoulin, and A. Courville. Film: Visual reasoning with a general conditioning layer. In Proc. AAAI, 2018. [64] S. Ravi and H. Larochelle. Optimization as a model for few-shot learning. In Proc. ICLR, 2017. [65] S.-A. Rebuffi, H. Bilen, and A. Vedaldi. Learning multiple visual domains with residual adapters. In Neur IPS, 2017. [66] J. Requeima, J. Gordon, J. Bronskill, S. Nowozin, and R. E. Turner. Fast and flexible multi-task classification using conditional neural adaptive processes. In Neur IPS, 2019. [67] I. Rocco, M. Cimpoi, R. Arandjelovi c, A. Torii, T. Pajdla, and J. Sivic. Neighbourhood consensus networks. In Neur IPS, pages 1651 1662, 2018. [68] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, et al. Imagenet large scale visual recognition challenge. IJCV, 115(3), 2015. [69] T. Saikia, T. Brox, and C. Schmid. Optimized generic feature learning for few-shot classification across domains. ar Xiv preprint ar Xiv:2001.07926, 2020. [70] A. Santoro, S. Bartunov, M. Botvinick, D. Wierstra, and T. Lillicrap. Meta-learning with memoryaugmented neural networks. In Proc. ICML, 2016. [71] S. Savarese, J. Winn, and A. Criminisi. Discriminative object class models of appearance and shape by correlatons. In Proc. CVPR, 2006. [72] J. Schmidhuber. Evolutionary principles in self-referential learning, or on learning how to learn: the meta-meta-... hook. Ph D thesis, Technische Universität München, 1987. [73] J. Schmidhuber. Learning to control fast-weight memories: An alternative to dynamic recurrent networks. Neural Computation, 4(1):131 139, 1992. [74] J. Schmidhuber. A neural network that embeds its own meta-levels. In IEEE International Conference on Neural Networks, pages 407 412. IEEE, 1993. [75] J. Sivic and A. Zisserman. Video Google: A text retrieval approach to object matching in videos. In Proc. ICCV, 2003. [76] J. Snell, K. Swersky, and R. Zemel. Prototypical networks for few-shot learning. In Neur IPS, 2017. [77] P. Sprechmann, S. M. Jayakumar, J. W. Rae, A. Pritzel, A. P. Badia, B. Uria, O. Vinyals, D. Hassabis, R. Pascanu, and C. Blundell. Memory-based parameter adaptation. In Proc. ICLR, 2018. [78] J.-C. Su, S. Maji, and B. Hariharan. When does self-supervision improve few-shot learning? In Proc. ECCV, 2020. [79] F. Sung, Y. Yang, L. Zhang, T. Xiang, P. H. Torr, and T. M. Hospedales. Learning to compare: Relation network for few-shot learning. In Proc. CVPR, 2018. [80] S. Thrun. Lifelong learning algorithms. In Learning to learn. Springer, 1998. [81] S. Thrun and L. Pratt. Learning to learn. Springer Science & Business Media, 1998. [82] Y. Tian, D. Krishnan, and P. Isola. Contrastive multiview coding. ar Xiv preprint ar Xiv:1906.05849, 2019. [83] Y. Tian, Y. Wang, D. Krishnan, J. B. Tenenbaum, and P. Isola. Rethinking few-shot image classification: a good embedding is all you need? In Proc. ECCV, 2020. [84] P. Tokmakov, Y.-X. Wang, and M. Hebert. Learning compositional representations for few-shot recognition. In Proc. ICCV, 2019. [85] G. Tolias, T. Jenicek, and O. Chum. Learning and aggregating deep local descriptors for instance-level recognition. In Proc. ECCV, 2020. [86] E. Triantafillou, T. Zhu, V. Dumoulin, P. Lamblin, U. Evci, K. Xu, R. Goroshin, C. Gelada, K. J. Swersky, P.-A. Manzagol, and H. Larochelle. Meta-dataset: A dataset of datasets for learning to learn from few examples. In Proc. ICLR, 2020. [87] A. van den Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu. Wavenet: A generative model for raw audio. ar Xiv preprint ar Xiv:1609.03499, 2016. [88] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin. Attention is all you need. In Neur IPS, 2017. [89] R. C. Veltkamp. Shape matching: Similarity measures and algorithms. In International Conference on Shape Modeling and Applications, 2001. [90] R. Vilalta and Y. Drissi. A perspective view and survey of meta-learning. Artificial intelligence review, 18, 2002. [91] O. Vinyals, C. Blundell, T. Lillicrap, K. Kavukcuoglu, and D. Wierstra. Matching networks for one shot learning. In Neur IPS, 2016. [92] C. Vondrick, A. Shrivastava, A. Fathi, S. Guadarrama, and K. Murphy. Tracking emerges by colorizing videos. In Proc. ECCV, 2018. [93] C. Wah, S. Branson, P. Welinder, P. Perona, and S. Belongie. The Caltech-UCSD Birds-200-2011 dataset. 2011. [94] X. Wang, R. Girshick, A. Gupta, and K. He. Non-local neural networks. In Proc. CVPR, 2018. [95] Y.-X. Wang, D. Ramanan, and M. Hebert. Learning to model the tail. In Neur IPS, 2017. [96] Z. Wu, Y. Xiong, S. X. Yu, and D. Lin. Unsupervised feature learning via non-parametric instance discrimination. In Proc. CVPR, 2018. [97] W. Xie, L. Shen, and A. Zisserman. Comparator networks. In Proc. ECCV, 2018. [98] K. Yang, K. Qinami, L. Fei-Fei, J. Deng, and O. Russakovsky. Towards fairer datasets: Filtering and balancing the distribution of the people subtree in the imagenet hierarchy. In Conference on Fairness, Accountability, and Transparency, 2020. [99] A. S. Younger, S. Hochreiter, and P. R. Conwell. Meta-learning with backpropagation. In IJCNN 01. International Joint Conference on Neural Networks. Proceedings (Cat. No. 01CH37222), volume 3. IEEE, 2001. [100] S. Zagoruyko and N. Komodakis. Wide residual networks. In Proc. BMVC., 2016. [101] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena. Self-attention generative adversarial networks. In Proc. ICML, 2019. [102] J. Zhang, M. Marszałek, S. Lazebnik, and C. Schmid. Local features and kernels for classification of texture and object categories: A comprehensive study. IJCV, 73(2), 2007. [103] R. Zhang, P. Isola, and A. A. Efros. Colorful image colorization. In Proc. ECCV. Springer, 2016. [104] R. Zhang, P. Isola, and A. A. Efros. Split-brain autoencoders: Unsupervised learning by cross-channel prediction. In Proc. CVPR, 2017. [105] H. Zheng, J. Fu, T. Mei, and J. Luo. Learning multi-attention convolutional neural network for fine-grained image recognition. In Proc. ICCV, 2017.