# contextaware_zeroshot_learning_for_object_recognition__deab3b47.pdf Context-Aware Zero-Shot Learning for Object Recognition Eloi Zablocki * 1 Patrick Bordes * 1 Benjamin Piwowarski 1 Laure Soulier 1 Patrick Gallinari 1 2 Zero-Shot Learning (ZSL) aims at classifying unlabeled objects by leveraging auxiliary knowledge, such as semantic representations. A limitation of previous approaches is that only intrinsic properties of objects, e.g. their visual appearance, are taken into account while their context, e.g. the surrounding objects in the image, is ignored. Following the intuitive principle that objects tend to be found in certain contexts but not others, we propose a new and challenging approach, contextaware ZSL, that leverages semantic representations in a new way to model the conditional likelihood of an object to appear in a given context. Finally, through extensive experiments conducted on Visual Genome, we show that contextual information can substantially improve the standard ZSL approach and is robust to unbalanced classes. 1. Introduction Traditional Computer Vision models, such as Convolutional Neural Networks (CNNs) (Lecun et al., 1998), are designed to classify images into a set of predefined classes. Their performances have kept improving in the last decade, namely on object recognition benchmarks such as Image Net (Deng et al., 2009a), where state-of-the-art models (Zoph et al., 2017; Real et al., 2018) have outmatched humans. However, training such models requires hundreds of manually-labeled instances for each class, which is a tedious and costly acquisition process. Moreover, these models cannot replicate humans capacity to generalize and to recognize objects they have never seen before. As a response to these limitations, Zero-Shot Learning (ZSL) has emerged as an important research field in the last decade (Farhadi et al., 2009a; Mensink et al., 2012; Fu et al., 2015a; Kodirov et al., 2017). In the object recognition field, ZSL aims at labeling an instance of *Equal contribution 1Sorbonne Universit e, CNRS, Laboratoire d Informatique de Paris 6, LIP6, F-75005 Paris, France 2Criteo AI Lab, Paris. Correspondence to: Eloi Zablocki . Proceedings of the 36 th International Conference on Machine Learning, Long Beach, California, PMLR 97, 2019. Copyright 2019 by the author(s). a class for which no supervised data is available, by using knowledge acquired from another disjoint set of classes, for which corresponding visual instances are provided. In the literature, these sets of classes are respectively called target and source domains terms borrowed from the transfer learning community. Generalization from the source to the target domain is achieved using auxiliary knowledge that semantically relates classes of both domains, e.g. attributes or textual representations of the class labels. Previous ZSL approaches only focus on intrinsic properties of objects, e.g. their visual appearance, by the means of handcrafted features e.g. shape, texture, or color (Lampert et al., 2014) or distributed representations learned from text corpora (Akata et al., 2016; Long et al., 2017). The underlying hypothesis is that the identification of entities of the target domain is made possible thanks to the implicit principle of compositionality (a.k.a. Frege s principle (Pelletier, 2001)) an object is formed by the composition of its attributes and characteristics and the fact that other entities of the source domain share the same attributes. For example, if textual resources state that an apple is round and that it can be red or green, this knowledge can be used to identify apples in images because these characteristics ( round , red ) could be shared by classes of the source domain (e.g, round like a ball, red like a strawberry...). We believe that visual context, i.e. the other entities surrounding an object, also explains human s ability to recognize an object that has never been seen before. This assumption relies on the fact that scenes are compositional in the sense that they are formed by the composition of objects they contain. Some works in Computer Vision have exploited visual context to refine the predictions of classification (Mensink et al., 2014) or detection (Bell et al., 2016) models. To the best of our knowledge, context has not been exploited in ZSL because, for obvious reasons, it is impossible to directly estimate the likelihood of a context for objects from the target domain from visual data only. However, textual resources can be used to provide insights on the possible visual context in which an object is expected to appear. To illustrate this, knowing from language that an apple is likely to be found hanging on a tree or in the hand of someone eating it, can be very helpful to identify apples in images. In this paper, our goal is to leverage visual context as an additional source of knowledge for ZSL, by Context-Aware Zero-Shot Learning for Object Recognition exploiting the distributed word representations (Mikolov et al., 2013) of the object class labels. More precisely, we adopt a probabilistic framework in which the probability to recognize a given object is split into three components: (1) a visual component based on its visual appearance (which can be derived from any traditional ZSL approach), (2) a contextual component exploiting its visual context, and (3) a prior component, which estimates the frequency of objects in the dataset. As a complementary contribution, we show that separating prior information in a dedicated component, along with simple yet effective sampling strategies, leads to a more interpretable model, able to deal with imbalanced datasets. Finally, as traditional ZSL datasets lack contextual information, we design a new dedicated setup based on the richly annotated Visual Genome dataset (Krishna et al., 2017). We conduct extensive experiments to thoroughly study the impact of contextual information. 2. Related work Zero-shot learning While state-of-the-art image classification models (Zoph et al., 2017; Real et al., 2018) restrict their predictions to a finite set of predefined classes, ZSL bypasses this important limitation by transferring knowledge acquired from seen classes (source domain) to unseen classes (target domain). Generalization is made possible through the medium of a common semantic space where all classes from both source and target domains are represented by vectors called semantic representations. Historically, the first semantic representations that were used were handcrafted attributes (Farhadi et al., 2009a; Parikh & Grauman, 2011; Mensink et al., 2012; Lampert et al., 2014). In these works, the attributes of a given image are determined and the class with the most similar attributes is predicted. Most methods represent class labels with binary vectors of visual features (e.g, Is Black , Has Claws ) (Lampert et al., 2009; Liu et al., 2011; Fu et al., 2014; Lampert et al., 2014). However, attribute-based methods do not scale efficiently since the attribute ontology is often domainspecific and has to be built manually. To cope with this limitation, more recent ZSL works rely on distributed semantic representations learned from textual datasets such as Wikipedia, using Distributional Semantic Models (Mikolov et al., 2013; Pennington et al., 2014; Peters et al., 2018). These models are based on the distributional hypothesis (Harris, 1954), which states that textual items with similar contexts in text corpora tend to have similar meanings. This is of particular interest in ZSL: all object classes (from both source and target domains) are embedded into the same continuous vector space based on their textual context, which is a rich source of semantic information. Some models directly aggregate textual representations of class labels and the predictions of a CNN (Norouzi et al., 2013), whereas others learn a cross-modal mapping between image representations (given by a CNN) and pre-learned semantic embeddings (Akata et al., 2015; Bucher et al., 2016). At inference, the predicted class of a given image is the nearest neighbor in the semantic embedding space. The cross-modal mapping is linear in most of ZSL works (Palatucci et al., 2009; Romera Paredes & Torr, 2015; Akata et al., 2016; Qiao et al., 2016); this is the case in the present paper. Among these works, the De Vi SE model (Frome et al., 2013) uses a max-margin ranking objective to learn a cross-modal projection and finetune the lower layers of the CNN. Several models have built upon De Vi SE with approaches that learn non-linear mappings between the visual and textual modalities (Ba et al., 2015; Xian et al., 2016), or by using a common multimodal space to embed both images and object classes (Fu et al., 2015b; Long et al., 2017). In this paper, we extend De Vi SE in two directions: by additionally leveraging visual context, and by reformulating it as a probabilistic model that allows coping with an imbalanced class distribution. Visual context The intuitive principle that some objects tend to be found in some contexts but not others, is at the core of many works. In NLP, visual context of objects can be used to build efficient word representations (Zablocki et al., 2018). In Computer Vision, it can be used to refine detection (Chen et al., 2015; Chu & Cai, 2018) or segmentation (Zhang et al., 2018) tasks. Visual context can either be low-level (i.e. raw image pixels) or high-level (i.e. labeled objects). When visual context is exploited in the form of low-level information (Torralba, 2003; Wolf & Bileschi, 2006; Torralba et al., 2010), it often consists of global image features. For instance, in (He et al., 2004), a Conditional Random Field is trained at combining low-level image features to assign to each pixel a class. In high-level approaches, the referential meaning of the context objects (i.e. class labels) is used. For example, Rabinovich et al. (2007) show that high-level context can be used at the post-processing level to reduce the ambiguities of a pre-learned object classification model, by leveraging co-occurrence patterns between objects that are computed from the training set. Moreover, Yu et al. (2016) study the role of context to classify objects: they investigate the importance of contextual indicators, such as object co-occurrence, relative scale and spatial relationships, and find that contextual information can sometimes be more informative than direct visual cues from objects. Spatial relations between objects can also be used in addition to co-occurrences, as in (Galleguillos et al., 2008; Chen et al., 2018). In (Bengio et al., 2013), co-occurrences are computed using external information collected from web documents. The model classifies all objects jointly; it gives an inference method enabling a balance between an image coherence term (given by an image classifier) and a semantic term (given by a co-occurrence matrix). However, the approach is fully supervised, and this setting cannot be applied to ZSL. The Context-Aware Zero-Shot Learning for Object Recognition 1. tennis ball 2. apple 3. lemon 4. orange 5. pear Word embedding CNN representation CNN representation 1. fork 10. apple 979. tennis ball 1. apple 2. lemon 3. orange 708. tennis ball P(.|V, C) Context component Visual component cos(f V (V),w) 30. apple 45. tennis ball 79. fork Prior component f P w : label embedding of candidate class Figure 1. The goal is to find the class (in the target domain) of the object contained within the blue image region V. Its context is formed of labeled objects from the source domain (red plain boxes) and of unlabeled object from the target domain (red dashed boxes). context-aware zero-shot learning task is related to the graph generation tasks (Zellers et al., 2018; Yang et al., 2018) and visual relationship detection (Lu et al., 2016). In conclusion, while many works in NLP and Computer Vision show the importance of visual context, its use in ZSL remains a challenge, that we propose to tackle in this paper. 3. Context-aware Zero-Shot Learning Let O be the set of all object classes, divided in classes from the source domain S and classes from the target domain T . The goal of our approach context-aware ZSL is to determine the class i T of an object contained in an image I, given its visual appearance V and its visual context C. The image I is annotated with bounding boxes, each containing an object. Given the zone V, the context C consists of the surrounding objects in the image. Their classes can either belong to the source domain (C S) or to the target domain (C T ). Note that the class of an object of C T is not accessible in ZSL, only its visual appearance is. 3.1. Model overview We tackle this task by modeling the conditional probability P(i|V, C) of a class i given both the visual appearance V and the visual context C of the object of interest. Given the absence of data in the target domain, we need to limit the complexity of the model, for generalizability s purpose. Accordingly, we suppose that V and C are conditionally independent given the class i we show in the experiments (section 5) that this hypothesis is acceptable. This hypothesis leads to the following expression: P(i|V, C) P(V|i)P(C|i)P(i) (1) where each conditional probability expresses the probability of either the visual appearance V or the context C given class i, and P(i) denotes the prior distribution of the dataset. Each term of this equation is modeled separately. The intuition behind our approach is illustrated in Figure 1, where the blue box contains the object of interest. Here, the class is apple, which belongs to the target domain T . The visual component, which focuses on the zone V, recognizes a tennis ball due to its yellow and round appearance; apple is ranked second. The prior component indicates that apple is slightly more frequent than tennis ball, but the frequency discrepancy may not be high enough to change the prediction of the visual component. In that case, the context component is discriminant: it ranks objects that are likely to be found in a kitchen, and reveals that an apple is far more likely to be found than a tennis ball in this context. Precisely modeling P(C|.), P(V|.) and P(.) is challenging due to the ZSL setting. Indeed, these distributions cannot be computed for classes of the target domain because of the absence of corresponding training data. Thus, to transfer the knowledge acquired from the source domain to the target domain, we use a common semantic space, namely Word2Vec (Mikolov et al., 2013), where source and target class labels are embedded as vectors of Rd, with d the dimension of the space. It is worth noting that we propose to separately learn the prior class distribution P(.) with a ranking loss (in section 3.3). This allows dealing with imbalanced datasets, in contrast to ZSL models like De Vi SE (Frome et al., 2013). This intuition is experimentally validated in section 5.2. 3.2. Description of the model s components Due to both the ZSL setting and the variety of possible context and/or visual appearance of objects, it is not possible to estimate directly the different probabilities of equation 1. Hence, in what follows, we estimate quantities related to P(C|.), P(V|.) and P(.) using parametric energy functions (Le Cun et al., 2006). These quantities are learned separately, as described in section 3.3. Finally, we explain how we combine them to produce the global probability P(.|C, V) in section 3.4. Context-Aware Zero-Shot Learning for Object Recognition Visual component The visual component models P(V|i) by computing the compatibility between the visual appearance V of the object of interest, and the semantic representation wi of the class i. Following previous ZSL works based on cross-modal projections (Frome et al., 2013; Bansal et al., 2018), we introduce fθV , a parametric function mapping an image to the semantic space: fθV (V) = WV .CNN(V) + b V Rd where CNN(V) is a vector in Rdvisual, output by a pretrained CNN truncated at the penultimate layer, WV is a projection matrix ( Rd dvisual) and b V a bias vector in our experiments, dvisual = 2048. The probability that the image region V corresponds to the class i is set to be proportional to the cosine similarity between the projection fθV (V) of V and the semantic representation wi of i: log P(V|i; θV ) cos(fθV (V), wi) := log e Pvisual (2) Context component The context component models P(C|i) by computing a compatibility score between the visual context C, and the semantic representation wi of class i. More precisely, the conditional probability is written: log P(C|i; θC) fθC(C, wi) = fθ1 C hθ2 C(C) wi := log e Pcontext (3) where hθ2 C(C) Rd is a vector representing the context, θC = {θ1 C; θ2 C} are parameters to learn, and is the concatenation operator. To take non-linear and high-order interactions between hθ2 C(C) and wi into account, fθ1 C is modeled by a 2-layer Perceptron. We found that concatenating hθ2 C(C) with wi leads to better results than a cosine similarity, as done in equation 2 for the visual component. To specify the modeling of hθ2 C(C), we propose various context models depending on which context objects are considered and how they are represented. Specifically, a context model is characterized by (a) the domain of context objects that are considered (i.e. source S or target T ) and (b) the way these objects are represented, either by a textual representation of their class label or by a visual representation of their image regions. Accordingly, we distinguish: The low-level (L) approach that computes a representation from the image region Vk of a context object. This produces the following context models: SL = {WCCNN(Vk) + b C|k C S} TL = {WCCNN(Vk) + b C|k C T } The high-level (H) approach which considers semantic representations wk of the class labels k of the context objects (only available for entities of the source domain). This produces context models: SH = {wk|k C S} and TH = {wk|k C T } Note that TH is not defined in the zero-shot setting, since class labels of objects from the target domain are unknown; yet it is used to define Oracle models (section 4.3). These four basic sets of vectors can further be combined in various ways to form new context models (for instance: SL TL, SH SL, SH SL TL, etc.). At last, hθ2 C averages the representations of these vectors to build a global context representation. For example, hθ2 C(CSH TL) equals: 1 |CS| + |CT | (i,Vi) CS wi + X WC.CNN(Vj) + b C i where | | denotes the cardinality of a set of vectors. Prior component The goal of the prior component is to assess whether an entity is frequent or not in images. We estimate P(i) from the semantic representation wi of class i: log P(i; θP ) fθP (wi) := log e Pprior (4) where fθP is a 2-layer Perceptron that outputs a scalar. 3.3. Learning In this section, we explain how we learn the energy functions fθC, fθV and fθP . Each component (resp. context, visual, prior) of our model is assigned a training objective (resp. LC, LV , LP ). As the components are independent by design, they are learned separately. This allows for a better generalization in the target domain, as shown experimentally (section 5.2). Besides, ensuring that some configurations are more likely than others motivates us to model each objective by a max-margin ranking loss, in which a positive configuration is assigned a lower energy than a negative one, following the learning to rank paradigm (Weston et al., 2011). Unlike previous works (Frome et al., 2013), which are generally based on balanced datasets such as Image Net and thus are not concerned with prior information, we want to avoid any bias coming from the imbalance of the dataset in LC and LV , and learn the prior separately with LP . In other terms, the visual (resp. context) component should focus exclusively on the visual appearance (resp. visual context) of objects. This is done with a careful sampling strategy of the negative examples within the ranking objectives, that we detail in the following. To the best of our knowledge, such a discussion relative to prior modeling in learning objectives which is, in our view, paramount in imbalanced datasets such as Visual Genome has not been done in previous research. Positive examples are sampled among entities of the source domain from the data distribution P : they consist in a single object for LP , an object/box pair for LV , an object/context pair for LC. To sample negative examples j from the source domain, we distinguish two ways: (1) For the prior objective LP , negative object classes are sampled from the uniform distribution U: Context-Aware Zero-Shot Learning for Object Recognition LP = E i P E j U γP fθP (wi) + fθP (wj) Noting ji := fθP (wj) fθP (wi), the contribution of two given objects i and j to this objective is: P (i) γP + ji + + P (j) γP ji If P (i) > P (j), i.e. when object class i is more frequent than object class j, this term is minimized when ji = γP , i.e. fθP (wi) = fθP (wj) + γP > fθP (wj). Thus, e Pprior(.; θP ) captures prior information, as it learns to rank objects based on their frequency. (2) For the visual and context objectives, negative object classes are sampled from the prior distribution P (.): LV = E i,V P E j P γV fθV (V) wi + fθV (V) wj LC = E i,C P E j P γC fθC C, wi +fθC C, wj Similarly, the contribution of two given objects i, j and a context C to the objective LC is: P (i)P (j) h P (C|i) γV + fθC(C, wj) fθC(C, wi) +P (C|j) γV + fθC(C, wi) fθC(C, wj) Minimizing this term does not depend on the relative order between P (i) and P (j); thus, e Pcontext(C|.; θC) does not take prior information into account. Moreover, P (C|i) > P (C|j) implies that fθC(C, wi) > fθC(C, wj). The alternative, as done in De Vi SE (Frome et al., 2013), is to sample negative classes uniformly in the source domain in the objective LV . Thus, if the prior is uniform, De Vi SE directly models P(.|V); otherwise, LV cannot be analyzed straightforwardly. Besides, the contributions of visual and prior information are mixed. However, we show that learning the prior separately and imposing the context (resp. visual) component to exclusively focus on contextual (resp. visual) information is more efficient (section 5.2). 3.4. Inference In this section, we detail the inference process. The goal is to combine the predictions of the individual components of the model to form the global probability distribution P(.|V, C). In section 3.3, we detailed how to learn the functions fθC, fθV and fθP , from which log e Pcontext, log e Pvisual and log e Pprior are deduced respectively. However, the normalization constants in equations 2, 3 and 4, which depend on the object class i in the general case, are unknown. As a simplifying hypothesis, we suppose that these normalization constants are scalars that we respectively note αC, αV and αP . This leads to: P(.|V, C) = ( e Pcontext)αC | {z } P(C|.) . ( e Pvisual)αV | {z } P(V|.) . ( e Pprior)αP | {z } P(.) Source domain Target domain log e Pcontext log e Pvisual log e Pprior log e Pcontext log e Pprior log e Pvisual Figure 2. 3D visualization of the unnormalized log-probabilities of each component (N = 500). Context model SL SH TL. To see whether this hypothesis is reasonable, we did some post-hoc analysis of one of our model, and plotted in Figure 2 the values log e Pvisual, log e Pcontext and log e Pprior for positive (red points) and negative (blue points) configurations (i, V, C) of the test set of Visual Genome. We observe that positive and negative triplets are well separated, which empirically validates our initial hypothesis. Hyper-parameters αC, αV and αP are selected on the validation set to compute P(.|C, V). To build models that do not use a visual/contextual component, we simply select a subset of the probabilities and their respective hyperparameters. For example, P(.|C) = ( e Pcontext)αC( e Pprior)αP . 4. Experimental protocol To measure the role of context in ZSL, a dataset that presents annotated objects within a rich visual context is required. However, traditional ZSL datasets, such as Aw A (Farhadi et al., 2009b), CUB-200 (Wah et al., 2011) or LAD (Zhao et al., 2018), are made of images that contain a unique object each, with no or very little surrounding visual context. We rather use Visual Genome (Krishna et al., 2017), a large-scale image dataset (108K images) annotated at a fine-grained level (3.8M object instances), covering various concepts (105K unique object names). This dataset is of particular interest for our work, as objects have richly annotated contexts (31 object instances per image on average). In order to shape the data to our task, we randomly split the set of images of Visual Genome into train/validation/test sets (70%/10%/20% of the total size). To build the set O of all objects classes, we select classes which appear at least 10 times in Visual Genome and have an available Word2vec representation. O contains 4842 object classes; it amounts to 3.4M object instances in the dataset. This dataset is highly imbalanced as 10% of most represented classes amount to 84% of object instances. We define the level of supervision psup as the ratio of the size of the source domain over the total number of objects: psup = |S|/|O|. For a given psup ratio, the source S and target T domains are built by randomly splitting O accordingly. Every object is annotated with a bounding box and we use this supervision in our Context-Aware Zero-Shot Learning for Object Recognition model for entities of both source and target domains. To facilitate future work on context-aware ZSL, we publicly release data splits and annotations 1. 4.2. Evaluation methodology and metrics We adopt the conventional setting for ZSL, which implies entities to be retrieved only among the target domain T . Besides, we also evaluate the performance of the model to retrieve entities of the source domain S (with models tuned on the target domain). The model s prediction takes the form of a list of n classes, sorted by probability; the rank of the correct class in that list is noted r. Depending on the setting, n equals |T | or |S|. We define the First Relevant (FR) metric with FR = 2 n 1(r 1). To further evaluate the performance over the whole test set, the Mean First Relevant (MFR) metric is used (Fuhr, 2017). It is computed by taking the mean value of FR scores obtained on each image of the test set. Note that the factor 2 n 1 rescales the metric such that the MFR score of a random baseline is 100%, while the MFR of a perfect model would be 0%. The MFR metric has the advantage to be interval-scale-based, unlike more traditional Recall@k metrics or Mean Reciprocal Ranks metrics (Ferrante et al., 2017), and thus can be averaged; this allows for meaningful comparison with a varying psup. 4.3. Scenarios and Baselines Model scenarios Model scenarios depend on the information that is used in the probabilistic setting: , C, V or both C and V. When contextual information is involved, a context model is specified to represent C, which we note C . The different context models are {SH, SL, TL, SL TL, SH TL, SL SH TL}. For clarity s sake, we note our model M. For example, M(CSH TL, V) models the probability P(CSH TL|.)P(V|.)P(.) as explained in 3.4, M(V) models P(V|.)P(.), and M( ) models P(.). Oracles To evaluate upper-limit performances for our models, we define Oracle baselines where classes of target objects are used, which is not allowed in the zero-shot setting. Note that every Oracle leverages visual information. True Prior: This Oracle uses, for its prior component, the true prior distribution P (i) = #i M computed for all objects of both source and target domains on the full dataset, where #i is the number of instances of the i-th class in images and M is the total number of images. Visual Bayes: This Oracle uses P (.) for its prior component as well. Its context component uses co-occurrence statistics between objects computed on the full dataset: P im(C|i) = Q c C Pco-oc(c|i) where Pco-oc(c|i) = #(c,i)M #c#i is the probability that objects c and i co-occur in images, with #(c, i) the number of co-occurrences of c and i. 1https://data.lip6.fr/context_aware_zsl/ Textual Bayes: Inspired by (Bengio et al., 2013), this Oracle is similar to Visual Bayes, except that its prior P text(.) and context component P text(.|C) are based on textual cooccurrences instead of image co-occurrences: Pco-oc(c|i) is computed by counting co-occurrences of words c and i in windows of size 8 in the Wikipedia dataset, and P text(i) is computed by summing the number of instances of the i-th class divided by the total size of Wikipedia. Semantic representations for all objects: M(CSH TH, V) uses word embeddings of both source and target objects. Baselines M(C V): To study the validity of the hypothesis about the conditional independence of C and V, we introduce a baseline where we directly model P(C, V|.)P(.). To do so, we replace, in the expression of LV (equation 6), fθV (V) by the concatenation of h(C) and fθV (V) projected in Rd with a 2-layer Perceptron. De Vi SE(V): To evaluate the impact of our Bayesian model (equation 1) and our sampling strategy (section 3.3), we compare against De Vi SE (Frome et al., 2013). De Vi SE(V) is different from M(V) because negative examples in LV are uniformly sampled, and the prior P(.) is not learned. De Vi SE(C V): similarly to M(C V), we define a baseline that does not rely on the conditional independence of C and V, using the same sampling strategy as De Vi SE. M(CI, V): To understand the importance of context supervision, i.e. annotations of context objects (boxes and classes), we design a baseline where no context annotations are used. The context is the whole image without the zone V of the object, which is masked out. The associated context model is = I with h(CI) = gθI(I \ V) ; gθI is a parametric function to be learned. This baseline is inspired from (Torralba et al., 2010), where global image features are used to refine the prediction of an image model. 4.4. Implementation details For each objective LC, LV and LP , at each iteration of the learning algorithm, 5 negative entities are sampled per positive example. Word representations are vectors of R300, learned with the Skip-Gram algorithm (Mikolov et al., 2013) on Wikipedia. Image regions are cropped, rescaled to (299 299), and fed to CNN, an Inception-v3 CNN (Szegedy et al., 2016), whose weights are kept fixed during training. This model is pretrained on Image Net (Deng et al., 2009b). As a result, every Image Net class that belongs to the total set of objects O was included in the source domain S. Models are trained with Adam (Kingma & Ba, 2014) and regularized with a L2-penalty; the weight of this penalty decreases when the level of supervision increases, as the model is less prone to overfitting. All hyper-parameters are cross-validated on classes of the target domain, on the validation set. Context-Aware Zero-Shot Learning for Object Recognition Table 1. Evaluation of various information sources, with varying levels of supervision. MFR scores in %. δC is the relative improvement (in %) of M(CSH , V) over M(V). Target domain T Source domain S psup 10% 50% 90% 10% 50% 90% Domain size 4358 2421 484 484 2421 4358 Random 100 100 100 100 100 100 M( ) 38.6 23.7 13.8 12.0 10.6 11.2 M(V) 20.5 10.7 6.0 1.5 2.6 3.6 M(CSH) 28.7 14.4 9.1 4.2 4.3 4.4 M(CSH, V) 18.1 9.0 5.2 1.1 1.9 2.4 δC (%) 11.6 16.4 12.1 23.7 27.3 31.5 5.1. The importance of context In this section, we evaluate the contribution of contextual information, with varying levels of supervision psup. We fix a simple context model ( = SH) and report MFR results with psup = 10, 50, 90% in Table 1 for every combination of information sources: , V, C and (C, V) we observe similar trends for the other context models. Results highlight that contextual knowledge acquired from the source domain can be transferred to the target domain, as M(CSH) significantly outperforms the Random baseline. As expected, it is not as useful as visual information: M(V) MFR < M(CSH), where MFR < means lower MFR scores, i.e. better performances. However, Table 1 demonstrates that contextual and visual information are complementary: using M(CSH, V) outperforms both M(CSH) and M(V). Interestingly, as the learned prior model M( ) is also able to generalize, we show that visual frequency can somehow be learned from textual semantics, which extends previous work where word embeddings were shown to be a good predictor of textual frequency (Schakel & Wilson, 2015). When psup increases, we observe that all models are better at retrieving objects of the target domain (i.e. MFR decreases), which is intuitive because models are trained on more data and thus generalize better to recognize entities from the target domain. Besides, when psup increases, the context is also more abundant. This explains: (1) the decreasing MFR values for model M(CSH) on T , (2) the increasing relative improvement δC of M(CSH, V) over M(V) on S. However, on the target domain, we note that δC does not monotonously increase with psup. A possible explanation is that the visual component improves faster than the context component, so the relative contribution brought by context to the final model M(CSH, V) decreases after psup = 50%. Since the highest relative improvement δC (in T ) is attained with psup = 50%, we fix the standard level of supervision psup = 50% in the rest of the experiments; this amounts to 2421 classes in both source and target domains. Table 2. MFR performances (given in %) for all baselines and scenarios. psup = 50%. Oracle results, written in italics, are not taken into account to determine the best scores, written in bold. Model Probability T S Textual Bayes P text(C|.)P (V|.)P text(.) 14.54 6.73 M(CSH TH , V) P (CSH TH |.)P (V|.)P (.) 7.57 2.53 True Prior P (V|.)P (.) 4.92 2.63 Visual Bayes P im(C|.)P (V|.)P (.) 3.40 2.11 De Vi SE(V) P (.|V) 10.73 3.62 De Vi SE(CSH V) P (.|CSH , V) 10.11 3.11 M(CSH V) P (CSH , V|.)P (.) 10.07 1.85 M(CI, V) P (CI|.)P (V|.)P (.) 9.19 2.13 M(V) P (V|.)P (.) 10.72 2.64 M(CSL, V) P (CSL|.)P (V|.)P (.) 9.01 2.05 M(CTL, V) P (CTL|.)P (V|.)P (.) 9.00 2.13 M(CSH , V) P (CSH |.)P (V|.)P (.) 8.96 1.92 M(CSH SL, V) P (CSH SL|.)P (V|.)P (.) 8.71 1.88 M(CSL TL, V) P (CSL TL|.)P (V|.)P (.) 8.60 1.93 M(CSH TL, V) P (CSH TL|.)P (V|.)P (.) 8.52 1.86 M(CSH SL TL, V) P (CSH SL TL|.)P (V|.)P (.) 8.31 1.79 5.2. Modeling contextual information In this section, we compare the different context models; results are reported in Table 2. First, underlying hypotheses of our model are experimentally tested. (1) Modeling context and prior information with semantic representations (models M(C , V)) is far more efficient than using direct textual co-occurrences, as shown by the Textual Bayes baseline, which is the weaker model despite being an Oracle. (2) Moreover, we show that the hypothesis on the conditional independence of C and V is acceptable, as separately modeling C and V gives better results than jointly modeling them (i.e. M(CSH, V) MFR < M(CSH V)). (3) Furthermore, we observe that our approach M(V) is more efficient to capture the imbalanced class distribution of the source domain, compared to De Vi SE(V); indeed, True Prior M(V) , whereas True Prior MFR < De Vi SE(V) on S. Even if the improvement is only significant for the source domain S, it indicates that separately using information sources is clearly a superior approach to further integrate contextual information. Second, as observed in the case of the context model SH (section 5.1), using contextual information is always beneficial. Indeed, all models with context M(C , V) improve over M(V) which is the model with no contextual information both on target and source domains. In more details, we observe that performances increase when additional information is used: (1) when the bounding boxes annotations are available: all of our models that use both C and V outperform the baseline M(CI, V), which could also be explained by the useless noise outside the object boxes in the image and the difficulty of computing a global context from raw image, (2) when context objects are labeled and high-level features are used instead of low-level features, e.g. SH MFR < SL and SH TH MFR < SH TL, (3) when more context objects are considered (e.g. SL TL MFR < SL), (4) when Context-Aware Zero-Shot Learning for Object Recognition person man helmet baseball shirt door cabinet fridge wall tooth grass zebra tree building giraffe tree window grass light man grass player man helmet ground 0 handle field player dirt house FR (%) M(V) M(C, V) Most common co-occurring Figure 3. Boxplot representing the distribution of the correct ranks (First Relevant in %) for five randomly selected classes of the target domain, with the context model SL SH TL. Below are listed, by order of frequency, the classes that co-occur the most with the object of interest (classes of T in green; S in red). low-level information is used complementarily to high-level information (e.g. SL SH TL MFR < SL TL). As a result, the best performance is attained for M(CSL SH TL, V), with a 22% (resp. 32%) relative improvement in the target (resp. source) domain compared to M(V). We note that there is still room for improvement to approach ground-truth distributions for objects of the target domain (e.g, towards word embeddings able to better capture visual context). Indeed, even if our models outperform True Prior and Visual Bayes on the source domain, these Oracle baselines are still better on the target domain, hence showing that learning the visual context of objects from textual data is challenging. 5.3. Qualitative Experiments To gain a deeper understanding of contextual information, we compare in Figure 3 the predictions of M(V) and the global model M(C, V). We randomly select five classes of the target domain and plot, for all instances of these classes in the test set of Visual Genome, the distribution of the predicted ranks of the correct class (in percentage); we also list the classes that appear the most in the context of these classes. We observe that, for certain classes (player, handle and field), contextual information helps to refine the predictions; for others (house and dirt), contextual information degrades the quality of the predictions. First, we can outline that visual context can guide the model towards a more precise prediction. For example, a player, without context, could be categorized as person, man or woman; but visual context provides important complementary information (e.g, helmet, baseball) that grounds person in a sport setting, and thus suggests that the person could be playing. Visual context is also particularly relevant when the 2. connector 3. indicator 4. carnations 1. jetliner 4. airplane 2. fingertip 376. filters 3. jetliner 11. airplane 1. airplane 256. airplane 1717. jetliner stems grass building (iii) Figure 4. Qualitative examples where the global model M(CSL SH TL, V) correctly retrieves the class (T classes only). object of interest has a generic shape. For example, handle, without context, is visually similar to many round objects; but the presence of objects like door or fridge in the context helps determine the nature of the object of interest. To get a better insight on the role of context, we cherrypicked examples where the visual or the prior component is inacurrate and the context component is able to counterbalance the final prediction (Figure 4). In (i), for example, the visual component ranks flower at position 223. However, the context component assesses flower to be highly probable in this context, due to the presence of source objects like vase, water, stems or grass, but also target objects like the other flowers around. At the inference phase, probabilities are aggregated and flower is ranked first. It is worth noting that our work is not without limitations. Indeed, some classes (such as house and dirt) have a wide range of possible contexts; in these cases, context is not a discriminating factor. This is confirmed by a complementary analysis: the Spearman correlation between the number of unique context objects and δC, the relative gain of M(CSH, V) over M(V), is ρ = 0.31. In other terms, contextual information is useful for specific objects, which appear in particular contexts; for objects that are too generic, adding contextual information can be a source of noise. 6. Conclusion In this paper, we introduced a new approach for ZSL: context-aware ZSL, along with a corresponding model, using complementary contextual information that significantly improves predictions. Possible extensions could include spatial features of objects, and, more importantly, removing the dependence on the detection of object boxes to make it fully applicable to real-world images (e.g. by using a Region Proposal Network (Ren et al., 2015)) Finally, designing grounded word embeddings that include more visual context information would also benefit such models. Context-Aware Zero-Shot Learning for Object Recognition Acknowledgments This work is partially supported by the CHIST-ERA EU project MUSTER 1 (ANR-15-CHR2-0005) and the Labex SMART (ANR-11-LABX-65) supported by French state funds managed by the ANR within the Investissements d Avenir program under reference ANR-11-IDEX-0004-02. Akata, Z., Reed, S. E., Walter, D., Lee, H., and Schiele, B. Evaluation of output embeddings for fine-grained image classification. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, June 7-12, 2015, pp. 2927 2936, 2015. doi: 10.1109/ CVPR.2015.7298911. URL https://doi.org/10. 1109/CVPR.2015.7298911. Akata, Z., Perronnin, F., Harchaoui, Z., and Schmid, C. Label-embedding for image classification. IEEE Trans. Pattern Anal. Mach. Intell., 38(7):1425 1438, 2016. doi: 10.1109/TPAMI.2015.2487986. URL https://doi. org/10.1109/TPAMI.2015.2487986. Ba, L. J., Swersky, K., Fidler, S., and Salakhutdinov, R. Predicting deep zero-shot convolutional neural networks using textual descriptions. In 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, December 7-13, 2015, pp. 4247 4255, 2015. doi: 10.1109/ICCV.2015.483. URL https://doi.org/ 10.1109/ICCV.2015.483. Bansal, A., Sikka, K., Sharma, G., Chellappa, R., and Divakaran, A. Zero-shot object detection. In Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part I, pp. 397 414, 2018. doi: 10.1007/978-3-030-01246-5\ 24. URL https://doi.org/10.1007/978-3-03001246-5_24. Bell, S., Zitnick, C. L., Bala, K., and Girshick, R. B. Insideoutside net: Detecting objects in context with skip pooling and recurrent neural networks. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 2874 2883, 2016. doi: 10.1109/CVPR.2016.314. URL https://doi.org/10.1109/CVPR.2016.314. Bengio, S., Dean, J., Erhan, D., Ie, E., Le, Q. V., Rabinovich, A., Shlens, J., and Singer, Y. Using web cooccurrence statistics for improving image categorization. Co RR, abs/1312.5697, 2013. URL http://arxiv. org/abs/1312.5697. Bucher, M., Herbin, S., and Jurie, F. Improving semantic embedding consistency by metric learning for zeroshot classiffication. In Computer Vision - ECCV 2016 - 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part V, pp. 730 746, 2016. doi: 10.1007/978-3-319-46454-1\ 44. URL https://doi.org/10.1007/978-3-31946454-1_44. Chen, Q., Song, Z., Dong, J., Huang, Z., Hua, Y., and Yan, S. Contextualizing object detection and classification. IEEE Trans. Pattern Anal. Mach. Intell., 37(1):13 27, 2015. doi: 10.1109/TPAMI.2014.2343217. URL https: //doi.org/10.1109/TPAMI.2014.2343217. Chen, X., Li, L., Fei-Fei, L., and Gupta, A. Iterative visual reasoning beyond convolutions. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pp. 7239 7248, 2018. doi: 10.1109/CVPR.2018.00756. URL http://openaccess.thecvf.com/content_ cvpr_2018/html/Chen_Iterative_Visual_ Reasoning_CVPR_2018_paper.html. Chu, W. and Cai, D. Deep feature based contextual model for object detection. Neurocomputing, 275:1035 1042, 2018. doi: 10.1016/j.neucom.2017. 09.048. URL https://doi.org/10.1016/j. neucom.2017.09.048. Deng, J., Dong, W., Socher, R., Li, L., Li, K., and Li, F. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pp. 248 255, 2009a. doi: 10.1109/CVPRW.2009.5206848. URL https://doi. org/10.1109/CVPRW.2009.5206848. Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei Fei, L. Image Net: A Large-Scale Hierarchical Image Database. In CVPR09, 2009b. Farhadi, A., Endres, I., Hoiem, D., and Forsyth, D. A. Describing objects by their attributes. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pp. 1778 1785, 2009a. doi: 10.1109/CVPRW.2009.5206772. URL https://doi. org/10.1109/CVPRW.2009.5206772. Farhadi, A., Endres, I., Hoiem, D., and Forsyth, D. A. Describing objects by their attributes. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pp. 1778 1785, 2009b. doi: 10.1109/CVPRW.2009.5206772. URL https://doi. org/10.1109/CVPRW.2009.5206772. Ferrante, M., Ferro, N., and Pontarollo, S. Are IR evaluation measures on an interval scale? In Proceedings of the ACM SIGIR International Conference on Context-Aware Zero-Shot Learning for Object Recognition Theory of Information Retrieval, ICTIR 2017, Amsterdam, The Netherlands, October 1-4, 2017, pp. 67 74, 2017. doi: 10.1145/3121050.3121058. URL https: //doi.org/10.1145/3121050.3121058. Frome, A., Corrado, G. S., Shlens, J., Bengio, S., Dean, J., Ranzato, M., and Mikolov, T. Devise: A deep visual-semantic embedding model. In Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013, Lake Tahoe, Nevada, United States., pp. 2121 2129, 2013. URL http: //papers.nips.cc/paper/5204-devise-adeep-visual-semantic-embedding-model. Fu, Y., Hospedales, T. M., Xiang, T., and Gong, S. Learning multimodal latent attributes. IEEE Trans. Pattern Anal. Mach. Intell., 36(2):303 316, 2014. doi: 10.1109/TPAMI.2013.128. URL https://doi.org/ 10.1109/TPAMI.2013.128. Fu, Y., Yang, Y., Hospedales, T. M., Xiang, T., and Gong, S. Transductive multi-label zero-shot learning. Co RR, abs/1503.07790, 2015a. URL http://arxiv.org/ abs/1503.07790. Fu, Z., Xiang, T. A., Kodirov, E., and Gong, S. Zero-shot object recognition by semantic manifold distance. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, June 7-12, 2015, pp. 2635 2644, 2015b. doi: 10.1109/CVPR.2015.7298879. URL https://doi.org/10.1109/CVPR.2015. 7298879. Fuhr, N. Some common mistakes in IR evaluation, and how they can be avoided. SIGIR Forum, 51(3):32 41, 2017. doi: 10.1145/3190580.3190586. URL https: //doi.org/10.1145/3190580.3190586. Galleguillos, C., Rabinovich, A., and Belongie, S. J. Object categorization using co-occurrence, location and appearance. In 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2008), 24-26 June 2008, Anchorage, Alaska, USA, 2008. doi: 10.1109/CVPR.2008.4587799. URL https://doi. org/10.1109/CVPR.2008.4587799. Harris, Z. S. Distributional structure. Word, 10(2-3):146 162, 1954. He, X., Zemel, R. S., and Carreira-Perpi n an, M. A. Multiscale conditional random fields for image labeling. In 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2004), with CD-ROM, 27 June - 2 July 2004, Washington, DC, USA, pp. 695 702, 2004. doi: 10.1109/CVPR.2004. 173. URL http://doi.ieeecomputersociety. org/10.1109/CVPR.2004.173. Kingma, D. P. and Ba, J. Adam: A method for stochastic optimization. Co RR, abs/1412.6980, 2014. URL http: //arxiv.org/abs/1412.6980. Kodirov, E., Xiang, T., and Gong, S. Semantic autoencoder for zero-shot learning. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 4447 4456, 2017. doi: 10.1109/CVPR.2017.473. URL https: //doi.org/10.1109/CVPR.2017.473. Krishna, R., Zhu, Y., Groth, O., Johnson, J., Hata, K., Kravitz, J., Chen, S., Kalantidis, Y., Li, L., Shamma, D. A., Bernstein, M. S., and Fei-Fei, L. Visual genome: Connecting language and vision using crowdsourced dense image annotations. International Journal of Computer Vision, 123(1):32 73, 2017. doi: 10.1007/s11263016-0981-7. URL https://doi.org/10.1007/ s11263-016-0981-7. Lampert, C. H., Nickisch, H., and Harmeling, S. Learning to detect unseen object classes by between-class attribute transfer. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pp. 951 958, 2009. doi: 10.1109/CVPRW.2009. 5206594. URL https://doi.org/10.1109/ CVPRW.2009.5206594. Lampert, C. H., Nickisch, H., and Harmeling, S. Attributebased classification for zero-shot visual object categorization. IEEE Trans. Pattern Anal. Mach. Intell., 36(3):453 465, 2014. doi: 10.1109/TPAMI.2013.140. URL https: //doi.org/10.1109/TPAMI.2013.140. Lecun, Y., Bottou, L., Bengio, Y., and Haffner, P. Gradientbased learning applied to document recognition. In Proceedings of the IEEE, pp. 2278 2324, 1998. Le Cun, Y., Chopra, S., and Hadsell, R. A tutorial on energybased learning. Predicting Structured Data, 1:0, 2006. Liu, J., Kuipers, B., and Savarese, S. Recognizing human actions by attributes. In The 24th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2011, Colorado Springs, CO, USA, 20-25 June 2011, pp. 3337 3344, 2011. doi: 10.1109/CVPR.2011.5995353. URL https://doi.org/10.1109/CVPR.2011. 5995353. Long, Y., Liu, L., Shao, L., Shen, F., Ding, G., and Han, J. From zero-shot learning to conventional supervised classification: Unseen visual data synthesis. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, Context-Aware Zero-Shot Learning for Object Recognition CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 6165 6174, 2017. doi: 10.1109/CVPR.2017.653. URL https://doi.org/10.1109/CVPR.2017.653. Lu, C., Krishna, R., Bernstein, M. S., and Li, F. Visual relationship detection with language priors. In Computer Vision - ECCV 2016 - 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part I, pp. 852 869, 2016. doi: 10.1007/978-3-31946448-0\ 51. URL https://doi.org/10.1007/ 978-3-319-46448-0_51. Mensink, T., Verbeek, J. J., Perronnin, F., and Csurka, G. Metric learning for large scale image classification: Generalizing to new classes at near-zero cost. In Computer Vision - ECCV 2012 - 12th European Conference on Computer Vision, Florence, Italy, October 7-13, 2012, Proceedings, Part II, pp. 488 501, 2012. doi: 10.1007/978-3-642-33709-3\ 35. URL https:// doi.org/10.1007/978-3-642-33709-3_35. Mensink, T., Gavves, E., and Snoek, C. Costa: Cooccurrence statistics for zero-shot classification. 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2441 2448, 2014. Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., and Dean, J. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013, Lake Tahoe, Nevada, United States., pp. 3111 3119, 2013. URL http://papers.nips.cc/paper/ 5021-distributed-representationsof-words-and-phrases-and-theircompositionality. Norouzi, M., Mikolov, T., Bengio, S., Singer, Y., Shlens, J., Frome, A., Corrado, G., and Dean, J. Zero-shot learning by convex combination of semantic embeddings. Co RR, abs/1312.5650, 2013. URL http://arxiv. org/abs/1312.5650. Palatucci, M., Pomerleau, D., Hinton, G. E., and Mitchell, T. M. Zero-shot learning with semantic output codes. In Advances in Neural Information Processing Systems 22: 23rd Annual Conference on Neural Information Processing Systems 2009. Proceedings of a meeting held 7-10 December 2009, Vancouver, British Columbia, Canada., pp. 1410 1418, 2009. URL http://papers.nips.cc/paper/3650zero-shot-learning-with-semanticoutput-codes. Parikh, D. and Grauman, K. Interactively building a discriminative vocabulary of nameable attributes. In The 24th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2011, Colorado Springs, CO, USA, 20-25 June 2011, pp. 1681 1688, 2011. doi: 10.1109/CVPR.2011.5995451. URL https://doi. org/10.1109/CVPR.2011.5995451. Pelletier, F. J. Did frege believe frege s principle? Journal of Logic, Language and information, 10(1):87 114, 2001. Pennington, J., Socher, R., and Manning, C. D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, October 2529, 2014, Doha, Qatar, A meeting of SIGDAT, a Special Interest Group of the ACL, pp. 1532 1543, 2014. URL http://aclweb.org/anthology/D/D14/ D14-1162.pdf. Peters, M. E., Neumann, M., Iyyer, M., Gardner, M., Clark, C., Lee, K., and Zettlemoyer, L. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, New Orleans, Louisiana, USA, June 1-6, 2018, Volume 1 (Long Papers), pp. 2227 2237, 2018. Qiao, R., Liu, L., Shen, C., and van den Hengel, A. Less is more: Zero-shot learning from online textual documents with noise suppression. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 2249 2257, 2016. doi: 10.1109/CVPR.2016.247. URL https: //doi.org/10.1109/CVPR.2016.247. Rabinovich, A., Vedaldi, A., Galleguillos, C., Wiewiora, E., and Belongie, S. J. Objects in context. In IEEE 11th International Conference on Computer Vision, ICCV 2007, Rio de Janeiro, Brazil, October 14-20, 2007, pp. 1 8, 2007. doi: 10.1109/ICCV.2007.4408986. URL https: //doi.org/10.1109/ICCV.2007.4408986. Real, E., Aggarwal, A., Huang, Y., and Le, Q. V. Regularized evolution for image classifier architecture search. Co RR, abs/1802.01548, 2018. URL http://arxiv. org/abs/1802.01548. Ren, S., He, K., Girshick, R. B., and Sun, J. Faster R-CNN: towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, December 7-12, 2015, Montreal, Quebec, Canada, pp. 91 99, 2015. URL http://papers.nips.cc/ paper/5638-faster-r-cnn-towards-realtime-object-detection-with-regionproposal-networks. Context-Aware Zero-Shot Learning for Object Recognition Romera-Paredes, B. and Torr, P. H. S. An embarrassingly simple approach to zero-shot learning. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015, pp. 2152 2161, 2015. URL http://jmlr.org/proceedings/ papers/v37/romera-paredes15.html. Schakel, A. M. J. and Wilson, B. J. Measuring word significance using distributed representations of words. Co RR, abs/1508.02297, 2015. URL http://arxiv.org/ abs/1508.02297. Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. Rethinking the inception architecture for computer vision. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 2818 2826, 2016. doi: 10. 1109/CVPR.2016.308. URL https://doi.org/10. 1109/CVPR.2016.308. Torralba, A. Contextual priming for object detection. International Journal of Computer Vision, 53(2):169 191, 2003. doi: 10.1023/A:1023052124951. URL https: //doi.org/10.1023/A:1023052124951. Torralba, A., Murphy, K. P., and Freeman, W. T. Using the forest to see the trees: exploiting context for visual object detection and localization. Commun. ACM, 53(3):107 114, 2010. doi: 10.1145/1666420.1666446. URL http: //doi.acm.org/10.1145/1666420.1666446. Wah, C., Branson, S., Welinder, P., Perona, P., and Belongie, S. The caltech-ucsd birds-200-2011 dataset. 2011. Weston, J., Bengio, S., and Usunier, N. WSABIE: scaling up to large vocabulary image annotation. In IJCAI 2011, Proceedings of the 22nd International Joint Conference on Artificial Intelligence, Barcelona, Catalonia, Spain, July 16-22, 2011, pp. 2764 2770, 2011. doi: 10.5591/978-1-57735-516-8/IJCAI11460. URL https://doi.org/10.5591/978-157735-516-8/IJCAI11-460. Wolf, L. and Bileschi, S. M. A critical view of context. International Journal of Computer Vision, 69(2):251 261, 2006. doi: 10.1007/s11263-006-7538-0. URL https: //doi.org/10.1007/s11263-006-7538-0. Xian, Y., Akata, Z., Sharma, G., Nguyen, Q. N., Hein, M., and Schiele, B. Latent embeddings for zero-shot classification. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 69 77, 2016. doi: 10. 1109/CVPR.2016.15. URL https://doi.org/10. 1109/CVPR.2016.15. Yang, J., Lu, J., Lee, S., Batra, D., and Parikh, D. Graph R-CNN for scene graph generation. In Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part I, pp. 690 706, 2018. doi: 10.1007/978-3-030-01246-5\ 41. URL https://doi.org/10.1007/978-3-03001246-5_41. Yu, R., Chen, X., Morariu, V. I., and Davis, L. S. The role of context selection in object detection. In Proceedings of the British Machine Vision Conference 2016, BMVC 2016, York, UK, September 19-22, 2016, 2016. URL http://www.bmva.org/bmvc/2016/ papers/paper133/index.html. Zablocki, E., Piwowarski, B., Soulier, L., and Gallinari, P. Learning Multi-Modal Word Representation Grounded in Visual Context. In Association for the Advancement of Artificial Intelligence (AAAI), New Orleans, United States, February 2018. URL https://hal.archivesouvertes.fr/hal-01632414. Zellers, R., Yatskar, M., Thomson, S., and Choi, Y. Neural motifs: Scene graph parsing with global context. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pp. 5831 5840, 2018. doi: 10.1109/CVPR.2018.00611. URL http://openaccess.thecvf.com/content_ cvpr_2018/html/Zellers_Neural_Motifs_ Scene_CVPR_2018_paper.html. Zhang, H., Dana, K. J., Shi, J., Zhang, Z., Wang, X., Tyagi, A., and Agrawal, A. Context encoding for semantic segmentation. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pp. 7151 7160, 2018. doi: 10.1109/CVPR.2018.00747. URL http://openaccess.thecvf.com/ content_cvpr_2018/html/Zhang_Context_ Encoding_for_CVPR_2018_paper.html. Zhao, B., Chang, B., Jie, Z., and Sigal, L. Modular generative adversarial networks. Co RR, abs/1804.03343, 2018. URL http://arxiv.org/abs/1804.03343. Zoph, B., Vasudevan, V., Shlens, J., and Le, Q. V. Learning transferable architectures for scalable image recognition. Co RR, abs/1707.07012, 2017. URL http://arxiv. org/abs/1707.07012.