# multilabel_classification_with_label_graph_superimposing__c705894c.pdf The Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20) Multi-Label Classification with Label Graph Superimposing Ya Wang,$ Dongliang He, Fu Li, Xiang Long, Zhichao Zhou, Jinwen Ma,$ Shilei Wen $School of Mathematical Sciences and LMAM, Peking University, China Department of Computer Vision Technology (VIS), Baidu Inc., Beijing, China {wangyachn, jwma}@math.pku.edu.cn, {hedongliang01, lifu, longxiang, zhouzhichao01, wenshilei}@baidu.com Images or videos always contain multiple objects or actions. Multi-label recognition has been witnessed to achieve pretty performance attribute to the rapid development of deep learning technologies. Recently, graph convolution network (GCN) is leveraged to boost the performance of multi-label recognition. However, what is the best way for label correlation modeling and how feature learning can be improved with label system awareness are still unclear. In this paper, we propose a label graph superimposing framework to improve the conventional GCN+CNN framework developed for multi-label recognition in the following two aspects. Firstly, we model the label correlations by superimposing label graph built from statistical co-occurrence information into the graph constructed from knowledge priors of labels, and then multilayer graph convolutions are applied on the final superimposed graph for label embedding abstraction. Secondly, we propose to leverage embedding of the whole label system for better representation learning. In detail, lateral connections between GCN and CNN are added at shallow, middle and deep layers to inject information of label system into backbone CNN for label-awareness in the feature learning process. Extensive experiments are carried out on MSCOCO and Charades datasets, showing that our proposed solution can greatly improve the recognition performance and achieves new state-of-the-art recognition performance. Introduction Multi-label is a natural property of images or videos, it is usually the case that a image or video contains multiple objects or actions. In the computer vision community, multilabel recognition is a fundamental and practical task, and has attracted increasing research efforts. Given the great success of single label image/video classification brought by deep convolutional networks (He et al. 2015; Carreira and Zisserman 2017; He et al. 2016a; Feichtenhofer et al. 2018; Wu et al. 2019), multi-label recognition can achieve pretty performance by naively treating each label as an independent individual and applying multiple binary classification equal contribution. This work was done when Ya Wang was a full-time research intern at Baidu. Corresponding author Copyright c 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. (a) Examples on MS-COCO (b) Examples on Charades Figure 1: Examples of label relationship in multi-label datasets. (a) illustrates the co-occurrence of Sports Ball and Tennis Racket on the MS-COCO datasets, we can see the frequency that Tennis Racket co-occurs with Sports Ball is as high as 0.42. Similarly, (b) showcases an example of Sitting on Couch and Watching Television from the Charades dataset. to predict whether a label presents or not. However, we argue that the following two aspects should be taken into consideration for such a task. First of all, labels co-occur in images or videos with priors. As illustrated in Figure 1, with great chance, Sports Ball comes together with Tennis Racket and a man Sitting on Couch is Watching Television simultaneously. Then, a question is naturally raised, how to model the re- lation among labels to leverage such priors for better performance? Secondly, given input X, the common practice for predicting its labels can be formulated as a two-stage mapping y = F1 F0(X), where F0 : X f denotes the CNN feature extraction process and F1 : f y is the mapping from feature space to label space. Labels are only explicitly involved in the last stage as supervision in the training phase. Therefore, the further question is, for a specific multi-label classification task, whether and how the mutual-related label space can explicitly help the feature learning process F0? To take into account the label correlations, some approaches have been proposed. For example, probabilistic graph model was used in (Li et al. 2016; Li, Zhao, and Guo 2014) and RNN was used in (Wang et al. 2016a) to capture dependencies among labels. However, probabilistic graph models may suffer from scalability issues given their computational cost. RNN model relies on predefined or learned label sequential order and fails to well capture the global dependencies. Recently, graph convolutional network (Kipf and Welling 2016), aka GCN, has witnessed prevailing success in modeling relationship among vertices of a graph. Such a tool was leveraged to model the relation of the label system for multi-label recognition in (Chen et al. 2019). Meanwhile, the label graph was built simply by utilizing the frequency of label co-occurrence. Another direction is to implicitly model label correlations via local image regions attention, as was done in (Wang et al. 2017; Zhu et al. 2017a). In addition, all the aforementioned solutions follow the conventional practice of two-stage mapping and the whole structure of label system is ignored in learning the feature space. In this paper, we attempt to find possible answers for the two questions. We propose a label graph superimposed deep convolution network called KSSNet for this task. The superimposing means the following two folds in our framework: (1) to model the priors of co-occurrence of labels following the GCN paradigm, instead of using statistics of label co-occurrence alone to build the relation graph of the label system, we propose to superimpose knowledge based graph into statistics based graph for constructing the final one. (2) In order to learn better feature representations for a specific multi-label recognition task anchored on its label structures, we design a novel superimposed CNN and GCN network to extract label structure aware descriptors. Specifically, we first construct two adjacency matrices AS RN N and AK RN N to denote correlation graphs of labels, which is constructed by co-occurrence statistics and a knowledge graph named Concept Net (Speer, Chin, and Havasi 2017) respectively. The initial embedding of all nodes (namely, labels) is extracted from Concept Net. The final adjacency matrix is a superimposed version. Then we apply multi-layer graph convolution on the final superimposed graph to model the label correlation. Besides, different from conventional graph augmented CNN solutions which utilize information of label system at the final recognition stage, we add lateral connections between CNN and GCN at shallow, middle and deep layers to inject information of the label system into backbone CNN for the purpose of labels awareness in feature learning. We have carried out extensive experiments on MS-COCO dataset (Lin et al. 2014) for multi-label image recognition and Charades (Sigurdsson et al. 2016) for multi-label video classification. Results show that our solution obtains absolute m AP improvement of 6.4% and 12.0% in MS-COCO and Charades with very limited computation cost overhead, when compared to its plain CNN counterpart. Our model achieves new state-of-the-art and outperforms current state-of-the-art solution by 1.3% and 2.4% in m AP on MS-COCO and Charades, respectively. Related Work State-of-the-art image or video classification frameworks (He et al. 2016a; Carreira and Zisserman 2017; Feichtenhofer et al. 2018; He et al. 2019; Wu et al. 2019) can be directly applied for multi-label classification by replacing the cross-entropy loss with multi-binary classification loss. The straightforward extension leaves label correlation unexplored thus degrading the recognition performance. We propose our solution to alleviate this problem and it is closely related with the following jobs. Many existing works on multi-label classification proposed to capture label relationship for performance improvement. The co-occurrence of labels can be well formulated by probabilistic graph models, in the literature, there have many methods based on such mathematical theory to model the labels (Li et al. 2016; Li, Zhao, and Guo 2014). To tackle the problem of computation cost burden of probabilistic graph models, the neural network based solution is becoming prevalence recently. In (Wang et al. 2016a), recurrent network was used to encode labels into embedding vectors for label correlation modeling purpose. Context gating strategy was utilized in (Lin, Xiao, and Fan 2018) to integrate the post processing of label re-ranking into the whole network architecture. There are also works done by leveraging the attention mechanism in order for modeling label relationship. In (Wang et al. 2017) and (Zhu et al. 2017a), either image region-level spatial attention map or attentive semantic-level label correlation modeling was used to boost the final recognition performance. (Wang, Jia, and Breckon 2019) proposed to improve the performance by model ensemble. Graph has been proved to be more effective for label structure modeling. Tree-structure label graph built with maximum spanning tree algorithm in (Li, Zhao, and Guo 2014) and knowledge graph for describing label dependency in (Lee et al. 2018) are two typical label graph solutions. Recently, GCN was introduced in (Kipf and Welling 2016) and it has been successfully utilized for non-grid structured data modeling. Researchers have leveraged GCN for many computer vision tasks and great performance was achieved. For instance, it was leveraged in (Yan, Xiong, and Lin 2018; Gao et al. 2018) to model the relationship of skeletons of humans bodies for human action recognition and knowledgeaware GCN was applied for zero-shot video classification in (Gao, Zhang, and Xu 2019). Our work mostly relates to the one proposed in (Chen et al. 2019), which used GCN to propagate information among labels and merges label information with CNN features at the final classification stage. Differently, our work builds GCN by superimposing the Figure 2: The overview of KSSNet with backbone of Inception-I3D. LC is our proposed lateral connection, S and L denote Sigmoid and Leaky Re LU operations, respectively. Inc. is the Inception block in I3D (Carreira and Zisserman 2017). KSSNet takes videos and initial label embeddings as input, and outputs the predicted labels of these videos. GConv is the abbreviation of Graph Convolution . graph built from statistical co-occurrence information into the graph built with knowledge priors. The label information is absorbed into the backbone network for better feature learning. Approach In this paper, We propose a knowledge and label graph superimposing framework for multi-label classification. We provide a new label correlation modeling method of superimposing statistical label graph and knowledge prior oriented label graph. Better feature learning network architecture by absorbing label structure information generated by GCN at shallow, middle and deep layers of backbone CNN is designed. We call our model as KSSNet (Knowledge and Statistics Superimposing Network). Taking the KSSNet with backbone of Inception-I3D (Carreira and Zisserman 2017) designed for multi-label video classification as example, we show its block-diagram in Figure 2. When it comes to multi-label image classification, the framework can be easily constructed by superimposing GCN with stateof-the-art 2D CNN such as Res Net (He et al. 2016a). In the following subsections, we firstly introduce in detail how label graph are constructed and superimposed, and then we show what is our proposed GCN and CNN superimposing. Graph Construction Our final graph is constructed by superimposing statistical label graph into knowledge prior oriented graph. Graph constructed with such statistical information as label cooccurrence frequencies and conditional probabilities of different labels is termed as statistical graph in our paper. Statistical information is determined by the distribution of samples in training set. The statistical graph can be influenced significantly by noise and disturbance. Meanwhile, knowledge graph, such as Concept Net (Speer, Chin, and Havasi 2017), is built with human knowledge by several methods, Figure 3: A subgraph with five nodes on MS-COCO. The number on each edge denotes its weight. Yellow dashed lines with red numbers nearby highlight the redundant edges when taking the threshold of 0.2. such as expert-created resources and games with a purpose. It is more authentic for representing the relationship of labels, especially for small scale datasets. However it has three drawbacks: Firstly, the graph is so dense that it represents too much trivial relationship of nodes. When used into deeper GCNs, it will result in more heavy negative effect of over-smoothed label embeddings, compared with sparse graphs. Secondly, it is datasets independent and neglects the characteristics of specific tasks. Thirdly, as knowledge graph can hardly contain all labels in a dataset, the edges of these labels are lost. Our proposed method combines statistical information and human knowledge, which can overcome their drawbacks to some extent. We formally present its details as follows. A graph is usually denoted as G = (V, E, A), where V, E, A are the set of nodes, set of edges and adjacency matrix. A is an N N matrix with (i, j) entry equaling to the weight of edges between nodes Vi and Vj. N = |V| is the number of vertices. E RN F denotes the feature (label embeddings in our case) matrix for all N nodes. We denote the statistical graph as GS = (V, ES, AS), knowledge graph as GK = (V, EK, AK), where AS and AK are adjacency matrices obtained with statistical information and knowledge priors respectively. AS is constructed by following (Chen et al. 2019). AK is obtained according to the human created knowledge graph Concept Net (Speer, Chin, and Havasi 2017). Specifically, [AK]ij = max{wr|r Sij}, if |Sij| > 0 0, if |Sij| = 0 (1) where Sij is a set of relations (such as used for and is a ) between nodes Vi and Vj extracted from Concept Net. wr is the weight of relation r. |Sij| is the number of elements in Sij. Denoting A S and A K as the normalized versions of AS and AK, respectively. The normalized AS is A S = D 1/2 S ASD 1/2 S , where DS is diagonal and [DS]ii = j[AS]ij. AS is normalized analogously. Weighted average of A S and A K is used to superimpose the prior knowledge into statistical graph and the resulted new adjacency matrix is normalized. A = λA S + (1 λ)A K (2) where λ [0, 1] is a weight coefficient. Meanwhile, as the elements of A S and A K are nonnegative, A has more nonzero elements compared with AS and AK. That is, the graph constructed with A has more redundant edges than GS or GK, as is illustrated in Figure 3. In order to suppress these edges, we use a threshold τ R to filter the elements of A [Aτ]ij = 0, if Aij < τ Aij, if Aij τ (3) As is known to us, when the number of GCN layers increases, the performance of models drops in some tasks. The reason is possibly the over-smoothing of deeper GCN layers (Chen et al. 2019). Inspired by such fact, we further adjust the entries in the adjacency matrix of the superimposed graph and obtain the final matrix AKS: AKS = ηAτ + (1 η)IN (4) where IN is an N N identity matrix. η R is a weight coefficient. With the adjacency matrix AKS, we construct the set of edges as EKS = {(Vi, Vj)|[AKS]ij = 0, and 0 i, j N} (5) (Vi, Vj) denotes the edge (directed or undirected) of nodes Vi and Vj. The graph we proposed is defined as GKS = (V, EKS, AKS), which is called KS graph. Superimposing of GCN and CNN Unlike conventional convolutions, GCN is designed for non Euclidean topological structure. In GCN, the label embeddings of each node is a mixture of the embeddings of its neighbors from the previous layer. We follow a common practice as was done in (Kipf and Welling 2016; Chen et al. 2019) to apply graph convolution. Every GCN layer can be formulated as a non-linear function: E(l+1) = σ(A KSE(l)W (l)), (6) where A KS is the normalized adjacency matrix. E(l) RN C(l) denotes the label embedding at the l-th layer for all N nodes. Note that E(0) is the initial label embeddings and it is extracted from semantic networks like Concept Net (Speer, Chin, and Havasi 2017). W (l) RC(l) C(l+1) is a transformation matrix and is learnable in the training phase. σ( ) denotes a non-linear activation operation. Instead of only superimposing information of label relationship at the final recognition stage, we propose to inject label information into backbone 2D/3D CNNs at different stages by lateral connection (LC operation). Figure 4 shows 2D and 3D versions of our proposed LC operation. Take 3D version for example, we define an LC operation in deep neural networks as: y = g(RN T H W (RT HW C(x) σ(ET ))) + x (7) Here x RC T H W is CNN feature, C is the number of channels. T, H and W denote the frames, height and width of feature tensor. N is the number of labels. E RN C indicates the hidden label embeddings of GCN. g is a 1 1 1 convolution g : RN T H W RC T H W , whose parameters are to be learnt for the downstream tasks. denotes matrix multiplication and ( )T is transpose operation. σ( ) denotes a non-linear activation operation. Both RN T H W ( ) and RT HW C( ) are defined as reshape operations, which rearrange the input array as the shape noted at their subscripts. The motivation of LC is to push the CNN network to learn label-system anchored feature representations for better recognition. As stated in (7), it first calculates crosscorrelation of CNN features and label embeddings and outputs how each CNN feature point is correlated with a label embedding. Such correlation tensor is then mapped to a hidden space by 1x1x1 convolution to encode the relationship of CNN features and label embeddings. At last, the relationship tensor generated from 1 1 1 convolution are added into the original CNN feature tensor. With the lateral connection, the relationship of label system and CNN feature maps is modeled and the learned CNN feature is kind of label-system anchored. Our KSSNet superimposes labels embeddings into CNN features not only in the classification layer but also in hidden layers. There are several advantages of this strategy. (1) The hidden embeddings in GCN can help the feature learning process of CNN, making hidden CNN features aware of label relationship. (2) As for the learning process of hidden embeddings, the extra gradients from LC operation can Table 1: Performance comparisons between baselines and KSSNet on MS-COCO. KSSNet is based on our proposed KS graph and has four GCN layers. Method m AP CP CR CF1 OP OR OF1 CNN-RNN (Wang et al. 2016a) 61.2 - - - - - - SRN (Zhu et al. 2017a) 77.1 81.6 65.4 71.2 82.7 69.9 75.8 Res Net101 (He et al. 2016b) 77.3 80.2 66.7 72.8 83.9 70.8 76.8 Multi-Evidence (Ge, Yang, and Yu 2018) 80.4 70.2 74.9 85.2 72.5 78.4 ML-GCN (Chen et al. 2019) 82.4 84.4 71.4 77.4 85.8 74.5 79.8 KSSNet 83.7 84.6 73.2 77.2 87.8 76.2 81.5 Figure 4: The block diagram of LC operation. R , ( )T , and + denote matrix reshape, transpose, multiplication and sum operations respectively. x(l) and E(l) are CNN feature and GCN feature at the lth GCN layer. The shape of each tensor is marked in gray annotation. be seen as a special regularization, which forces hidden embeddings more adapt to CNN features. It can overcome the over-smoothing of deeper GCN to some extent. Experiment In this section, we conduct experiments to show that our proposed solution can achieve pretty good performance in both image and video multi-label recognition tasks. Then, we carry out ablation studies to evaluate the effectiveness of the proposed graph construction method in our KSSNet. Datasets and Evaluation Metrics MS-COCO MS-COCO (Lin et al. 2014) is a static image dataset, which is widely used for many tasks, such as multilabel image recognition, object localization and semantic segmentation. It contains about 82K images for training, 41K for validation and 41K for test. All images are involved with 80 object labels in the multi-label image recognition task. On average, each image has 2.9 labels. We evaluate all the methods on validation set, since the ground-truth labels of the test set are not available. Charades Charades (Sigurdsson et al. 2016) is a multilabel video dataset containing around 9.8K videos, among which about 8K for training and 1.8K for validation. The average length of videos in Charades is about 30 seconds. It has 157 action labels and 66.5K annotated activities, about 6.8 labels per video. Each action label is composed of a noun (object) and a verb (action). In total, there are 38 different nouns and 33 different verbs. We also evaluate different methos using its validation set. Evaluation Metrics In order for evaluating our model on MS-COCO comprehensively and for convenience of comparison with other solutions, we report the average per-class precision (CP), recall (CR), F1 (CF1), the average overall precision (OP), overall recall (OR), overall F1 (OF1) and mean average precision (m AP), as is done in (Chen et al. 2019). As for the Charades, we evaluate all models with m AP (Sigurdsson et al. 2016) to show their effectiveness. Besides, we also report the value of FLOPs that each model consumes to depict model complexity. Implementation Details Experiment on MS-COCO For image recognition, we choose state-of-the-art Res Net101 (He et al. 2016b) as the backbone of our KSSNet, which is pre-trained on Image Net. The GCN of KSSNet is built from four successive graph convolution layers and the number of channels of their outputs is 256, 512, 1024 and 2048, respectively. In order to deal with the dead Re LU problem, we use Leaky Re LU as activation operation for graph convolution layers, with negative slope of 0.2. Three 2D version LC operations between GCN and the backbone Res Net101 are used and the label embeddings of four graph convolution layers are injected to res2, res3, res4 and res5 of Res Net101. The activation function in the LC operation is set to be Tanh. We adopt 300-dimentional Glo Ve text model (Pennington, Socher, and Manning 2014) to generate the initial label embeddings of labels. As for the labels whose names contain multiple words and have no corresponding keys in Glo Ve, we obtain the label representation by averaging embeddings of all words. In the process of constructing statistical matrix GS, we use the strategy proposed in (Chen et al. 2019). We set λ in (2) to be 0.4, τ in (3) to be 0.02 and η in (4) to be 0.4. During training, the same data preprocessing procedure as (Chen et al. 2019) is adopted. Adam is used as the optimizer with a momentum of 0.9, weight decay of 10 4 and batch size of 80. The initial learning rate of Adam is 0.01. All models are trained for 100 epochs in total. Table 2: Quantitative results of baselines and KSSNet on Charades validation set. The KSSNet bellow has 4 GCN layers and its adjacency matrix is from our proposed KS graph. Method Backbone Modality Pretrain m AP GFLOPs Two-stream (Wu et al. 2018) VGG16 RGB+Flow Image Net,UCF101 14.3 Co Vi AR (Wu et al. 2018) Compressed (Wu et al. 2018) ILSVRC2012-CLS 21.9 Co Vi AR (Wu et al. 2018) Compressed+Flow ILSVRC2012-CLS 24.1 Asyn-TF (Sigurdsson et al. 2017) VGG16 RGB+Flow Image Net 22.4 Multi Scale (TR) (Zhou et al. 2018) Inception-I3D RGB Image Net 25.2 I3D (Carreira and Zisserman 2017) Inception RGB Kinetics-400 32.9 108 Res Net-101(NL) (Wang et al. 2018) Res Net101-I3D RGB Kinetics-400 37.5 544 STRG (NL) (Wang and Gupta 2018) Res Net101-I3D RGB Kinetics-400 39.7 630 Slow Fast (Feichtenhofer et al. 2018) Res Net101 RGB Kinetics-400 42.1 213 Slow Fast(NL) (Feichtenhofer et al. 2018) Res Net101 RGB Kinetics-400 42.5 234 LFB(NL) (Wu et al. 2019) Res Net101-I3D RGB Kinetics-400 42.5 KSSNet Inception-I3D RGB Image Net 44.9 127 Experiment on Charades Inception-I3D of KSSNet is initialized following the inflating mechanism proposed in I3D (Carreira and Zisserman 2017) with BN-Inception pretrained on Image Net. We fine-tune our models using 64frame input clips. These clips are sampled following the strategy of (Wang et al. 2016b), where each clip consists of 64 snippets and each snippet contain only one frame. The spatial size is 224 224, randomly cropped from a scaled video whose spatial size is 256 256. λ, η and τ are set to 0.6, 0.4 and 0.03, respectively. We train all models with mini-batch size of 16 clips. Adam is used as the optimizer, starting with a momentum of 0.9 and weight decay of 10 4. The weight decays of all bias are set to zero. Dropout (Hinton et al. 2012) with a ratio of 0.5 is added after the average pooled CNN features. The initial learning rate of GCN parameters is set to be 0.001, while others are set to be 10 4. We use the strategy proposed in (He et al. 2015) to initialize the GCN and initial label embeddings are extracted with Concept Net (Speer, Chin, and Havasi 2017). During inference, we evenly extract 64 frames from the original fulllength video. Comparison with Baselines In this part, we present comparisons with several state-ofthe-arts on MS-COCO and Charades, respectively to show the effectiveness of our proposed solution. Results on MS-COCO We compare our KSSNet with the state-of-the-art methods, including CNN-RNN (Wang et al. 2016a), SRN (Zhu et al. 2017b), Res Net101 (He et al. 2016b), Multi-Evidence (Ge, Yang, and Yu 2018) and MLGCN (Chen et al. 2019). Table 1 records the quantitative results of all models on MS-COCO validation set. ML-GCN is a GCN+CNN framework based on statistical label graph and it is the current state-of-the-art. It can be observed that our KSSNet obtains the best performance at almost all evaluation matrices. Specially, compared with ML-GCN, its m AP is 1.3% higher, the improvement of overall precision is improved from 85.8% to 87.8%, the gain of overall recall is 1.7% and new state-of-the-art overall F1 score of 81.5% is achieved. The result demonstrates the effectiveness of our KSSNet framework. The comparison of KSSNet and its backbone Res Net101 shows that the absolute improvement in m AP is up to 6.4% and evidences that the label embeddings of GCN can explicitly take advantage of the label relationship, which is hard to be learned by plain CNN or even ignored by many frameworks. Results on Charades Table 2 shows the comparison with state-of-the-art models for our proposed KSSNet on Charades. Compared with backbone I3D model, KSSNet provides 12.0% higher m AP at the cost of very little computation overhead (from 108 GFLOPs to 127 GFLOPs). We can see that the gain is even larger than what achieved on MSCOCO (12.0% v.s. 6.4%). The origin beneath is possibly the characteristics of Charades dataset. On the one hand, each video has 6.8 labels on average, which is even more than MS-COCO. The correlation among different labels has significant influence on multi-label video recognition task. On the other hand, the dataset is not sufficiently large, so the impact of extra label correlation introduced by GCN is more obvious. It can be concluded from such observation that our proposed GCN and CNN superimposing framework can significantly improve baseline result, especially when the training data is not so sufficient. We can also see that although no pretraining on extra large scale video dataset, KSSNet (KS graph) achieves the best performance, which is 2.4% higher than the current state-of-the-art method LFB and Slow Fast(NL) which are pretrained on Kinetics-400. It should be noted that the GFLOPs of our KSSNet is much smaller than Slow Fast(NL), which means KSSNet has remarkable potential in fast multi-label video classification. Ablation Studies In this section, we perform ablation studies to evaluate the effectiveness of our KS graph and to analyze the influence of GCN depth in KSSNet framework. Label graphs of KSSNet In order to evaluate the influence of different graph, we implement three versions of KSSNet with statistical graph, knowledge graph and our proposed KS graph. All of them have the same framework with four GCN layers. Table 3 summarizes the results of KSSNet (statistical graph), KSSNet (knowledge graph) and KSSNet (KS graph). The experiment on MS-COCO shows that knowledge graph performs worse than statistical graph Table 3: Performance comparisons of different label graphs on MS-COCO and Charades. KSSNet (statistical graph) , KSSNet (knowledge graph) and KSSNet (KS graph) are three versions of our proposed KSSNet which have the same framework and different graphs on each dataset. All our variants have four GCN layers. Methods MS-COCO Charades Backbone m AP CP CR CF1 OP OR OF1 Backbone m AP KSSNet (statistical graph) Res Net101 83.1 84.2 73.1 77.6 87.2 76.4 81.2 Inception-I3D 40.7 KSSNet (knowledge graph) Res Net101 81.0 82.8 69.5 75.6 84.5 73.4 78.6 Inception-I3D 41.1 KSSNet (KS graph) Res Net101 83.7 84.6 73.2 77.2 87.8 76.2 81.5 Inception-I3D 44.9 Table 4: Performance of different GCN depths of KSSNet. On each experiment, all versions of KSSNet use KSS graph as adjacency matrix. Methods MS-COCO Charades Backbone m AP #Params Backbone m AP GFLOPs #Params KSSNet (2 layers) Res Net101 82.9 172.1MB Inception-I3D 41.9 110 47.6MB KSSNet (3 layers) Res Net101 83.5 173.2MB Inception-I3D 43.8 115 49.1MB KSSNet (4 layers) Res Net101 83.7 173.8MB Inception-I3D 44.9 127 49.8MB and KS graph, which is caused by the relationship missing of uncovered labels in knowledge graph and by the oversmoothing impact introduced by the presence of many trivial edges. However, the experiment on Charades exhibits a contrary result, KSSNet (knowledge graph) outperforms KSSNet (statistical graph) by a m AP of 0.4. The cause can attribute to the characteristics of Charades. In Charades, labels are more complex and training samples are not so sufficient. Graph constructed from co-occurrence information is not so reliable while knowledge priors are always valid, so the contradiction between complex label relationship modeling and the lack of samples in Charades makes knowledge graph more effective than statistical graph. Both experiments show that our KS graph performs the best, which validates the effectiveness of superimposing statistical graph and knowledge graph. Influence of GCN Depth in KSSNet As we know, conventional GCN suffers from over-smoothing. In this part, we conduct multi-label image and video recognition experiments to demonstrate that our KSSNet framework can deal with this problem effectively. The backbone of KSSNet is Res Net101 and Inception-I3d for MS-COCO and Charades, respectively. In this experiment, we modify the GCN pathway of KSSNet to be with three and two graph convolution layers. The modification can be simply done in two steps: 1) delete the first one or two graph convolution layer(s) and the corresponding LC operation(s) from the GCN pathway; 2) then the first graph convolution layer of the rest ones is adapted to take the initial label embeddings E(0) as input by adjusting its number of input channel C to the channel number of E(0). Experimental results are shown in Table 4. It is obvious, with our KSSNet, more GCN layers lead to better classification results at the cost of small increase of computational cost and model size. KSSNet (3 layers) achieves better performance than KSSNet (2 layers) by absolute m AP improvements of 0.6% and 1.9% in MS-COCO and Charades. KSSNet (4 layers) outperforms KSSNet (3 layers) by 0.2% and 1.1% in m AP. As is reported in ML-GCN (Chen et al. 2019), when GCN has no less than 2 layers, performance of conventional GCN+CNN solution degrades as long as the number of graph convolution layers gets larger. Our model performs on the contrary. This is because that (1) more GCN layers bring more LC operations which guide CNN to learn better label structure aware features at shallow, middle and higher CNN layers. (2) The extra gradients from LC operation can regularize the learning of label embeddings in GCN. (3) We have proposed such strategies as redundant removal to tackle the over-smoothing issue of GCN. Impact of Super-Parameters This experiment is conducted on Charades. When λ varies from 0 to 1 by step of 0.2 and keep other parameters as described above, m AP is 41.05, 41.7, 43.44, 44.93, 44.83 and 41.57. When we fix λ as 0.6, τ varies from 0.01 to 0.04 by step of 0.01, the m AP is 44.6, 44.62, 44.93 and 44.77. Conclusion Capturing label relationship takes a crucial position on multi-label recognition. In order to better model this information, we propose to construct the KS graph for label correlation modeling by superimposing knowledge graph into statistical graph. Then the LC operation is presented for injecting GCN embeddings into CNN features, resulting in a novel neural network KSSNet. LC operation acts as label-feature correlation modeling and helps the model learn label-anchored feature representations. The KSSNet is proven to be capable of learning better feature representations for a specific multi-label recognition task anchored on its label relationship. Experiments on MS-COCO and Charades have demonstrated the effectiveness of our proposed KS graph and KSSNet for both multi-label image and video recognition tasks. Acknowledgement This work was supported by the Joint Laboratory of Intelligent Sports of China Institute of Sport Science (CISS). References Carreira, J., and Zisserman, A. 2017. Quo vadis, action recognition? a new model and the kinetics dataset. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 6299 6308. Chen, Z.-M.; Wei, X.-S.; Wang, P.; and Guo, Y. 2019. Multilabel image recognition with graph convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 5177 5186. Feichtenhofer, C.; Fan, H.; Malik, J.; and He, K. 2018. Slowfast networks for video recognition. ar Xiv preprint ar Xiv:1812.03982. Gao, X.; Hu, W.; Tang, J.; Pan, P.; Liu, J.; and Guo, Z. 2018. Generalized graph convolutional networks for skeleton-based action recognition. ar Xiv preprint ar Xiv:1811.12013. Gao, J.; Zhang, T.; and Xu, C. 2019. I know the relationships: Zero-shot action recognition via two-stream graph convolutional networks and knowledge graphs. In AAAI. Ge, W.; Yang, S.; and Yu, Y. 2018. Multi-evidence filtering and fusion for multi-label classification, object detection and semantic segmentation based on weakly supervised learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1277 1286. He, K.; Zhang, X.; Ren, S.; and Sun, J. 2015. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision, 1026 1034. He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016a. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 770 778. He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016b. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 770 778. He, D.; Zhou, Z.; Gan, C.; Li, F.; Liu, X.; Li, Y.; Wang, L.; and Wen, S. 2019. Stnet: Local and global spatial-temporal modeling for action recognition. In AAAI. Hinton, G. E.; Srivastava, N.; Krizhevsky, A.; Sutskever, I.; and Salakhutdinov, R. R. 2012. Improving neural networks by preventing co-adaptation of feature detectors. ar Xiv preprint ar Xiv:1207.0580. Kipf, T. N., and Welling, M. 2016. Semi-supervised classification with graph convolutional networks. ar Xiv preprint ar Xiv:1609.02907. Lee, C.-W.; Fang, W.; Yeh, C.-K.; and Frank Wang, Y.-C. 2018. Multi-label zero-shot learning with structured knowledge graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1576 1585. Li, Q.; Qiao, M.; Bian, W.; and Tao, D. 2016. Conditional graphical lasso for multi-label image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2977 2986. Li, X.; Zhao, F.; and Guo, Y. 2014. Multi-label image classification with a probabilistic label enhancement model. In UAI, volume 1, 3. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Doll ar, P.; and Zitnick, C. L. 2014. Microsoft coco: Common objects in context. In European conference on computer vision, 740 755. Springer. Lin, R.; Xiao, J.; and Fan, J. 2018. Nextvlad: An efficient neural network to aggregate frame-level features for large-scale video classification. In Proceedings of the European Conference on Computer Vision (ECCV), 0 0. Pennington, J.; Socher, R.; and Manning, C. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 1532 1543. Sigurdsson, G. A.; Varol, G.; Wang, X.; Farhadi, A.; Laptev, I.; and Gupta, A. 2016. Hollywood in homes: Crowdsourcing data collection for activity understanding. In European Conference on Computer Vision. Sigurdsson, G. A.; Divvala, S.; Farhadi, A.; and Gupta, A. 2017. Asynchronous temporal fields for action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 585 594. Speer, R.; Chin, J.; and Havasi, C. 2017. Conceptnet 5.5: An open multilingual graph of general knowledge. In Thirty-First AAAI Conference on Artificial Intelligence. Wang, X., and Gupta, A. 2018. Videos as space-time region graphs. In Proceedings of the European Conference on Computer Vision (ECCV), 399 417. Wang, J.; Yang, Y.; Mao, J.; Huang, Z.; Huang, C.; and Xu, W. 2016a. Cnn-rnn: A unified framework for multi-label image classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2285 2294. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; and Van Gool, L. 2016b. Temporal segment networks: Towards good practices for deep action recognition. In European conference on computer vision, 20 36. Springer. Wang, Z.; Chen, T.; Li, G.; Xu, R.; and Lin, L. 2017. Multi-label image recognition by recurrently discovering attentional regions. In Proceedings of the IEEE international conference on computer vision, 464 472. Wang, X.; Girshick, R.; Gupta, A.; and He, K. 2018. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 7794 7803. Wang, Q.; Jia, N.; and Breckon, T. P. 2019. A baseline for multilabel image classification using an ensemble of deep convolutional neural networks. Wu, C.-Y.; Zaheer, M.; Hu, H.; Manmatha, R.; Smola, A. J.; and Kr ahenb uhl, P. 2018. Compressed video action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 6026 6035. Wu, C.-Y.; Feichtenhofer, C.; Fan, H.; He, K.; Krahenbuhl, P.; and Girshick, R. 2019. Long-term feature banks for detailed video understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 284 293. Yan, S.; Xiong, Y.; and Lin, D. 2018. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Thirty-Second AAAI Conference on Artificial Intelligence. Zhou, B.; Andonian, A.; Oliva, A.; and Torralba, A. 2018. Temporal relational reasoning in videos. In Proceedings of the European Conference on Computer Vision (ECCV), 803 818. Zhu, F.; Li, H.; Ouyang, W.; Yu, N.; and Wang, X. 2017a. Learning spatial regularization with image-level supervisions for multi-label image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 5513 5522. Zhu, F.; Li, H.; Ouyang, W.; Yu, N.; and Wang, X. 2017b. Learning spatial regularization with image-level supervisions for multi-label image classification. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).