# learning_sound_events_from_webly_labeled_data__456d11ad.pdf Learning Sound Events from Webly Labeled Data Anurag Kumar , Ankit Shah , Alexander Hauptmann and Bhiksha Raj Language Technologies Institute, School of Computer Science, Carnegie Mellon University argxkr@gmail.com, {aps1, alex, bhiksha}@cs.cmu.edu In the last couple of years, weakly labeled learning has turned out to be an exciting approach for audio event detection. In this work, we introduce webly labeled learning for sound events which aims to remove human supervision altogether from the learning process. We first develop a method of obtaining labeled audio data from the web (albeit noisy), in which no manual labeling is involved. We then describe methods to efficiently learn from these webly labeled audio recordings. In our proposed system, Webly Net, two deep neural networks co-teach each other to robustly learn from webly labeled data, leading to around 17% relative improvement over the baseline method. The method also involves transfer learning to obtain efficient representations. 1 Introduction As artificial intelligence becomes an increasingly integral part of our life, it is imperative that automated understanding of sounds too gets integrated into everyday systems and devices. Sound event detection and understanding has a wide range of applications [Kumar2018], and hence, in the past few years, the field has received considerable attention in the broader areas of machine learning and audio processing. One long-standing problem in audio event detection (AED) has been the availability of labeled data. Labeling sound events in an audio stream require marking their beginnings and ends. Annotating audio recordings with times of occurrences of events is a laborious and resource intensive task. Weakly-supervised learning for sound events [Kumar and Raj2016] addressed this issue by showing that it is possible to train audio event detectors using weakly labeled data: audio recordings, here, are tagged only with presence or absence of the events as opposed to the time stamp annotations in strongly labeled audio data. Weakly labeled AED has gained significant attention since it was first proposed and has become the preferred and the most promising approach for scaling AED. Several weakly labeled methods have been proposed in last couple of years e.g. [Kumar et al.2018,Chou et al.2018,Mc Fee et al.2018,Lu et al.2018, Xu et al.2017], to mention a few. Weak labeling has enabled the collection of audio-event datasets much larger than before [Gemmeke et al.2017,Fonseca et al.2017]. Moreover, learning from weak labels features in the annual DCASE challenge for sound events detection as well 1. Being able to work with weak labels is, however, only half the story. Even weak labeling, when done manually, becomes challenging on large scale; tagging a large number of audio recordings for a large number of sound classes is non-trivial. Datasets along the lines of Audio Set [Gemmeke et al.2017] are not easy to create and require considerable resources. However, a big advantage of weakly labeled learning is that it opens up the possibility of learning from the data on the web without employing manual annotation, thereby allowing large scale learning without laborious human supervision. The web provides us with a rich resource from which weakly-labeled data could be easily derived. It removes the resource intensive processing of creating the training data manually and opens up the possibility of completely automated training. However, this brings up a new problem the weak labels associated with these recordings, having been automatically obtained through some means, are likely to be noisy. The challenge now extends to being able to learn from weakly and noisily labeled web data. We call such data webly labeled. This paper proposes solutions to learning from webly-labeled data. There have been several works on webly supervised learning of visual objects and concepts [Chen and Gupta2015, Liang et al.2016, Divvala et al.2014]. However, learning sound events from webly labeled data has received little to no attention. The main prior work here is [Kumar and Raj2017], where webly labeled data have been employed; however, to counter the noise in the labels, strongly labeled data is used to provide additional supervision. Needless to say, the strong labels which act as the supporting data are manually obtained. Our objective in this paper is to eliminate human supervision altogether from the learning process by proposing webly supervised learning of sounds. Webly labeled data by default are weakly labeled and hence our proposed methods are designed for weakly labeled audio recordings. Our motivation then is to introduce a learning scheme which can effectively counter additional challenges of webly labeled data. We first present the challenges of webly labeled learning of sounds and then an outline of the proposed system in next section. 1http://dcase.community/ Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) 1.1 Challenges in Webly Labeled Learning Webly labeled learning involves several challenges. The first one is obtaining the webly labeled data itself. The challenge of obtaining quality exemplars from the web has been well documented in several computer vision works [Chen and Gupta2015, Fergus et al.2010, Xia et al.2014, Divvala et al.2014]. This applies for sound events as well and is, in fact, harder due to the complex and intricate ways we describe sounds [Kumar2018]. Often sound related terms are not mentioned in videos and hence text-based retrieval can lead to a much inferior collection of exemplars. Nevertheless, a collection of exemplars for sound events obtained through automated methods will contain incorrect exemplars or label noise. This is a major challenge from the learning perspective. Learning from noisy labels has been a known problem [Fr enay and Verleysen2014] and in recent years effectively training deep learning models from noisy labels has also started to get attention. However, it remains an open problem. Even here most of the work on learning from noisy labels has been in the domain of computer vision. The next challenge is the presence of signal noise. Manual annotations keep in check the overall amount of signal noise in the data. Even if the source of data is the web (e.g Audio Set), manual labeling ensures that the sound is at least audible to most human subjects. However, for webly labeled data the signal noise is unchecked and the sound event, even if present, might be heavily masked by other sounds or noise. Signal noise in the webly labeled data is difficult to quantify and remains an open research topic for future works. In this work, we develop an entire framework to deal with such webly labeled data. We begin by collecting a webly labeled dataset using a video search engine as the source. We then propose a deep learning based system for effectively learning from this webly labeled data. Our primary idea is that two neural networks can co-teach each other to robustly learn from webly labeled data. The two networks use two different views of the data due to which they have different learning capabilities. Since the labels are noisy, we argue that one cannot rely only on the loss with respect to the labels to train the networks. Instead, the agreement between the networks can be used to address this problem. Hence, we introduce a method to factor the agreement between the networks in the learning process. Our system also includes transfer learning to obtain robust feature representations. 2 Webly Labeled Learning of Sounds 2.1 Webly Labeled Training Data Obtaining training audio recordings is the first step in the learning process and is a considerable hard open problem on its own. The most popular approach in webly supervised systems in vision has been text query based retrieval from search engines [Fergus et al.2010, Xia et al.2014, Chen and Gupta2015]. Our approach is along similar lines where we use text queries to retrieve potential exemplars from a video search engine, You Tube. We must first select a vocabulary of sounds terms used to describe sounds. In this paper, we use a subset of 40 sound events from Audio Set, chosen based on several factors. These Figure 1: False Positive for the 5 sound classes with the highest FP. include preciseness in event names and definitions, the quality of metadata-based retrieval of videos from You Tube, the retention of sound hierarchies, and the number of exemplars in Audio Set (larger is better). 2 Obtaining Webly Labeled Data Using only the sound name itself as a text query on You Tube leads to extremely noisy retrieval. [Kumar et al.2017] argued that humans often use the phrase sound of in texts before referring to a sound. Based on this intuition we augment the search query with the phrase sound of . This leads to a dramatic improvement in the retrieval of relevant examples. For example, using sound of dog instead of dog improves the relevant results (sound event actually present in the recordings) by more than 60% in the top 50 retrieved videos. Hence, we use the phrase sound of as the search query for retrieving example recordings of each class. We formed two datasets using the above strategy. The first one referred to as Webly-2k uses top 50 retrieved videos for each class and has around 1,900 audio recordings. The second one, Webly-4k, uses the top 100 retrieved videos for each class and contains around 3,800 recordings. Note that some recordings are retrieved for multiple classes, and hence, the datasets are multi-labeled, similar to Audio Set. Only recordings under 4 minutes duration are considered. Analysis of the Dataset The average duration of the Webly-2k set is around 111 seconds resulting in a total of around 60 hours of data. Webly4k is around 108 hours of audio with an average recording duration of 101 seconds. As mentioned before, label noise is expected in these datasets. To analyze this, we manually verified the positive exemplars of each class and estimated the number of false positives (FP) for each class. Clearly, the larger Webly-4k contains far more noisy labels than Webly-2k. Figure 1 shows the FP counts for 5 classes with the highest false positives. Note that for these classes 30-50% of the examples are wrongly labeled to contain the sound when it is actually not present. However, FP values can also be low for some classes, e.g., Piano and Crowd. Estimating false negatives requires one to manually check all of the recordings for all classes, which makes the task considerably difficult. Even the Audio Set dataset has not been assessed for false negatives (FN) and we also keep FN estimation out of scope of this paper. 2For more details visit github page mentioned in Section 3. Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) Figure 2: Webly Net System: Network 1 (N1) is a deep CNN with first view of data as input. Network 2 (N2) takes in the second view of data obtained through transfer learning. The networks are trained together to co-teach each other. 2.2 Proposed Approach: Webly Net The manual verification in the previous section was done only for analysis; the actual goal is to learn from noisily-labeled Webly-4k (or -2k) directly. Robustly training neural networks with noisy labels remains a very hard problem [Tanaka et al.2018]. Several methods have been proposed, especially in the visual domain [Goldberger and Ben-Reuven2016, Chen and Gupta2015, Liang et al.2016, Reed et al.2014, Tanaka et al.2018,Fr enay and Verleysen2014]. These methods include bootstrapping, self-training, pseudo-labeling and curriculum learning, to mention a few. Another set of approaches try to estimate a noise transition matrix to estimate the distribution of noise in labels [Goldberger and Ben-Reuven2016]. However, estimating the noise transition matrix is not an easy problem as it is dependent on the representation and input features. Conventionally, ensemble learning has also been useful in handling noisy labels [Fr enay and Verleysen2014]. In supervised learning, neural networks are trained on some divergence measure between the output produced by the network and the ground truth label. As the network is trained, the noise in the labels will lead to wrong updates in parameters which can affect the generalization capabilities of the network [Zhang et al.2016a]. Some recent approaches have used the idea of having two networks working together to address this problem [Malach and Shalev-Shwartz2017, Han et al.2018]. [Malach and Shalev-Shwartz2017] gives a when to update rule where networks are updated when they disagree. In [Han et al.2018], networks co-teach each other by sampling instances for training within a minibatch. Our approach is fundamentally based on the idea of training multiple networks together, where the agreement (or disagreement) between the networks are used for improved learning. The method incorporates ideas from co-training and multi-view learning. Multi-view learning methods (e.g., cotraining [Blum and Mitchell1998] [Sun2013]) are primarily semi-supervised learning methods where learners are trained on different views of the data, and the goal is to maximize their agreement on the unlabeled data. Our proposed method exploits this central idea of the agreement between classifiers to address the challenges of webly labeled data. The intuition is as follows: Two (or more) independent classifiers operating on noisily labeled data are likely to agree with the provided label when it is correct. When the given label is incorrect, the classifiers are unlikely to agree with it. They are, however, likely to agree with one another if both of them independently identify the correct label. Hence, the networks can inform each other on the errors they are making and help in filtering out those which are coming from noisy labels, thereby improving the overall robustness of the networks. In contrast to prior work such as [Han et al.2018], our proposed method explicitly ties in the co-teaching of the networks by having a disagreement measure in the loss term. Moreover, in our method, the two networks are operating on different views of the data and hence have different learning abilities. As a result, they will not fall in the degenerate situation where both networks essentially end up learning the same thing. This allows us to combine the classifiers outputs during the prediction phase which further improves the performance. We refer to our overall system as Webly Net. Furthermore, our method is easily extended to more than two networks. The central idea remains the same, a divergence measure captures disagreement measure between any two given pair of networks. Given K networks in the system, K(K 1)/2 pairs of disagreements can be measured. These disagreement measures along with losses with respect to the available ground truths are then used to update the network parameters. Each divergence measure can be appropriately weighed by a scalar α to reflect the weight given to that particular pair of networks. Algorithm 1 outlines this procedure. In this work, we work with only two networks. Figure 2 shows an overview of the proposed method. The two networks Network 1 and Network 2 take as input two different views of the data. The networks are trained jointly by combining their individual loss functions and a third divergence term which explicitly measures the agreement between the two networks. The individual losses provide supervision Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) Algorithm 1 Webly Net system Input: Networks N1 to NK, Representation R1 to RK of audio recordings for different networks and labels Y of the recordings, learning rates η1 to ηK, divergence weights α1 to αK(K 1)/2, number of epochs nepochs Output: Jointly trained networks 1: for n = 1, 2, ....nepochs do 2: for k = 1, 2, ....K do 3: Compute loss, Lk(Nk, Y ) w.r.t label Y for network Nk 4: end for 5: for k = 1, 2, ....K(K 1)/2 do 6: Compute divergence D(Ni(Ri), Nj(Rj)) between each pair of networks 7: Weigh each divergence by its corresponding hyperparameter α 8: end for 9: Combine all loss terms L() and divergence terms D() 10: Update networks based on the combined loss. 11: end for from given labels, and the mutual-agreement provides supervision when the labels are noisy. Two Views of the Data Our primary representation of audio recordings are embeddings provided by Google [Hershey et al.2017]. The embeddings are 128-dimensional quantized vectors for each 1 second of audio. Hence, an audio recording R, in this first view, is represented by a feature matrix X1 RNX128, where N depends on the duration of the audio. The temporal structure of the audio is maintained by stacking the embedding sequentially in X1. The first network N1 is trained on these features. Several methods exist in the literature for generating multiple views from a single view [Sun2013]. In this work, we propose to use multiple non-linear transforms through a neural network to generate the second view (X2) of the data. To this end, we use a network trained on first view, X1, to obtain the second view of the data. This network is trained on another large scale sound events dataset (different from the webly labeled set we work with). One motivation is that, given the noisy nature of webly labeled data, a network trained on a large scale dataset such as Audioset can provide robust feature representations (as in transfer learning approaches [Kumar et al.2018]). We first train a network (N1 with C = 527) on the Audio Set dataset and then use this trained model to obtain feature representations for our webly labeled data. More specifically, the F2 layer is used to obtain 1024-dimensional representations for the audio recordings by averaging the outputs across all 1-second segments. This representation learning through knowledge transfer, as shown empirically later, is significantly useful in webly labeled data where a higher level of signal noise and intra-class variation is expected. Network Architectures: N1 and N2 N1 is trained on the first (X1) audio representations. It is a deep CNN. The layer blocks from B1 to B4 consists of two convolutional layers followed by a max-pooling layer. The number of filters in both convolutional layers of these blocks are, { B1:64, B2:128, B3:256, B4:256 }. The convolutional filters are of size 3 3 in all cases, and the convolution operation is done with a stride of 1. Padding of 1 is also applied to inputs of all convolutional layers. The max-pooling in these blocks are done using a window of size 1 2, moving by the same amount. Layer F1 and F2 are again convolutional layers with 1024 filters of size 1 8 and 1024 filters of size 1 1 respectively. All convolutional layers from B1 to F2 consists includes batch-normalization [Ioffe and Szegedy2015] and Re LU (max(0, x)) activations. The layer represented as C is the segment level output layer. It consists of C filters of size 1 1, where C is the number of classes in the dataset. This layer has a sigmoid activation function. The segment level outputs are pooled through a mapping function in the layer marked as G, to produce the recording level output. We use the average function to perform this mapping. The network architecture N1 achieves state-of-the-art results on Audio Set. In other words, the same architecture with C = 527 (527 classes in Audio Set), achieves the state-of-theart result on Audio Set. Hence, we believe it is a good base network for our webly labeled learning. The network N2 (with X2 as inputs) consists of 3 fully connected hidden layers with 2048, 1024 and 1024 neurons respectively. The output layer contains C number of neurons. A dropout of 0.4 is applied after first and second hidden layers. Re LU activation is used in all hidden layers and sigmoid in the output layer. Training Webly Net Given the multi-label nature of datasets, we first compute the loss with respect to each class. The output layer of both N1 and N2 gives posterior outputs for each class. We use the binary cross-entropy loss, defined with respect to cth class as, l(yc, pc) = yc log(pc) (1 yc) log(1 pc). Here, yc and pc = N(X) are the target and the network output for cth class, respectively. The overall loss function with respect to the target is the mean of losses over all classes, as shown in Eq 1 L(N(X), y) = 1 c=1 l(yc, pc) (1) In the Webly Net system, N1 and N2 co-teach each other through the following loss function L(X1, X2, y) = L(N1(X1), y) + L(N2(X2), y)+ α D(N1(X1), N2(X2)) (2) The first two terms in Eq2 are losses for the two networks with respect to the target. D(N1(X1), N2(X2)) is the divergence measure between the outputs of the two networks. The divergence of opinion between the networks provides additional information beyond the losses with respect to target labels and helps reduce the impact of noisy labels on the learning process. The α term in Eq 2 is a hyperparameter and controls the weight given to the divergence measure in the total loss. This can be set through a grid search and validation. The divergence, D(N1(X1), N2(X2)), between the networks can be measured through a variety of functions. We found that the generalized KL-divergence worked best [Banerjee et al.2005]. Note that, the outputs from the two Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) Method MAP Method MAP WLAT 21.3 Res Net-SPDA 21.9 [Kumar et al.2018] [Zhang et al.2016b] Res Net-Attention 22.0 Res Net-mean pooling 21.8 [Xu et al.2017] [Chou et al.2018] M&mmnet-MS 22.6 Ours N1 22.9 [Chou et al.2018] Table 1: MAP of N1 compared with state-of-the-art on whole Audio Set (527 Sound Events, Training: Balanced Set, Test: Eval Set) networks do not sum to 1. The generalized KL divergence is defined as DKL(x||y) = Pd i=1 xilog( xi yi ) Pd i=1 xi + Pd i=1 yi. DKL(x||y) is non-symmetric and is not a distance measure. We use D(x, y) = DKL(x||y) + DKL(y||x) to measure the divergence between the outputs of the two networks. This measure is symmetric with respect to x and y. If o N1 and o N2 are outputs from N1 and N2 respectively then D(N1(X1), N2(X2)) is D(N1(X1), N2(X2)) = i=1 (o N1 i o N2 i )log o N1 i o N2 i (3) During the inference stage, the prediction from Webly Net is the average of the outputs from the networks N1 and N2. 3 Experiments and Results Webly Net is trained on the Webly-4k and Webly-2k training sets. Audio recordings are represented through X1 and X2 views (Sec. 2.2). To the best of our knowledge, no other work has done an extensive study of webly supervised learning for sound events. Previous work, [Kumar and Raj2017], is computationally not scalable to over 100 hours of data we use in this work. Moreover, it also relies on strongly labeled data in the learning process, which is not available in our case. All recordings from the Eval set of Audio Set are used as the test set. This set contains around 4500 recordings corresponding to our set of 40 sound events. A subset of recordings from the Unbalanced set of Audio Set is used for validation. Furthermore, to compare our webly supervised learning with a manually labeled set, we also create Audio Set-40 training set. Audio Set-40 is obtained from the balanced set of the Audio Set by taking all recordings corresponding to the 40 sound events in our vocabulary with over 4,600 recordings. All experiments are done in Py Torch toolkit. Hyperparameters are tuned using the validation set. The network is trained through Adam optimization [Kingma and Ba2014]. Similar to other works [Gemmeke et al.2017, Kumar et al.2018], we use Average Precision (AP) as the performance metric for each class and then Mean Average Precision (MAP) of all classes as the metric for comparison. Please visit https: //github.com/anuragkr90/webly-labeled-sounds for webly labeled data, codebase and additional analysis. Full Audioset performance. We begin by assessing the performance of N1 (with C=527) on Audio Set. The primary motivation behind this analysis is to show that the architecture of N1 is capable of obtaining state-of-the-art results on a standard well-known weakly labeled dataset. Table 1 shows Methods MAP N1-Self (Baseline) (4k) 38.7 N2-Self (4k) 41.4 Webly Net (4k) 45.3 Methods MAP N1-Self (Baseline) (2k) 38.0 N2-Self (2k) 41.2 Webly Net (2k) 44.0 Methods MAP Methods MAP N1-Self (Baseline) 38.7 N1 (Co-trained) 43.6 N2-Self 41.4 N2 (Co-trained) 43.5 N1-Self and N2-Self (Averaged) 43.5 Webly Net 45.3 N2 [Reed et al.2014] 42.6 Table 2: Upper Tables: Comparison of systems on Webly-4k (L) and Webly-2k (R). Lower: N1 and N2 co-teach each other in Webly Net leading to improvement in their individual performances over training them separately. Results shown on Webly-4k. See Sec. 3.1 comparison with state-of-the-art. Our N1 is able to achieve state-of-art-performance on Audio Set. [Chou et al.2018] reports a slightly better MAP of 23.2 using an ensemble of M&mmnet-MS. Ensemble can improve our performance as well. The performance of N1 on Audio Set shows that it can serve as a good base architecture for our Webly Net system. Hence, it also serves the baseline method for comparison. 3.1 Evaluation of Webly Supervised Learning We first train N1 alone on the webly labeled dataset, and this performance (N1-Self) is taken as the baseline. We also train N2 alone on X2 features (N2-Self) to assess the significance of the second view obtained through knowledge transfer in the webly supervised setting. To compare our method with a noisy label learning method, we apply the well-known approach described in [Reed et al.2014] for training N2. The upper two tables in Table-2 shows results for different systems on Webly-4k and Webly-2k training sets. We observe that Webly Net leads to an absolute improvement of 6.6% (17% relative) over the baseline method on the Webly-4k training set. Moreover, X2 representations from pre-trained Audioset leads to considerable improvement over the baseline performance; around 7% and 8.5% relative improvements on the Webly-4k and Webly-2k training sets respectively. The lower table in Table-2 shows how our proposed system in which both networks co-teach each other leads to an improvement in the performance of the individual networks. First, we see that a simple combination of the two networks (N1-Self and N2-Self (Averaged) also leads to improved results. Once the Webly Net system has been trained, we consider the output from individual N1 (or N2) as the output of the system. These are referred to as N1 (Co-trained) and N2 (Co-trained) respectively in Table 2. We can observe that the performances of both networks are improved by a considerable amount, over 12.7% for N1 (38.7 to 43.6) and over 5% for N2 (41.4 to 43.5). The overall Webly Net system leads to 45.3 MAP. Moreover, the noisy label learning approach from [Reed et al.2014] leads to an improvement in N2 (to 42.6), 1% absolute less than the improvement in N2 obtained with our co-training approach (43.6). This shows that learning together with agreement is a more effective compared to bootstrapping in [Reed et al.2014]. Moreover, our overall system is 2.7% absolute (6.4% relative) better than that obtained through [Reed et al.2014]. Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) Figure 3: AP for sound events on Webly-4k training. Comparison of baseline (N1-Self) and Webly Net System Method MAP Method MAP Method MAP Audio Set-40 54.3 Webly-4k 38.7 Webly-4k 45.3 N1-Self (Clean data) N1-Self Webly Net Table 3: Webly labeled training vs. manual labeling (Audio Set-40) Sounds Webly-4k Webly-2k N1-Self Webly Net N1-Self Webly Net Vehicle 28.1 46.9 34.8 36.4 Singing 51.3 52.7 47.8 50.5 Animal 29.7 35.2 29.0 31.1 Water 51.8 59.6 50.4 51.9 Tools 31.9 36.5 40.7 40.6 Avg. 38.6 46.2 40.5 42.1 Table 4: APs for 5 classes with high label noise in webly sets. Comparison with manual labeling. Table 3 shows comparison of N1 trained on Audio Set-40 (which is manually labeled) with systems trained on Webly-4k set. Note, the test set is the same for all cases, only the training set is changing. A considerable difference of 15.6%, exists between N1 trained on Audio Set-40 and that trained on Webly-4k. Webly Net improves the webly supervised learning, by reducing this gap to 9.0%. Such differences in performances between human-supervised data and non-human supervised webly data has been discussed in computer vision as well. Often, human supervision is hard to beat even by using even orders of magnitude more data [Chen and Gupta2015]. Class specific results. Figure 3 shows the comparison of class-specific result comparison between baseline (N1-Self) and Webly Net. Webly Net improves over the baseline N1Self for most of the classes (32 out of 40). Further, we analyze the performance of five classes with high label noise (Fig. 1). Table 4 shows the AP for these sounds. For all of these events, the improvements are considerable, e.g., around 67% and 19% relative improvements for Vehicle and Animal sounds, respectively. Interestingly, N1 s overall performance goes down for these 5 events as we increase the size of the dataset, from Webly-2k to Webly-4k. A considerable drop in performances are seen for Vehicle and Tools sounds whereas only small improvements Animal and Water sounds are seen. Deep learning methods are expected to improve as we increase the amount of training dataset. However, it is clear that the larger Webly-4k contains too many noisy labels which adversely affects the performances in certain cases. The proposed Webly Net system is able to address this problem. On an average Webly Net gives 7.6% absolute (20% relative) over the baseline method when trained on Webly-4k set. Effect of divergence measure. To ensure that the divergence is playing a role in improving the system, we ran a sanity check experiment with α = 0 in Eq 2. The Webly Net system, in this case, produces a MAP of 43.5, same as the simple combination of the individually trained networks. This is expected as the networks are not tied together anymore and one network will have no impact on the learning of the other. 4 Conclusions Human supervision comes at a considerable cost, and hence we need to build methods which rely on human supervision to the least possible extent. In this paper, we presented webly supervised learning of sound events as the solution. We presented a method for mining web data and then a robust deep learning method to learn from webly labeled data. We showed that our proposed method in which networks co-teach each other leads to a considerable improvement in performance while learning from challenging webly labeled data. The method is extendable to more than two networks and in the future, we aim to explore this for efficient training of deep networks. Furthermore, we also need better methods to mine the web for sound events. This can involve clever natural processing techniques to associate metadata with different sound events and then assigning labels based on that. We keep this as part of our future works. Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) [Banerjee et al., 2005] Arindam Banerjee, Srujana Merugu, Inderjit S Dhillon, and Joydeep Ghosh. Clustering with bregman divergences. Journal of machine learning research, 6(Oct), 2005. [Blum and Mitchell, 1998] Avrim Blum and Tom Mitchell. Combining labeled and unlabeled data with co-training. In Proceedings of the eleventh annual conference on Computational learning theory, pages 92 100. ACM, 1998. [Chen and Gupta, 2015] Xinlei Chen and Abhinav Gupta. Webly supervised learning of convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, 2015. [Chou et al., 2018] Szu-Yu Chou, Jyh-Shing Roger Jang, and Yi Hsuan Yang. Learning to recognize transient sound events using attentional supervision. In IJCAI, pages 3336 3342, 2018. [Divvala et al., 2014] Santosh K Divvala, Ali Farhadi, and Carlos Guestrin. Learning everything about anything: Webly-supervised visual concept learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3270 3277, 2014. [Fergus et al., 2010] Rob Fergus, Li Fei-Fei, Pietro Perona, and Andrew Zisserman. Learning object categories from internet image searches. Proceedings of the IEEE, 98(8):1453 1466, 2010. [Fonseca et al., 2017] Eduardo Fonseca, Jordi Pons Puig, Xavier Favory, Frederic Font Corbera, Dmitry Bogdanov, Andres Ferraro, Sergio Oramas, Alastair Porter, and Xavier Serra. Freesound datasets: a platform for the creation of open audio datasets. In Hu X, Cunningham SJ, Turnbull D, Duan Z, editors. Proceedings of the 18th ISMIR Conference; 2017 oct 23-27; Suzhou, China.[Canada]: International Society for Music Information Retrieval; 2017. p. 486-93., 2017. [Fr enay and Verleysen, 2014] Benoˆıt Fr enay and Michel Verleysen. Classification in the presence of label noise: a survey. IEEE transactions on neural networks and learning systems, 2014. [Gemmeke et al., 2017] Jort F Gemmeke, Daniel PW Ellis, Dylan Freedman, Aren Jansen, Wade Lawrence, R Channing Moore, Manoj Plakal, and Marvin Ritter. Audio set: An ontology and human-labeled dataset for audio events. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 776 780. IEEE, 2017. [Goldberger and Ben-Reuven, 2016] Jacob Goldberger and Ehud Ben-Reuven. Training deep neural-networks using a noise adaptation layer. 2016. [Han et al., 2018] Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and Masashi Sugiyama. Coteaching: robust training deep neural networks with extremely noisy labels. 2018. [Hershey et al., 2017] Shawn Hershey, Sourish Chaudhuri, Daniel PW Ellis, Jort F Gemmeke, Aren Jansen, R Channing Moore, Manoj Plakal, Devin Platt, Rif A Saurous, Bryan Seybold, et al. Cnn architectures for large-scale audio classification. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 131 135. IEEE, 2017. [Ioffe and Szegedy, 2015] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. ar Xiv preprint ar Xiv:1502.03167, 2015. [Kingma and Ba, 2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. [Kumar and Raj, 2016] Anurag Kumar and Bhiksha Raj. Audio event detection using weakly labeled data. In 24th ACM International Conference on Multimedia. ACM Multimedia, 2016. [Kumar and Raj, 2017] Anurag Kumar and Bhiksha Raj. Audio event and scene recognition: A unified approach using strongly and weakly labeled data. In Neural Networks (IJCNN), 2017 International Joint Conference on, pages 3475 3482. IEEE, 2017. [Kumar et al., 2017] Anurag Kumar, Bhiksha Raj, and Ndapandula Nakashole. Discovering sound concepts and acoustic relations in text. submitted IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2017. [Kumar et al., 2018] Anurag Kumar, M. Khadkevich, and C. Fugen. Knowledge transfer from weakly labeled audio using convolutional neural network for sound events and scenes. In Acoustics, Speech and Signal Processing (ICASSP), 2018 IEEE International Conference on. IEEE, 2018. [Kumar, 2018] Anurag Kumar. Acoustic Intelligence in Machines. Ph D thesis, Carnegie Mellon University, 2018. [Liang et al., 2016] Junwei Liang, Lu Jiang, Deyu Meng, and Alexander Hauptmann. Learning to detect concepts from weblylabeled video data. IJCAI, 2016. [Lu et al., 2018] Xugang Lu, Peng Shen, Sheng Li, Yu Tsao, and Hisashi Kawai. Temporal attentive pooling for acoustic event detection. Proc. Interspeech 2018, pages 1354 1357, 2018. [Malach and Shalev-Shwartz, 2017] Eran Malach and Shai Shalev Shwartz. Decoupling when to update from how to update . In Advances in Neural Information Processing Systems, pages 960 970, 2017. [Mc Fee et al., 2018] Brian Mc Fee, Justin Salamon, and Juan Pablo Bello. Adaptive pooling operators for weakly labeled sound event detection. ar Xiv preprint ar Xiv:1804.10070, 2018. [Reed et al., 2014] Scott Reed, Honglak Lee, Dragomir Anguelov, Christian Szegedy, Dumitru Erhan, and Andrew Rabinovich. Training deep neural networks on noisy labels with bootstrapping. ar Xiv preprint ar Xiv:1412.6596, 2014. [Sun, 2013] Shiliang Sun. A survey of multi-view machine learning. Neural Computing and Applications, 23(7-8):2031 2038, 2013. [Tanaka et al., 2018] Daiki Tanaka, Daiki Ikami, Toshihiko Yamasaki, and Kiyoharu Aizawa. Joint optimization framework for learning with noisy labels. ar Xiv preprint ar Xiv:1803.11364, 2018. [Xia et al., 2014] Yan Xia, Xudong Cao, Fang Wen, and Jian Sun. Well begun is half done: Generating high-quality seeds for automatic image dataset construction from web. In European Conference on Computer Vision, pages 387 400. Springer, 2014. [Xu et al., 2017] Yong Xu, Qiuqiang Kong, Qiang Huang, Wenwu Wang, and Mark D Plumbley. Attention and localization based on a deep convolutional recurrent model for weakly supervised audio tagging. ar Xiv preprint ar Xiv:1703.06052, 2017. [Zhang et al., 2016a] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. ar Xiv preprint ar Xiv:1611.03530, 2016. [Zhang et al., 2016b] Han Zhang, Tao Xu, Mohamed Elhoseiny, Xiaolei Huang, Shaoting Zhang, Ahmed Elgammal, and Dimitris Metaxas. Spda-cnn: Unifying semantic part detection and abstraction for fine-grained recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1143 1152, 2016. Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19)