# explanationbased_data_augmentation_for_image_classification__224e751d.pdf Explanation-based Data Augmentation for Image Classification Sandareka Wickramanayake Mong Li Lee Wynne Hsu School of Computing National University of Singapore {sandaw, leeml, whsu}@comp.nus.edu.sg Existing works have generated explanations for deep neural network decisions to provide insights into model behavior. We observe that these explanations can also be used to identify concepts that caused misclassifications. This allows us to understand the possible limitations of the dataset used to train the model, particularly the under-represented regions in the dataset. This work proposes a framework that utilizes concept-based explanations to automatically augment the dataset with new images that can cover these under-represented regions to improve the model performance. The framework is able to use the explanations generated by both interpretable classifiers and post-hoc explanations from black-box classifiers. Experiment results demonstrate that the proposed approach improves the accuracy of classifiers compared to state-of-the-art augmentation strategies. 1 Introduction Machine learning models learn decision boundaries based on the training dataset. Ideally, a training dataset should provide sufficient variations in each class so that the model can learn the correct decision boundaries. However, when there are under-represented regions in the training dataset, the model will learn sub-optimal decision boundaries, leading to misclassification of data points [1, 2]. One way to address this problem is to augment the training dataset. A common data augmentation approach is to apply different transformations to the training samples [3] or to mix existing samples [4, 5]. However, such data augmentations only explore the neighborhood of these samples and may not cover the under-represented regions. Another approach is to use annotations such as part landmarks to obtain images from image resources such as Flickr or Google [6]. This approach may introduce out-of-distribution images and degrade the classifier performance. Identifying the appropriate samples to add to the training dataset is crucial. Here, we propose to use model decision explanations to facilitate the search for samples from image repositories and augment the training dataset with samples from the under-represented regions to enable the model to learn the correct decision boundaries. Figure 1a shows the image of a juvenile Black Tern that has been misclassified as White Breasted Nuthatch. A closer examination of the training dataset reveals that it contains very few images of juvenile Black Terns (5 out of 30 images). These juvenile birds have white heads with black skullcap and black earmark, unlike the adult Black Terns [7] (see Figure 1b). In fact, the juvenile Black Tern in Figure 1a is more similar to White Breasted Nuthatch as shown in Figure 1c, leading to the misclassification. Clearly, there is a need to augment the training data to include more images of juvenile Black Tern in order to learn the correct decision boundary between Black Terns and White Breasted Nuthatch. Methods to explain model decisions include pixel-level attributions [8, 9, 10], and high-level semantic concepts such as coarse-grained saliency maps [11, 12], interpretable basis decomposition [13], linguistic explanations [14, 15] and prototypes [16]. The work in [17] designs a Comprehensive 35th Conference on Neural Information Processing Systems (Neur IPS 2021). (a) Juvenile black tern. (b) Images of adult black tern. (c) Images of white breasted nuthatch. Figure 1: Example of juvenile black tern that is misclassified as white breasted nuthatch. Convolutional Neural Network called CCNN that explains the model decision in terms of word phrases, e.g., the bird in Figure 1a is classified as White Breasted Nuthatch due to the concepts white breast , white belly , long beak . This paper introduces a framework called BRACE (Bette R Accuracy from Concept-based Explanation), which utilizes concept-based explanations to identify candidate samples from image repositories for data augmentation. The key idea is to identify concepts that have led to misclassifications and augment the dataset with images that contain these concepts. We design a utility function that computes the degree of match between the visual features in an image and the set of concepts in the under-represented regions. The proposed approach enables us to select samples with high utility scores and increase the representativeness of the training dataset. Experiment results on multiple datasets demonstrate that BRACE outperforms state-of-the-art augmentation methods. 2 Related Work Data augmentation is a widely used pre-processing step in training deep neural networks. The objective is to inflate the training dataset with label preserving small transformations to avoid overfitting. Techniques include geometric transformations, color space augmentations, kernel filters, mixing images, feature space augmentation, adversarial training, generative adversarial networks, neural style transfer, and meta-learning [3]. However, these augmentation methods only explore the neighborhood of existing images and are not able to add samples from under-represented regions [18]. The work in [6] taps into the large online image repositories to augment the training dataset. Object bounding boxes and part landmarks annotations in the original dataset are used to select relevant and diverse images. Since the images retrieved are typically noisy, [19] trains two models, to select in-distribution samples and to re-label the noisy images in the selected set. Both these works may not increase the representativeness of the dataset as they do not consider the under-represented regions in the dataset when deciding which samples to add. Existing explanation methods for deep neural network based image classifiers include pixel-level attribution methods and high-level semantic explanation methods. Pixel-level attribution methods [8, 9, 10] assign an importance score to each pixel in the image based on the contribution of that pixel towards the model decision. On the other hand, semantic explanation methods explain model decisions in terms of human-friendly concepts and can be categorized into post-hoc methods versus fully interpretable classifiers. Post-hoc methods include Grad-CAM [11], FLEX [15], IBD [13], TCAV [20], and ACE [21]. The first three approaches are image-level explanation mechanisms explaining the model s prediction of the given input. In contrast, TCAV and ACE are class-level explanation mechanisms that explain the characteristics of the entire class. Fully interpretable classifiers such as Proto PNet [16] and CCNN [17] explain model decisions in terms of prototypical parts and word phrases, respectively. The proposed BRACE framework is able to utilize explanations from both post-hoc methods and fully interpretable classifiers to obtain new samples for data augmentation. 3 Proposed Framework Machine learning models often make wrong classifications for classes that have insufficient variance in training data. Adding more diverse samples to the mix will enable the model to improve its accuracy. Intuitively, we want to give more training samples to classes with higher misclassifications so that the model is able to learn the real-world variations. Let C be the set of class labels, and D = {(xi, yi)} be the original training set, where yi C is the class label of image xi. For each class c C, we want to augment D with a set of new samples that is proportional to the number of misclassifications for that class. Given a deep neural network model M, we compute the ratio of misclassifications rc for class c on the validation set Dv as follows: (x,c) Dv I(M(x) = c) where m is the number of validation samples with class label c, and I is a function that returns 1 when its argument is true and 0 otherwise. We will augment D with |Dc| rc number of new samples where Dc D is the set of images with class label c. Let Xc be the set of samples obtained using class c as the keyword query on image repositories. This set is typically noisy, which may degrade model performance. As such, we will assess whether the new image is in the under-represented region and whether it contains concepts that led to the misclassification. We say that x is an under-represented sample if it has visual features of class c, but the model s confidence that x belongs to a different class c = c is high. We compute the under-representation score as follows: β(x, c, c) = fx.fc fx fc e P ( c|x) (2) where fx is the feature vector of x and fc is the average feature vector of all the images in class c, P( c|x) is the predicted probability of x belonging to class c. Feature vectors are extracted from model M. A high β value indicates that x is similar to some images in class c, but the model has classified it as class c. Such samples are more informative, and using them for training may help M generalize better. Further, the sample x should not only possess features of class c but should also contain specific concepts that caused the model M to misclassify it as class c. Adding images with these specific concepts will help the model to better differentiate class c from class c. Let Sc c be the set of concepts that caused M to misclassify images of class c into c, and C = S c. Let (Sc c, x) be the function that computes the degree of match between the visual features in an image x and the concepts in Sc c. The derivation of Sc c and (.) differ depending on the explanation mechanism used (see Section 4). The utility score of a sample x is computed as follows: utility(x) = X c C [β(x, c, c) (Sc c, x))] (3) Then |Dc| rc samples with the highest utility scores are added to the original train dataset. We repeat this process for all the classes in D. We fine-tune the weights of M with the augmented training dataset. Note that adding new samples may result in data imbalance, and this is resolved by using weighted cross-entropy loss during fine-tuning where the weight of class c is given by 1 |Dc|(1+rc). With this, there will be minimal change to the weights of the classes where M already performs well. Algorithm 1 gives the details of the BRACE framework. 4 Derivation of Sc c and (.) In this section, we describe how we derive the concepts that cause the misclassification, Sc c, for two popular explanation mechanisms: the fully interpretable classifier with linguistic explanation and the black-box classifier with post-hoc explanation using heat maps. We also show how to compute the degree of match (.) in these cases. 4.1 Fully Interpretable Classifier with Linguistic Explanation The work in [17] introduces an additional concept layer to the CNN-based architecture to guide the learning of the associations between visual features and word phrases extracted from image descriptions. This Comprehensible Convolutional Neural Network (CCNN) explains its decisions in word phrases corresponding to different visual concepts. The training objective function considers concept uniqueness which encourages each learned concept to correspond to only one word phrase, and mapping consistency which aims to preserve the distance between the learned concept and its corresponding word phrase in a joint embedding space. Together with classification accuracy loss, CCNN is currently the state-of-the-art fully interpretable classifier. Here, we discuss how the proposed BRACE framework can further improve CCNN classification accuracy. Algorithm 1: BRACE input :Classification model M, Original training dataset D, Set of class labels C output :Fine-tuned weights θ θ Weight initialization. Initialize θ with pre-trained weights of M for iter [0, max_iterations] do D D for c C do rc Calculate miss-classification ratio as in Eq. 1. Xc Obtain images given the class name of c as the query term. for x Xc do β(x, c, c) Calculate under-representation score as in Eq. 2. Sc c Derive concepts caused misclassifications. (Sc c, x) Calculate availability of concepts Sc c. end utility(x) = P c C [β(x, c, c) (Sc c, x))] end D D S {|Dc| rc samples with the highest utility scores} end θ finetune_weights(M, D ) end Let V = {v1, ..., vn} be the set of visual features extracted by the concept layer of CCNN for an image x. Each vi V corresponds to a concept described by a word phrase p P = {p1, ..., pn}. CCNN employs a Global Average Pooling (GAP) layer to express its classification decision as a weighted sum of the learned concepts. Suppose the output of the GAP layer is {o1, ..., on} where oi corresponds to concept i in x. Then the contribution of i to the class decision c is given by oi wi,c W, where W Rn C is the weight matrix of the fully connected layer in CCNN. We apply the softmax function to obtain the percentage contributions and select the set of concepts whose contribution is greater than some threshold. We will take the top-5 if there are more than five concepts that satisfy the criteria. Here, we set the threshold to be 1 n. For instance, CCNN classifies the bird in Figure 2 as Nashville Warbler and explains that the concepts "Yellow breast", "Grey crown" and "White eyering" accounted for 47%, 19%, and 14% of the decision, respectively. Predicted class: Nashville Warbler Concepts contributed: Yellow breast (0.47), Grey crown (0.19) , White eyering (0.14) Figure 2: Concepts contributed to CCNN s predictions. Given Xc c, the set of images with class label c that have been misclassified as c, we compute the number of images in Xc c that contains the concept i. The top k concepts with the highest number form the set Sc c. We set k = 3 in our experiments. Recall, each visual feature vi in CCNN concept layer is mapped to a word phrase pi describing a concept, and vi is activated only when the concept occurs in the image. As such, given the set of concepts Sc c, we utilize their corresponding visual feature activations to determine the degree of match of a concept i in an image x as follows: αi = log 1 1 1 + e oi + ϵ (4) where oi is the normalized oi obtained by subtracting the mean activation of the GAP layer. The negative log function gives more weight to an image with one highly activated concept than one with multiple weakly activated concepts. With this, the degree of match between the set Sc c and image x is given by: (Sc c, x) = 4.2 Post-hoc Explanation BRACE requires image-level explanation to identify concepts that caused misclassification. One widely accepted image-level post-hoc explanation mechanism is Grad-CAM [11]. Grad-CAM computes gradients of the predicted class score with respect to the last convolutional layer feature maps. The gradients are averaged channel-wise and multiplied with the respective feature maps to create a coarse-grained saliency map. Since feature maps in the last convolutional layer correspond to semantic concepts [22, 23], we consider the image regions highlighted by Grad CAM to cover semantic concepts. In our implementation, we normalize the pixels in the saliency map to have values between 0 and 1, and regions whose values are greater than the threshold of 0.5 correspond to semantic concepts that are responsible for the model decision. Let Xc c be the set of images whose class label has been misclassified from c to c. Then Sc c is the set of regions in Xc c obtained from the saliency maps of Grad-CAM. The regions in Sc c are passed through M, and the output of the last layer before the classification layer form their visual features, denoted as U. Given an image x, we use a state-of-the-art RCNN model [24] to derive salient region proposals that correspond to an object or object part. These region proposals are passed through M to obtain the corresponding visual features denoted as W. The degree of match between Sc c, and x is given by the similarity between visual features in U and W. (Sc c, x) = X u U zu where zu = max w W log 1 u.w u w + ϵ (6) An image with high (Sc c, x) indicates that it contains many concepts that led the model to misclassify. Such images help the model to learn correct decision boundaries and are selected to augment the training dataset. For each concept u that causes misclassifications, zu indicates the degree of match between u and some region w in x. We take the region with the maximum similarity as it indicates the highest probability of u being present in the image. Using a negative log allows us to choose an image containing at least one high confidence concept that causes misclassification over another image with multiple low confidence concepts. This approach can be utilized when we use BRACE with any explanation method that provides super-pixel-based explanations, e.g., ACE [21] and IBD [13]. 5 Performance Evaluation We carry out experiments to show that BRACE can effectively improve the accuracy of general image classification tasks as well as fine-grained image classification. We also demonstrate the broad applicability of BRACE on a wide range of network architectures such as Res Net[25] and Dense Net[26]. All the codes are implemented using Py Torch, and the experiments are run on NVIDIA Tesla V100 GPUs. The following datasets are used: 1. Caltech UCSD Birds (CUB) [27]. This dataset has 11,788 images of birds belonging to 200 classes. The dataset is divided into a train set of 3994 images, a validation set of 2000 images, and a test set of 5794 images. Each image has ten sentences describing the bird collected by [28]. 2. CUB-Families [29]. This dataset groups the 200 species of birds in CUB into 37 families comprised of multiple bird species. We create under-represented regions by removing some species in the test split of [30] from each family in the training dataset. No species are removed in the validation and test sets. The resultant dataset contains 4585 training images, 2343 validation images, and 4860 test images. 3. Tiny Image Net. Tiny Image Net contains 110,000 images with 200 classes. In the standard dataset split, each class has 500 training images (400 for training the model and 100 for tuning the hyper-parameters) and 50 validation images. We report our results on the validation set. In our experiments, we use the samples collected by [31] for CUB and CUB-Families. Since [31] has not collected samples for Tiny Image Net, we use Flicker API to obtain 250 new samples per class using the class name as the search query. These samples are used by all the methods in our comparative study. The number of samples added to each class c is based on the ratio of misclassifications rc (recall Section 3). To ensure that no two images are highly similar, we remove images similar to the ones in the training image set and duplicates obtained from the online repositories. Since CCNN requires an indicator vector per image for training, we extract 398 and 378 word phrases from the image descriptions of CUB and CUB-Families, respectively, to construct these vectors depicting the concepts present in each image. The union of the indicator vectors in the same class are called the class indicator vectors. For each dataset, we set the number of nodes in the concept layer to be the same as the number of extracted word phrases and train CCNN using the same hyper-parameters settings in [17]. For new samples without image descriptions, we use the corresponding class indicator vector as the indicator vectors. 5.1 Comparison with Data Augmentation Methods We compare BRACE and the following state-of-the-art data augmentation methods: Cut Mix [5]. This mixed-based data augmentation method cuts out one image patch, pastes it on another image, and mixes their labels according to the area proportion. Snap Mix [4]. This is another mixed-based data augmentation method. It combines images similar to Cut Mix but uses the semantic composition of the resultant image to derive the label. WS-DAN [32]. WS-DAN is an attention-based approach that uses attention-cropping and attention-dropping. Attention-cropping refers to cropping the region of the image attended by the model to create a new sample. Attention-dropping refers to erasing the attended region to create a new sample. Part-based [6]. This method obtains samples from image repositories by using object bounding boxes and part landmark annotations. Metaset-based [19]. This method also obtains samples from image repositories. It trains two models to ensure the samples are in-distribution and to correct noisy labels. We train a black-box classifier using Res Net-34 architecture. Table 1 shows the results on CUB, CUB-Families, and Tiny Image Net. We observe that training using the BRACE-augmented datasets consistently achieves the highest accuracy for all datasets. In contrast, Cut-mix and Snap-mix fall short as they only combine existing images without considering the under-represented regions. Though WS-DAN uses the attention weights to determine the regions for cropping/dropping, the samples created may not cover the under-represented regions, leading to a lower accuracy compared to BRACE. For the CUB dataset, Part-based augmentation added a total of 106,000 samples while Metaset-based added 18,000 samples. Despite the large number of new samples, the improvements are lower compared BRACE, indicating the importance of adding only samples that are likely to cover the under-represented regions. Table 1: Performance of Res Net-34 black-box classifier with Grad CAM. Method CUB CUB-Families Tiny Image Net Original dataset 85.5 82.6 76.1 Cut-mix 85.7 85.9 77.0 Snap-mix 87.1 86.3 77.5 WS-DAN 87.2 83.5 76.6 Part-based 86.3* - - Metaset-based 86.8 89.3 73.1 BRACE 87.7 90.0 78.3 *taken from corresponding references. Next, we examine the performance of a fully interpretable classifier using these augmentation techniques. We train a CCNN based on the Res Net-34 architecture. Note that when mixed-based augmentations are used with CCNN, the indicator vector of the original image might not match the resultant image. Hence, we combine the class indicator vectors of mixed classes according to their proportions and apply a threshold to create the indicator vector of the resultant image. Further, CCNN requires image descriptions that are available in the CUB and CUB-Families datasets. We omit the Tiny Image Net dataset as it does not have image descriptions. Table 2 shows the results on CUB and Table 2: Performance of fully interpretable CCNN classifier based on Res Net-34. Method CUB CUB-Families Original dataset 84.3 83.8 Cut-mix 80.6 79.0 Snap-mix 82.4 79.9 WS-DAN 81.6 81.8 Metaset-based 85.1 88.1 BRACE 86.1 88.7 CUB-Families. Again, we see that BRACE-augmented datasets achieve the highest accuracy. The improvement is bigger in CUB-Families, where there are more under-represented regions. 5.2 Comparison of Sample Selection Methods BRACE ranks the samples based on the utility scores and selects those with top scores to augment the training dataset. In this set of experiments, we compare BRACE with the following methods: Random. Here, the samples are selected randomly. Confidence. The samples obtained for each class are passed through the classifier. We rank the samples based on the classifier s scores, and the top scored samples are selected. Core-set [33]. k representative samples are selected using the K-Center-Greedy algorithm. L-loss [34]. The classification loss of a sample is predicted, and samples with the top-k losses are returned. Table 3 shows the classification accuracies for Res Net-101 and Dense Net-161 based classifiers using post-hoc explanation. We see that Brace(utility) has the greatest improvement in accuracy in all datasets. The results for L-loss and Core-set are mixed when compared to Random and Confidence methods. This is because both L-loss and Core-set tend to select more noisy images (e.g., out-ofdistribution images) as compared to Random and Confidence leading to lower performance accuracy. See Figure 4. Table 3: Comparison of classification accuracies using post-hoc explanation. Method CUB CUB-Families Tiny Image Net Dense-161 Res-101 Dense-161 Res-101 Dense-161 Res-101 Original dataset 86.6 87.4 86.5 85.4 80.1 81.3 Core-set 84.8 85.9 87.2 88.6 80.2 77.8 L-loss 85.2 84.8 86.5 88.4 78.3 77.4 Random 86.5 86.4 87.0 88.3 80.3 76.8 Confidence 87.3 87.0 86.5 86.6 80.6 81.3 BRACE(utility) 88.0 89.2 93.0 91.2 81.1 81.7 Table 4: Comparison of classification accuracies using linguistic explanation mechanism. Method CUB CUB-Families Dense-161 Res-101 Dense-161 Res-101 Original dataset 84.5 86.6 85.8 85.7 Core-set 85.0 84.5 87.1 86.1 L-loss 82.0 84.7 84.7 85.5 Random 85.8 87.0 88.6 88.0 Confidence 85.4 86.7 85.8 85.8 BRACE(utility) 86.8 88.4 91.9 92.2 Table 4 shows the classification accuracy when we use the linguistic explanation mechanism. We observe that the proposed BRACE(utility) consistently gives the highest accuracy. We also investigate the performance of BRACE(utility) in the under-represented classes by selecting five classes in the CUB-Families dataset that have the lowest classification accuracy. Table 5 shows that BRACE(utility) significantly improves the classification accuracies of these classes compared to other methods. Note that Confidence has a lower accuracy compared to Random. One possible reason is that the samples selected by Confidence tend to be similar to the training images resulting in model over-fitting. Table 5: Class-level classification accuracies on different categories in CUB-Families. Method Hirundinidae Mimidae Fringillidae Cuculidae Icteridae Original dataset 41.1 46.7 52.3 53.1 55.4 Core-set 54.7 67.6 72.7 56.3 67.7 L-loss 52.8 66.7 68.2 49.2 72.1 Random 67.1 70.2 63.9 63.3 70.8 Confidence 32.9 44.4 56.9 53.1 50.0 BRACE(utility) 72.6 71.1 75.6 67.3 82.2 0 1 2 3 4 5 Iteration Classification Accuracy Baseline Utility Confidence Random Core-set L-loss 0 1 2 3 4 5 Iteration Classification Accuracy Baseline Utility Confidence Random Core-set L-loss (b) CUB-Families Figure 3: Classification accuracy of Dense Net-based CCNN. Category Removed images Samples selected by method Utility Random Confidence Core-set L-loss Hirundinidae Fringillidae Figure 4: Comparison of samples selected by different methods. Figure 3 shows the classification accuracy of the Dense Net-161-based CCNN for the different iterations as we add samples to the training dataset. We see that with just one iteration, BRACE(Utility) is able to achieve higher accuracies compared to the other methods. This indicates that the utility score can identify the under-represented, and hence more informative samples are added to the training dataset, thus enabling the learning of such under-represented class boundaries. This performance gap widens in the CUB-Families dataset. Finally, we visualize the samples selected for the "Hirundinidae" and "Fringillidae" categories in CUB-Families. Figure 4 shows that the samples selected by BRACE(utility) are the most similar to the images in the subcategories that have been removed. This assures us that BRACE(utility) can select samples containing concepts of the under-represented regions. This allows the model to learn the correct boundaries for these under-represented regions, leading to higher accuracies. 5.3 Ablation Study Next, we examine the effects of β and by implementing two variants of BRACE with CCNN: BRACE does not take into account the degree of match with the concepts that caused misclassification. In other words, the utility score is given by P c C β(x, c, c) BRACE β does not consider if a sample falls in the under-represented region and computes the utility score as P c C (Sc c, x). Table 6 shows the classification accuracy on CUB and CUB-Families datasets. We observe that the highest accuracy is achieved when both β and are used in the computation of utility score. This is because β ensures that the samples selected fall into the under-represented regions, while ensures that the concepts that caused misclassifications are present in the selected samples. Further, we see that plays a greater role in improving the classification accuracy compared to β. Table 6: Results of ablation study. Method CUB CUB-Families Dense-161 Res-101 Dense-161 Res-101 CCNN 84.8 86.6 85.8 85.7 BRACE 85.9 87.1 91.2 91.4 BRACE β 86.3 87.6 91.3 91.6 BRACE 86.8 88.4 91.9 92.2 Class Sample images Images removed BRACE BRACE BRACE β Alcedinidae Concepts : blue crown, white neck, red crown Stercorariidae Concepts: large wingspan, black crown, grey wing Figure 5: Comparison of samples selected by variants of BRACE. Figure5 shows samples selected by BRACE variants for two classes in CUB-Families given the top-3 concepts responsible for the misclassifications for each class. We observe that BRACE has selected images from under-represented regions. They either belong to the subcategories that were excluded from the training dataset, have unfamiliar backgrounds, or the discriminative features of the class are hidden due to the pose or occlusions. However, the images selected by BRACE do not include the concepts caused for misclassifications. On the other hand, BRACE β selects images clearly showing one or more concepts that have led to misclassifications, e.g., Alcedinidaes with a white neck and Stercorariidae with a large wingspan. However, some of these images do not fall into under-represented regions. In contrast, BRACE has selected images from the under-represented regions that contain concepts contributed for misclassifications. Such images are informative and help the model to learn correct boundaries, leading to higher accuracies. We compute the "overlapping rate", defined as the intersection over the union of sets of the samples selected by BRACE variants. The overlapping rate of BRACE and BRACE β is 0.09, indicating that the two variants tend to select different image sets. In addition, the overlapping rate between BRACE and BRACE is 0.20, which is lower than the 0.37 overlapping rate between BRACE and BRACE β. This suggests that has a greater influence on the set of images selected by BRACE. 5.4 Generalizability with BRACE A representative dataset facilitates the learning of more generalizable features. As a result, a classifier trained on such a representative dataset can recognize a wide variety of instances of a class even when some of those variations were not part of the training data for that class. We demonstrate the generalizability of BRACE in this final set of experiments. We use Res Net-34 trained on CUB-Families and test on a subset of NAbirds [35]. This subset consists of classes that are found in CUB-Families. These classes have a wide variety of sub-categories that may or may not be present in the CUB-Families. We call this subset NAbirds-Sub. For Res Net-34 trained on Tiny Image Net, we test it on the subset of images from Image Net-V2 [36]. This subset consists of classes that are in Tiny Image Net, and we call it Image Net-V2-Sub. Table 7 shows that BRACE(utility) achieves the highest accuracy, confirming that BRACE enables the classifier to learn features that are generalizable to handle more diverse images. Note that when collecting new images for Tiny Image Net using Ficklr API, we set the API parameters different from those used in [36] to ensure there is no overlap between the images we collect and Image Net-V2. Table 7: Comparison on generalizability. Method NAbirds-Sub Image Net-V2-Sub Original dataset 81.5 54.4 Cut-mix 72.1 37.2 Snap-mix 76.0 42.6 WS-DAN 67.0 56.1 Metaset-based 83.5 42.4 Core-set 83.0 61.9 L-loss 79.7 63.9 Random 82.4 65.1 Confidence 81.5 67.6 BRACE(utility) 84.9 70.0 6 Conclusion In this work, we have examined the potential of utilizing explanation mechanisms to augment training datasets to increase classification accuracy. In contrast to existing data augmentation approaches, we have proposed a framework called BRACE that uses the explanations from model decisions to understand the limitations in the training dataset and select informative images that are in the under-represented regions and contain visual concepts that have led to misclassifications to fine-tune the classifier. The assumption here is that the image repositories contain a sufficiently large number of diverse images. For applications with limited images (e.g., images of rare diseases), the performance of BRACE may be affected. Nevertheless, the experiment results on multiple datasets demonstrate that our proposed solution is able to learn features that are generalizable and improve the classification accuracy over state-of-the-art augmentation methods. Acknowledgement. This research is supported by the National Research Foundation, Singapore under its AI Singapore Programme (Award No: AISG-GC-2019-001, AISG-RP-2018-008). [1] A. Torralba and A. A. Efros, Unbiased look at dataset bias, in CVPR 2011, 2011. [2] R. Taori, A. Dave, V. Shankar, N. Carlini, B. Recht, and L. Schmidt, Measuring robustness to natural distribution shifts in image classification, Neur IPS, 2020. [3] C. Shorten and T. M. Khoshgoftaar, A survey on image data augmentation for deep learning, Journal of Big Data, 2019. [4] X. W. Shaoli Huang and D. Tao, Snapmix: Semantically proportional mixing for augmenting fine-grained data, in AAAI, 2021. [5] S. Yun, D. Han, S. J. Oh, S. Chun, J. Choe, and Y. Yoo, Cutmix: Regularization strategy to train strong classifiers with localizable features, in ICCV, 2019. [6] Z. Xu, S. Huang, Y. Zhang, and D. Tao, Webly-supervised fine-grained visual categorization via deep domain adaptation, IEEE transactions on pattern analysis and machine intelligence, 2016. [7] All about birds. https://www.allaboutbirds.org/guide/Black_Tern. Accessed: 2021-01-06. [8] K. Simonyan, A. Vedaldi, and A. Zisserman, Deep inside convolutional networks: Visualising image classification models and saliency maps, ar Xiv, 2014. [9] A. Binder, S. Bach, G. Montavon, K. R. Muller, and W. Samek, Layer-wise relevance propagation for deep neural network architectures, in Information Science and Applications, pp. 913 922, Springer, 2016. [10] M. Sundararajan, A. Taly, and Q. Yan, Axiomatic attribution for deep networks, in ICML, 2017. [11] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, Grad-cam: Visual explanations from deep networks via gradient-based localization, in ICCV, 2017. [12] B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba, Learning deep features for discriminative localization, in CVPR, 2016. [13] B. Zhou, Y. Sun, D. Bau, and A. Torralba, Interpretable basis decomposition for visual explanation, in ECCV, 2018. [14] L. A. Hendricks, Z. Akata, M. Rohrbach, J. Donahue, B. Schiele, and T. Darrell, Generating visual explanations, in ECCV, 2016. [15] S. Wickramanayake, W. Hsu, and M. L. Lee, Flex: Faithful linguistic explanations for neural net based model decisions, in AAAI, 2019. [16] C. Chen, O. Li, D. Tao, A. Barnett, C. Rudin, and J. K. Su, This looks like that: deep learning for interpretable image recognition, in Neur IPS, 2019. [17] S. Wickramanayake, W. Hsu, and M. L. Lee, Comprehensible convolutional neural networks via guided concept learning, in IJCNN, 2021. [18] T. Tran, T. Pham, G. Carneiro, L. Palmer, and I. Reid, A bayesian data augmentation approach for learning deep models, in Neur IPS, 2017. [19] C. Zhang, Y. Yao, X. Shu, Z. Li, Z. Tang, and Q. Wu, Data-driven meta-set based fine-grained visual recognition, in ACM-MM, 2020. [20] B. Kim, M. Wattenberg, J. Gilmer, C. Cai, J. Wexler, F. Viegas, et al., Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav), in ICML, 2018. [21] A. Ghorbani, J. Wexler, J. Y. Zou, and B. Kim, Towards automatic concept-based explanations, in Neur IPS, 2019. [22] M. D. Zeiler and R. Fergus, Visualizing and understanding convolutional networks, in ECCV, 2014. [23] D. Bau, B. Zhou, A. Khosla, A. Oliva, and A. Torralba, Network dissection: Quantifying interpretability of deep visual representations, in CVPR, 2017. [24] S. Ren, K. He, R. Girshick, and J. Sun, Faster r-cnn: Towards real-time object detection with region proposal networks, Neur IPS, 2015. [25] K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, in CVPR, 2016. [26] G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, Densely connected convolutional networks, in CVPR, 2017. [27] C. Wah, S. Branson, P. Welinder, P. Perona, and S. Belongie, Caltech-uscd birds-200-2011 dataset, California Institute of Technology, 2011. [28] S. Reed, Z. Akata, H. Lee, and B. Schiele, Learning deep representations of fine-grained visual descriptions, in CVPR, 2016. [29] T. Chen, W. Wu, Y. Gao, L. Dong, X. Luo, and L. Lin, Fine-grained representation learning and recognition by exploiting hierarchical semantic embedding, in ACM MM, 2018. [30] Y. Xian, C. H. Lampert, B. Schiele, and Z. Akata, Zero-shot learning a comprehensive evaluation of the good, the bad and the ugly, IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019. [31] C. Zhang, Y. Yao, H. Liu, G.-S. Xie, X. Shu, T. Zhou, Z. Zhang, F. Shen, and Z. Tang, Websupervised network with softly update-drop training for fine-grained visual classification, in AAAI, 2020. [32] T. Hu, H. Qi, Q. Huang, and Y. Lu, See better before looking closer: Weakly supervised data augmentation network for fine-grained visual classification, ar Xiv preprint ar Xiv:1901.09891, 2019. [33] O. Sener and S. Savarese, Active learning for convolutional neural networks: A core-set approach, in ICLR, 2018. [34] D. Yoo and I. S. Kweon, Learning loss for active learning, in CVPR, 2019. [35] G. Van Horn, S. Branson, R. Farrell, S. Haber, J. Barry, P. Ipeirotis, P. Perona, and S. Belongie, Building a bird recognition app and large scale dataset with citizen scientists: The fine print in fine-grained dataset collection, in CVPR, 2015. [36] B. Recht, R. Roelofs, L. Schmidt, and V. Shankar, Do imagenet classifiers generalize to imagenet?, in ICML, 2019.