# guided_dropout__7d5cdda2.pdf The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) Guided Dropout Rohit Keshari, Richa Singh, Mayank Vatsa IIIT-Delhi, India {rohitk, rsingh, mayank}@iiitd.ac.in Dropout is often used in deep neural networks to prevent over-fitting. Conventionally, dropout training invokes random drop of nodes from the hidden layers of a Neural Network. It is our hypothesis that a guided selection of nodes for intelligent dropout can lead to better generalization as compared to the traditional dropout. In this research, we propose guided dropout for training deep neural network which drop nodes by measuring the strength of each node. We also demonstrate that conventional dropout is a specific case of the proposed guided dropout. Experimental evaluation on multiple datasets including MNIST, CIFAR10, CIFAR100, SVHN, and Tiny Image Net demonstrate the efficacy of the proposed guided dropout. Introduction Better than a thousand days of diligent study is one day with a great teacher. Japanese proverb Deep neural network has gained a lot of success in multiple applications. However, due to optimizing millions of parameters, generalization of Deep Neural Networks (DNN) is a challenging task. Multiple regularizations have been proposed in the literature such as l1 norm (Nowlan and Hinton 1992), l2 norm (Nowlan and Hinton 1992), maxnorm (Srivastava et al. 2014), rectifiers (Nair and Hinton 2010), KL-divergence (Hinton, Osindero, and Teh 2006), drop-connect (Wan et al. 2013), and dropout (Hinton et al. 2012), (Srivastava et al. 2014) to regulate the learning process of deep neural networks consisting of a large number of parameters. Among all the regularizers, dropout has been widely used for the generalization of DNNs. Dropout (Hinton et al. 2012), (Srivastava et al. 2014) improves the generalization of neural networks by preventing co-adaptation of feature detectors. The working of dropout is based on the generation of a mask by utilizing Bernoulli and Normal distributions. At every iteration, it generates a random mask with probability (1 θ) for hidden units of the network. (Wang and Manning 2013) have proposed a Gaussian dropout which is a fast approximation of conventional Copyright c 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. dropout. (Kingma, Salimans, and Welling 2015) have proposed variational dropout to reduce the variance of Stochastic Gradients for Variational Bayesian inference (SGVB). They have shown that variational dropout is a generalization of Gaussian dropout where the dropout rates are learned. (Klambauer et al. 2017) have proposed alpha-dropout for Scaled Exponential Linear Unit (SELU) activation function. (Ba and Frey 2013) have proposed standout for a deep belief neural network where, instead of initializing dropout mask using Bernoulli distribution with probability p, they have adapted the dropout probability for each layer of the neural network. In addition to the conventional learning methods of dropout, (Gal and Ghahramani 2016) have utilized the Gaussian process for the deep learning models which allows estimating uncertainty of the function, robustness to over-fitting, and hyper-parameter tuning. They have measured model uncertainty by measuring the first and second moments of their approximate predictive distribution. (Gal, Hron, and Kendall 2017) have proposed Concrete Dropout which is a variant of dropout where concrete distribution has been utilized to generate the dropout mask. They have optimized the probability p via path-wise derivative estimator. In the literature, methods related to dropout have been explored in two aspects: 1) sampling dropout mask from different distributions and maintaining mean of the intermediate input while dropping nodes, and 2) adapting dropout probability. However, if prior information related to nodes of a Neural Network (NN) is available, nodes can be dropped selectively in such a way that generalization of NN is improved. Therefore, in this research, we propose strength parameter to measure the importance of nodes and featuremap for dense NN and CNN, respectively and use it for guiding dropout regularization. Figure 1 illustrates the graphical abstract of the paper guided dropout . For understanding the behavior of strength parameter t, a three hidden layer neural network with 8192 nodes is trained with strength parameter using Equation 2 (details discussed in the next section). After training, the accuracy is evaluated by removing low strength to high strength nodes one by one. The effect on the accuracy can be observed in Figure 2. It shows that removing up to almost 5000 low strength nodes has minimal affect on the network accuracy. Therefore, such nodes are considered to be inactive nodes, lying in the inactive region. Few epochs for strength NN Training with Guided Dropout High strength Medium strength Low strength Strength of nodes have improved NN Training with Dropout Testing Nodes are dropped randomly Guidance in Figure 1: Illustrating of the effect of conventional dropout and proposed guided dropout. In neural network training with dropout, nodes are dropped randomly from the hidden layers. However, in the proposed guided dropout, nodes are dropped based on their strength. (Best viewed in color). In the absence of inactive node, network performance does not affected Number of drop nodes in ascending order Figure 2: A Neural Network NN[8192, 3] is trained with strength parameter on CIFAR10 dataset. The bar graph represents the trained strength of the first layer of the NN. It can be observed that low strength nodes are not contributing in the performance and removing such nodes minimally affect the accuracy. Such nodes are termed as inactive nodes in the inactive region. Similarly, high strength nodes are contributing in the performance and removing such nodes affect the accuracy. Such nodes are termed as active node in the active region. (Best viewed in color). On removing nodes with high strength, the network accuracy reduces aggressively. Such nodes are considered to be active nodes in the active region. Our hypothesis is that in the absence of high strength nodes during training, low strength nodes can improve their strength and contribute to the performance of NN. To achieve this, while training a NN, we drop the high strength nodes in the active region and learn the network with low strength nodes. This is termed as Guided Dropout. As shown in Figure 1, during training the network generalizability is strengthened by nurturing inactive nodes. Once trained, more nodes are contributing towards making predictions thus improving the accuracy. The key contribution of this paper is: Strength parameter is proposed for deep neural networks which is associated with each node. Using this parameter, a novel guided dropout regularization approach is proposed. To the best of our knowledge, this is the first attempt to remove randomness in the mask generation process of dropout. We have also presented that conventional dropout is a special case of guided dropout, and is observed when the concept of active and inactive regions are not considered. Further, experimental and theoretical justifications are also presented to demonstrate that the proposed guided dropout performance is always equal to or better than the conventional dropout. Proposed Guided Dropout In dropout regularization, some of the nodes from the hidden layers are dropped at every epoch with probability (1 θ). Let l be the lth layer of a network, where the value of l ranges from 0 to L, and L is the number of hidden layers in the network. When the value of l is zero, it represents the input layer, i.e. a0 = X. Let the intermediate output of the network be z(l). Mathematically, it can be expressed as: z(l+1) j = w(l+1) j i al i + b(l+1) j a(l+1) j = f(z(l+1) j ) (1) where, i [1, ..., Nin], and j [1, ..., Nout] are in- dex variables for Nin and Nout at the (l + 1)th layer, respectively. f(.) is the RELU activation function. The conventional dropout drops nodes randomly using Bernoulli distribution and is expressed as: ea(l) = r(l) a(l), where ea is the masked output, a(l) is the intermediate output, is the element-wise multiplication, and r(l) is the dropout mask sampled from Bernoulli distribution. While dropping nodes from NN, expected loss E(.) increases which enforces regularization penalty on NN to achieve better generalization (Mianjy, Arora, and Vidal 2018). Introducing strength parameter As shown in Figure 3, in the initial few iterations of training with and without dropout, network performance is almost similar. The effectiveness of dropout can be observed after few iterations of training. Dropping some of the trained nodes may lead to more number of active nodes in the network. Hence, the performance of the network can be improved. Utilizing this observation and the discussion presented with respect to Figure 2 (about active/inactive nodes), we hypothesize that a guided dropout can lead to better generalization of a network. The proposed guided dropout utilizes the strength of nodes for generation of the dropout mask. In the proposed formulation, strength is learned by the network itself via Stochastic Gradient Descent (SGD) optimization. Mathematically, it is expressed as: a(l+1) j = t(l+1) j max 0, w(l+1) j i al i + b(l+1) j (2) where, t(l) is sampled from uniform distribution (assuming all nodes have equal contribution). It can also be used to measure the importance of the feature-map for CNN networks. Therefore, Equation 2 can be rewritten as: a(l+1) = t(l+1) max 0, al W(l+1) + b(l+1) (3) where, is a convolution operation, max(0, .) is a RELU operation1, and al is a three-dimensional feature map (for ease of understanding, subscript has been removed). Strength parameter in matrix decomposition: To understand the behavior of the proposed strength parameter t in a simpler model, let the projection of input x Rd2 on W Rd1 d2 represent label vector y ( Rd1). Matrix W can be linearly compressed using singular value decomposition (SVD), i.e., W = Udiag(t)V T . Here, top few entries of diag(t) can approximate the matrix W (Denton et al. 2014). This concept can be utilized in the proposed guided dropout. Strength parameter in two hidden layers of an NN: In an NN environment, let V Rd2 r and U Rd1 r be the weight matrices of the first and second hidden layers of NN, respectively. The hypothesis class can be represented as h U,V (x) = UV T x (Mianjy, Arora, and Vidal 1In the case of other activation functions, intermediate feature maps might have negative values. Therefore, |t| (mod of t ) can be considered as strength parameter. In this case, t value approaching to zero represents low strength and node associated with low strength can be considered as an inactive node. 2018). In case of Guided Dropout, hidden node is parameterized as h U,V,t(x) = Udiag(t)V T x which is similar to the SVD decomposition where parameter t is learned via backpropagation. Therefore, t can be considered as a strength of a node which is directly proportional to the contribution of the node in NN performance. In this case, the weight update rule for parameters U, V and t on the (s + 1)th batch can be written as: θUsdiag(ts rs)V T s xs ys x T s Vsdiag(ts rs) θx T s Vsdiag(ts rs)U T s y T s Usdiag(ts rs) diag(ts+1) diag(ts) θUsdiag(ts rs)V T s xs ys x T s Vs (4) where, {(xs, ys)}S 1 s=0 is the input data, (1 θ) is the dropout rate, η is the learning rate, and r is the dropout mask. For the initial few iteration, r is initialized with ones. However, after few iterations of training, r is generated using Equation 5. In the proposed algorithm active nodes are dropped in two ways: 1. Guided Dropout (top-k): Select (top-k) nodes (using strength parameter) to drop 2. Guided Dropout (DR): Drop Randomly from the active region. Proposed Guided Dropout (top-k): While dropping (top-k) nodes based on the strength, the mask for the proposed guided dropout can be represented as: rl = tl th, where, th = max N (1 θ) tl (5) max N (1 θ) is defined as the k large elements of t where, (1 θ) is the percentage ratio of nodes needed to be dropped and N is the total number of nodes. The generated mask rl is then utilized in equation ey(l) = r(l) y(l) to drop the nodes. Since the number of dropped nodes are dependent on the total number of nodes N and percentage ratio (1 θ), expected loss can be measured by Eb,x[||y 1 θUdiag(r)V T x||2]. If dropout mask rl drops top k nodes, the expected loss would be maximum with respect to conventional dropping nodes. Therefore, guided dropout (top k) would impose maximum penalty in NN loss. Proposed Guided Dropout (DR): The second way of generating guided dropout mask is to select the nodes from the active region, i.e., nodes are Dropped Randomly (DR) from the active region only. Since the number of inactive nodes are large and have a similar strength; therefore, to find the active or inactive region, number of elements in all the bins2 have been computed. The maximum number of elements among all the bins is considered as the count of inactive nodes fm. Thus, fm and (N fm) are the number of inactive and active nodes, respectively. Here, (1 θ) is the probability for sampling dropout mask for active region nodes using Bernoulli distribution. Probability with respect to the total number of nodes should be reduced to maintain the mean µ in the training phase. Therefore, new probability with respect to N is modified as (1 fm N (1 θ)). For the proposed guided dropout (DR), when the nodes are dropped randomly from the active region, in Equation 4, 1 θ will be modified as 1 fm N (1 θ). We have carefully men- tioned (1 θ) as the percentage ratio for the proposed guided dropout (top k). In case of guided dropout (top k), the generated mask might be fixed until any low strength node can replace the (top k) nodes. On the other hand, for the proposed guided dropout (DR), (1 θ) is the dropout probability. Why Guided Dropout Should Work? Dropout improves the generalization of neural networks by preventing co-adaptation of feature detectors. However, it is our assertion that guidance is essential while dropping nodes from the hidden layers. Guidance can be provided based on the regions where nodes are dropped randomly or top few nodes are dropped from the active region. To understand the generalization of the proposed guided dropout, we have utilized Lemma A.1 from (Mianjy, Arora, and Vidal 2018). In their proposed lemma: Let x Rd2 be distributed according to distribution D with Ex[xx T ] = I. Then, for L(U, V ) := Ex[||y UV T x||2] and f(U, V ) := Eb,x[||y 1 θUdiag(r)V T x||2], it holds that f(U, V ) = L(U, V ) + λ i=1 ||ui||2||vi||2 (6) Furthermore, L(U, V ) = ||W UV T ||2 F , where, diag(r) Rn n, V Rd2 n, U Rd1 n, W Rd1 d2, λ = 1 θ θ . (The proof of this lemma is given in (Mianjy, Arora, and Vidal 2018)). According to the above mentioned lemma, it can be observed that the guided dropout assists NN to have a better generalization. Let r be sampled from Bernoulli distribution at every iteration to avoid overfitting in conventional dropout. In guided dropout, mask r is generated based on the strength value t. For Equation L(U, V ) = ||W Udiag(t)V T ||2 F , high strength nodes can be chosen to form mask r . Therefore, loss Eb,x[||y 1 θUdiag(r )V T x||2] Eb,x[||y 1 θUdiag(r)V T x||2]. In this case, penalty would increase while dropping higher strength nodes. The expected loss would be same only if r = r . If r = r , the regularization imposed by the proposed guided dropout increases in the training process. Hence, optimizing the loss in the train- 2In this case, 100 equally spaced bins are chosen. ing process helps inactive nodes to improve their strength in the absence of higher strength nodes. From Equation 6, the path regularization term λ Pn i=1 ||ui||2||vi||2 regularizes the weights ui and vi of the inactive node such that the increase in loss due to dropping higher strength nodes can be minimized. Thus, the worst case of generalization provided by the proposed guided dropout should be equal to the generalization provided by the conventional dropout. Implementation Details Experiments are performed on a workstation with two 1080Ti GPUs under Py Torch (Paszke et al. 2017) programming platform. The program is distributed on both the GPUs. Number of epoch, learning rate, and batch size are kept as 200, [10 2, ..., 10 5], and 64, respectively for all the experiments. Learning rate is started from 10 2 and is reduced by a factor of 10 at every 50 epochs. For conventional dropout, the best performing results are obtained at 0.2 dropout probability. In the proposed guided dropout, 40 epochs have been used to train the strength parameter. Once the strength parameter is trained, dropout probabilities for guided dropout (DR) are set to 0.2, 0.15, and 0.1 for 60, 50, and 50 epochs, respectively. However, after strength learning, dropout ratio for guided dropout (top-k) are set to [0.2, 0.0, 0.15, 0.0, 0.1, 0.0] for [10, 40, 10, 40, 10, 50] epochs, respectively. Experimental Results and Analysis The proposed method has been evaluated using three experiments: i) guided dropout in neural network, ii) guided dropout in deep network (Res Net18 and Wide Res Net 2810), and iii) case study with small sample size problem. The databases used for evaluation are MNIST, SVHN, CIFAR10, CIFAR100, and Tiny Image Net. The proposed guided dropout is compared with stateof-art methods such as Concrete dropout3(Gal, Hron, and Kendall 2017), Adaptive dropout (Standout)4 (Ba and Frey 2013), Variational dropout5 (Kingma, Salimans, and Welling 2015), and Gaussian dropout5. Alpha-dropout (Klambauer et al. 2017) has also been proposed in literature. However, it is specifically designed for SELU activation function. Therefore, to have a fair comparison, results of the alpha-dropout are not included in Tables. Database and Experimental Protocol Protocol for complete database: Five benchmark databases including MNIST (Le Cun et al. 1998), CIFAR10 (Krizhevsky and Hinton 2009), CIFAR100 (Krizhevsky and Hinton 2009), SVHN (Netzer et al. 2011), and Tiny Image Net (Tiny Image Net 2018) have been used to evaluate the proposed method. MNIST is only used for benchmarking Neural Network with conventional dropout (Srivastava et al. 2014). The MNIST dataset 3https://tinyurl.com/yb5msqrk 4https://tinyurl.com/y8u4kzyq 5https://tinyurl.com/y8yf6vmo Table 1: Test accuracy (%) on CIFAR10 and CIFAR100 (Krizhevsky and Hinton 2009) databases using four different architectures of a three layer Neural Network. (Top two accuracies are in bold). Algorithm CIFAR10 CIFAR100 1024, 3 2048, 3 4096, 3 8192, 3 1024, 3 2048, 3 4096, 3 8192, 3 Without Dropout 58.59 59.48 59.72 59.27 28.86 30.01 30.73 32.02 With Dropout 58.77 59.61 59.62 59.86 31.52 31.63 31.37 31.63 Concrete Dropout 57.38 57.64 57.45 55.28 28.03 29.09 28.91 31.02 Adaptive Dropout 55.05 55.45 56.84 57.01 27.82 28.27 28.62 28.65 Variational Dropout 48.90 52.08 53.48 54.90 17.02 20.64 23.32 24.53 Gaussian Dropout 56.12 56.52 56.94 57.34 27.24 28.34 28.87 29.81 Strength only 58.30 58.92 59.21 59.49 29.66 30.20 30.84 31.12 Proposed Guided Dropout (top-k) 58.75 59.65 59.64 59.92 30.92 31.59 31.34 32.11 Proposed Guided Dropout (DR) 59.84 60.12 60.89 61.32 31.88 32.78 33.01 33.15 Table 2: Test accuracy (%) on SVHN (Netzer et al. 2011) and Tiny Image Net (Tiny Image Net 2018) databases using four different architectures of a three layer Neural Network (NN). (Top two accuracies are in bold). Algorithm SVHN Tiny Image Net 1024, 3 2048, 3 4096, 3 8192, 3 1024, 3 2048, 3 4096, 3 8192, 3 Without Dropout 86.36 86.72 86.82 86.84 12.42 13.74 14.64 15.21 With Dropout 85.98 86.60 86.77 86.79 16.39 14.28 14.69 14.44 Concrete Dropout 83.57 84.34 84.97 85.53 11.98 12.50 12.65 14.85 Adaptive Dropout 77.67 79.68 80.89 81.96 12.41 12.98 13.75 14.17 Variational Dropout 74.28 77.91 80.22 81.52 7.95 10.08 12.91 14.69 Gaussian Dropout 72.46 78.07 80.42 80.74 13.88 15.67 15.76 15.94 Strength only 85.76 85.92 85.91 86.83 12.11 13.52 13.95 14.63 Proposed Guided Dropout (top-k) 86.12 86.57 86.78 86.85 15.47 15.45 15.55 16.01 Proposed Guided Dropout (DR) 87.64 87.92 87.95 87.99 17.59 18.84 18.41 17.74 Gap between training and testing loss Figure 3: Illustrating of training and testing losses at every epoch. On the CIFAR10 dataset, the proposed method is compared with the conventional dropout method. It can be observed that the gap between training and testing loss is minimum in proposed guided dropout. (Best viewed in color). contains 70k grayscale images pertaining to 10 classes (28 28 resolution). The CIFAR10 dataset contains 60k color images belonging to 10 classes (32 32 resolution). The experiments utilize 50k training samples and 10k as the test samples. CIFAR100 has a similar protocol with 100 classes. The protocol for CIFAR100 also has 50k and 10k training-testing split. The SVHN dataset contains 73, 257 training samples and 26, 032 testing samples. Tiny Image Net dataset is a subset of the Image Net dataset with 200 classes. It has images with 64 64 resolution with 100k and 10k samples for training and validation sets, respectively. The test-set label is not publicly available. Therefore, validation-set is treated as test-set for all the experiments on Tiny Image Net. Protocol for small sample size problem: Recent literature has emphasized the importance of deep learning architecture working effectively with small sample size problems (Keshari et al. 2018). Therefore, the effectiveness of the proposed algorithm is tested for small sample size problem as well. The experiments are performed on Tiny Image Net database with three-fold cross validation. From the entire training set, 200, 400..., 1k, 2k, .., 5k samples are randomly chosen to train the network and evaluation is performed on the validation set. Evaluation of Guided Dropout in Dense Neural Network (NN) Architecture To showcase the generalization of the proposed method, training and testing loss at every epoch is shown in Figure 3. First Layer: strength value sorted in descending order 0 1000 2000 3000 4000 5000 6000 7000 8000 Number of node Node Strength NN with Guided Dropout NN with Dropout NN Figure 4: Illustrating the learned strength values of the first hidden layer nodes for the CIFAR10 database with NN[8192, 3]. Strength of nodes is improved by utilizing the proposed guided dropout in comparison to with/without conventional dropout. (Best viewed in color). Table 3: Test accuracy (%) on the MNIST (Le Cun et al. 1998) database using three layer Neural Network (NN). (Top two accuracies are in bold). Algorithm Number of nodes, Layers 1024, 3 2048, 3 4096, 3 8192, 3 Without Dropout 98.44 98.49 98.42 98.41 With Dropout 98.45 98.67 98.50 98.53 Concrete Dropout 98.66 98.60 98.62 98.59 Adaptive Dropout 98.31 98.33 98.34 98.40 Variational Dropout 98.47 98.55 98.58 98.52 Gaussian Dropout 98.35 98.43 98.47 98.44 Strength only 98.42 98.51 98.40 98.46 Proposed Guided Dropout (top-k) 98.52 98.59 98.61 98.68 Proposed Guided Dropout (DR) 98.93 98.82 98.86 98.89 A three layer NN with 8192 nodes at each layer is trained without dropout, with dropout, and with the two proposed guided dropout algorithms top k, and DR. It can be inferred that the proposed guided dropout approaches help to reduce the gap between the training and testing losses. The proposed guided dropout is evaluated on three layer Neural Network (NN) with four different architectures as suggested in (Srivastava et al. 2014). Tables 1 to 3 summarize test accuracies (%) on CIFAR10, CIFAR100, SVHN, Tiny Image Net, and MNIST databases. It can be observed that the proposed guided dropout (DR) performs better than existing dropout methods. In large parameter setting such as three layer NN with 8192 nodes, the proposed guided dropout (top-k) algorithm also shows comparable performance. For NN[1024, 3] architecture, conventional dropout is the second best performing algorithm on CIFAR10, CIFAR100, and Tiny Image Net databases. We have claimed that the strength parameter is an essential element in NN to measure the importance of nodes. Though the number of training parameters are increased but this overhead is less than 0.2% of the total number of parameters of a NN6. Figure 4 represents the learned strength value of the first 6For a NN with three hidden layers of 8192 nodes each, total number of learning parameters is only 8192 8192 + 8192 8192 = 134, 217, 728 and overhead of strength parameter is 24, 576. 1 2 3 4 5 6 7 8 0 Time per epoch (seconds) Figure 5: Illustration of time taken in per epoch. X-axis represents without dropout (1), with dropout (2), concrete dropout (3), adaptive dropout (4), variational dropout (5), Gaussian dropout (6), proposed guided dropout (top-k) (7), and proposed guided dropout (DR) (8) algorithms, respectively. hidden layer of NN[8196, 3]. It can be observed that the conventional dropout improves the strength of hidden layer nodes. However, the strengths are further improved upon by utilizing the proposed guided dropout. For understanding the computational requirements, a NN[8192, 3] has been trained and time taken without dropout, with dropout, concrete dropout, adaptive dropout, variational dropout, Gaussian dropout, proposed guided dropout (top-k), and proposed guided dropout (DR) and the results are reported for one epoch. Figure 5 summarizes the time (in seconds) for these variations, which clearly shows that applying the proposed dropout approach does not increase the time requirement. Evaluation of Guided Dropout in Convolutional Neural Network (CNN) Frameworks The proposed guided dropout is also evaluated on CNN architectures of Res Net18 and Wide-Res Net 28-10. On the same protocol, the proposed guided dropout performance is compared with existing state-of-the-art dropout methods. Table 4 summarizes test accuracies of four bench marking databases. It can be observed that on CIFAR10 (C10), dropout is providing second best performance after the pro- Table 4: Test accuracy (%) on CIFAR10, CIFAR100 (Krizhevsky and Hinton 2009) (in Table written as C10, C100), SVHN (Netzer et al. 2011), and Tiny Image Net (Tiny Image Net 2018) databases using CNN architectures of Res Net18 and Wide-Res Net 28-10. (Top two accuracies are in bold). Algorithm Res Net18 Wide-Res Net 28-10 C10 C100 SVHN Tiny Image Net C10 C100 SVHN Tiny Image Net Without Dropout 93.78 77.01 96.42 61.96 96.21 81.02 96.35 63.57 With Dropout 94.09 75.44 96.66 64.13 96.27 82.49 96.75 64.38 Concrete Dropout 91.33 74.74 92.63 62.95 92.63 75.94 92.79 Adaptive Dropout 90.45 73.26 92.33 61.14 79.04 52.12 90.40 62.15 Variational Dropout 94.01 76.23 96.12 62.75 96.16 80.78 96.68 64.36 Gaussian Dropout 92.34 75.11 95.84 60.33 95.34 79.76 96.02 63.64 Strength only 93.75 76.23 96.34 62.06 95.93 80.79 96.31 64.13 Proposed Guided Dropout (top-k) 94.02 76.98 96.62 64.11 96.22 82.31 96.42 64.32 Proposed Guided Dropout (DR) 94.12 77.52 97.18 64.33 96.89 82.84 97.23 66.02 C10 C100 Tiny Image Net Database Accuracy (%) Without Dropout With Dropout With Guided Dropout Figure 6: Classification accuracies on C10, C100, and Tiny Image Net databases. The performance is measured with Res Net152 CNN architecture without dropout, with traditional dropout, and with the proposed guided dropout. (Best viewed in color). posed algorithm. On CIFAR100 (C100), without dropout is providing second best performance and the proposed guided dropout (DR) is providing the best performance. In case of Wide-Res Net 28-10 which has larger parameter space than Res Net18, conventional dropout consistently performs second best after the proposed guided dropout (DR). It improves the Wide-Res Net 28-10 network performance by 0.62%, 0.35%, 0.48%, and 1.64% on C10, C100, SVHN, and Tiny Image Net databases, respectively. We have also computed the results using Res Net152 CNN architecture on C10, C100, and Tiny Image Net databases. As shown in Figure 6, even with a deeper CNN architecture, the proposed guided dropout performs better than the conventional dropout method. Small Sample Size Problem Avoiding overfitting for small sample size problems is a challenging task. A deep neural network, which has a large Accuracy (%) Number of training samples Figure 7: Results of small sample size experiments: Accuracies on varying training samples of the Tiny Image Net dataset. The performance has been measured on four different dense NN architectures. (Best viewed in color). number of parameters, can easily overfit on small size data. For measuring the generalization performance of models, (Bousquet and Elisseeff 2002) suggested to measure the generalization error of the model by reducing the size of the training dataset. Therefore, we have performed three fold validation along with varying the size of the training data. The experiments are performed with Res Net-18 and four dense neural network architectures. As shown in Figures 7 and 8, with varying training samples of the Tiny Image Net database, the proposed guided dropout (DR) yields higher accuracies compared to the conventional dropout. Discussion and Conclusion Dropout is a widely used regularizer to improve the generalization of neural network. In the dropout based training, a mask is sampled from Bernoulli distribution with Number of training samples Accuracy (%) Without dropout With dropout Guided dropout (top-k) Guided dropout (DR) Figure 8: Classification accuracies obtained with varying the training samples for the Tiny Image Net dataset. The performance is measured with Res Net18 CNN architecture. (Best viewed in color). (1 θ) probability which is used to randomly drop nodes at every iteration. In this research, we propose a guidance based dropout, termed as guided dropout, which drops active nodes with high strength in each iteration, as to force nonactive or low strength nodes to learn discriminative features. During training, in order to minimize the loss, low strength nodes start contributing in the learning process and eventually their strength is improved. The proposed guided dropout has been evaluated using dense neural network architectures and convolutional neural networks. All the experiments utilize benchmark databases and the results showcase the effectiveness of the proposed guided dropout. Acknowledgement R. Keshari is partially supported by Visvesvaraya Ph.D. fellowship. R. Singh and M. Vatsa are partly supported by the Infosys Center of Artificial Intelligence, IIIT Delhi, India. References Ba, J., and Frey, B. 2013. Adaptive dropout for training deep neural networks. In Burges, C. J. C.; Bottou, L.; Welling, M.; Ghahramani, Z.; and Weinberger, K. Q., eds., NIPS. Curran Associates, Inc. 3084 3092. Bousquet, O., and Elisseeff, A. 2002. Stability and generalization. JMLR 2(Mar):499 526. Denton, E. L.; Zaremba, W.; Bruna, J.; Le Cun, Y.; and Fergus, R. 2014. Exploiting linear structure within convolutional networks for efficient evaluation. In NIPS, 1269 1277. Gal, Y., and Ghahramani, Z. 2016. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In ICML, 1050 1059. Gal, Y.; Hron, J.; and Kendall, A. 2017. Concrete dropout. In NIPS, 3584 3593. Hinton, G. E.; Srivastava, N.; Krizhevsky, A.; Sutskever, I.; and Salakhutdinov, R. R. 2012. Improving neural networks by preventing co-adaptation of feature detectors. ar Xiv preprint ar Xiv:1207.0580. Hinton, G. E.; Osindero, S.; and Teh, Y.-W. 2006. A fast learning algorithm for deep belief nets. Neural computation 18(7):1527 1554. Keshari, R.; Vatsa, M.; Singh, R.; and Noore, A. 2018. Learning structure and strength of CNN filters for small sample size training. In CVPR, 9349 9358. Kingma, D. P.; Salimans, T.; and Welling, M. 2015. Variational dropout and the local reparameterization trick. In NIPS, 2575 2583. Klambauer, G.; Unterthiner, T.; Mayr, A.; and Hochreiter, S. 2017. Self-normalizing neural networks. In NIPS, 971 980. Krizhevsky, A., and Hinton, G. 2009. Learning multiple layers of features from tiny images. Technical report. Le Cun, Y.; Bottou, L.; Bengio, Y.; and Haffner, P. 1998. Gradientbased learning applied to document recognition. Proceedings of the IEEE 86(11):2278 2324. Mianjy, P.; Arora, R.; and Vidal, R. 2018. On the implicit bias of dropout. ar Xiv preprint ar Xiv:1806.09777. Nair, V., and Hinton, G. E. 2010. Rectified linear units improve restricted boltzmann machines. In ICML, 807 814. Netzer, Y.; Wang, T.; Coates, A.; Bissacco, A.; Wu, B.; and Ng, A. Y. 2011. Reading digits in natural images with unsupervised feature learning. In NIPS-W, volume 2011, 5. Nowlan, S. J., and Hinton, G. E. 1992. Simplifying neural networks by soft weight-sharing. Neural computation 4(4):473 493. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; De Vito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; and Lerer, A. 2017. Automatic differentiation in pytorch. In NIPS-W. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; and Salakhutdinov, R. 2014. Dropout: A simple way to prevent neural networks from overfitting. JMLR 15(1):1929 1958. Tiny Image Net. 2018. Tiny Image Net tiny imagenet visual recognition challenge. https://tiny-imagenet.herokuapp.com/. Accessed: 14th-May-2018. Wan, L.; Zeiler, M.; Zhang, S.; Le Cun, Y.; and Fergus, R. 2013. Regularization of neural networks using dropconnect. In ICML, 1058 1066. Wang, S., and Manning, C. 2013. Fast dropout training. In ICML, 118 126.