# fewshot_continual_active_learning_by_a_robot__cd08c630.pdf Few-Shot Continual Active Learning by a Robot Ali Ayub University of Waterloo Waterloo, ON N2L3G1, Canada a9ayub@uwaterloo.ca Carter Fendley Capital One New York, NY 10017, USA ccf5164@psu.edu In this paper, we consider a challenging but realistic continual learning problem, Few-Shot Continual Active Learning (Fo CAL), where a CL agent is provided with unlabeled data for a new or a previously learned task in each increment and the agent only has limited labeling budget available. Towards this, we build on the continual learning and active learning literature and develop a framework that can allow a CL agent to continually learn new object classes from a few labeled training examples. Our framework represents each object class using a uniform Gaussian mixture model (GMM) and uses pseudo-rehearsal to mitigate catastrophic forgetting. The framework also uses uncertainty measures on the Gaussian representations of the previously learned classes to find the most informative samples to be labeled in an increment. We evaluate our approach on the CORe-50 dataset and on a real humanoid robot for the object classification task. The results show that our approach not only produces state-of-the-art results on the dataset but also allows a real robot to continually learn unseen objects in a real environment with limited labeling supervision provided by its user1. 1 Introduction Continual learning (CL) [3, 4, 5, 6] has emerged as a popular area of research in recent years because of its limitless real-world applications, such as domestic robots, autonomous cars, etc. Most continual machine learning models [7, 8, 9, 10], however, are developed for constrained task-based continual learning setups, where a CL model continually learns a sequence of tasks, one at a time, with all the data of the current task labeled and available in an increment. Real world systems, particularly autonomous robots, do not have the luxury of getting a large amount of labeled data for each task. In contrast, robots operating in real-world environments mostly have to learn from supervision provided by their users [11, 5, 12]. Human teachers, however, would be unwilling answer a large number of questions or label a large amount of data for the robot. It would therefore be useful for robots to self-supervise their learning, and ask the human teachers to label the most informative training samples from the environment, in each increment. In this paper, we focus on this challenging problem, termed as Few-Shot Continual Active Learning (Fo CAL). One of the main problems faced by continual machine learning models is catastrophic forgetting, in which the CL model forgets the previous learned tasks when learning new knowledge. In recent years, several works in CL have focused on mitigating the catastrophic forgetting problem [13, 3, 6]. Most of these works, however, are developed for the task-based continual learning setup, where the model assumes that all the data for a task is available in an increment and it is fully labeled. These constraints are costly and limit the real-world application of CL models on robots. Active learning has emerged as an area of research in recent years, where machine learning models can choose the most informative samples to be labeled from a large corpus of unlabeled data, thus reducing the labelling 1A preliminary version [1, 2] of this work was presented at workshops in Ro Man 2020 and ICRA 2021. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Figure 1: Our overall framework for Fo CAL. In each increment t, the features extracted for unlabeled objects f(xt; θ) are passed through the acquisition function a(xt, M) to get k most informative samples xt , which are labeled by the oracle. The labeled feature vectors are used to update the GMM representation of the learned classes Y t. Pseudo-rehearsal is used to replay old class data, and the classifier model C(.; W) is trained on the pseudo-samples of the old classes and the labeled feature vectors in the tth increment. effort [14, 15]. Most active learning techniques use uncertainty sampling to request labels for the most uncertain objects [15, 14, 16]. These techniques, however, do not learn continually and thus would suffer from catastrophic forgetting. These issues related to the development of "close-world techniques for continual learning, active learning and open-set recognition have been explored in detail in [17]. In this paper, we consider Fo CAL for the online continual learning scenario for the image classification task. In this setup, a CL model (applied on a robot) receives a small amount of unlabeled image data of objects from the environment in an increment, where the objects can belong to the previously learned classes by the model, or new classes. The model is allowed to get a small number of object samples to be labeled by the user. As the model continues to learn from new training samples, it does not have access to the raw image data of the previously learned objects. Overall, Fo CAL is a combination of multiple challenging problems in machine learning, mainly Few-Shot Class Incremental Learning (FSCIL) [9, 10], Active Learning [14, 18], and online continual learning [19]. To solve Fo CAL, we get inspiration from the continual learning and active learning literature, to develop protocols for continual learning models so that they can actively choose informative samples in an increment. Particularly, we take inspiration from FSCIL literature to develop a new Fo CAL model, in which we learn and preserve the feature representation of the previously learned objects classes by modelling them as Gaussian mixture models. To mitigate catastrophic forgetting, we use pseudo-rehearsal [20] using the samples generated from the Gaussian distributions of the old classes, thus removing the need to store raw data for the classes. Further, to choose most informative samples from an unlabeled set, we use a combination of predictive entropy [21, 18] and viewpoint consistency metrics [14, 16] on the GMM representation of the previously learned classes. We perform extensive evaluations of our proposed approach on the CORe-50 dataset [22], and on a real humanoid robot in an indoor environment. Our approach outperforms state-of-the-art (SOTA) continual learning approaches for Fo CAL on the CORe-50 datast with significant margins. Further, our approach can also be integrated on a humanoid robot, and allow the robot to learn a large number of common household objects over a long period of time with limited supervision provided by the user. Finally, as a part of this work, we also release the object dataset collected by our robot as a benchmark for future evaluations for Fo CAL (available here: https://tinyurl.com/2vuwv8ye). 2 Few-Shot Continual Active Learning We define the Few-Shot Continual Active Learning (Fo CAL) problem as follows: Suppose that an AI agent (e.g. a robot) gets a stream of unlabeled data sets D1 pool, D2 pool, ..., Dt pool, ... over t increments, where Dt pool = {xt i} |Dt pool| i=1 . In each increment, a continual learning model M with parameters Θ can only obtain a small number (kt < |Dt pool|) of samples to be labeled. Given the model M, an acquisition function a(xt, M), where xt Dt pool, is used by the AI agent to find the most informative samples to be labeled in an increment t: xt = argmax Dt poola(xt, M). Therefore, in each increment t, the model M gets trained on small subsets of labeled data Dt = {(xt i , yt i)}|kt| i=1, where yt i Y t represents the class label of xt i and Y t is the set of classes in the t-th increment. Note that unlike most continual learning setups, Y i Y j = , i = j. After training on Dt, the model M is tested to recognize all the encountered classes so far Y 1, Y 2, ..., Y t. The main challenges of Fo CAL are three-fold: (1) avoid catastrophic forgetting, (2) prevent overfitting on the few training samples, (3) efficiently choose most informative samples in each increment. For Fo CAL for the task of object classification, we consider the model M (a CNN) as a composition of a feature extractor f(.; θ) with parameters θ and a classification model with weights W. The feature extractor transforms the input images into a feature space F Rn. The classification model takes the features generated by the feature extractor, and generates an output vector followed by a softmax function to generate multi-class probabilities. In this paper, we use a pre-trained feature extractor, therefore parameters θ are fixed. Thus, we incrementally finetune the classification model on D1, D2, ... and get parameters W 1, W 2, .... In an increment t, we expand the output layer by |Y t| neurons to incorporate new classes. Note that this setup does not alleviate the three challenges of Fo CAL mentioned above. The subsections below describe the main components of our framework (Figure 1) to transform this setup for Fo CAL. 2.1 GMM Based Continual Learning (GBCL) We aim to develop a model that not only helps the system with continual learning but is also motivated by the newness of an object. To accomplish this, we must evaluate how different an incoming object is from previously learned object classes, ideally without any additional supervision. To accomplish this, we consider a clustering-based approach to represent the distribution of object classes. Unlike previous works on clustering-based approaches for continual learning [10, 11] that represent the object classes as mean feature vectors (centroids), we estimate the distribution of the each object class using a uniform Gaussian mixture model (GMM). We believe that representing each class data as a GMM may better represent the true distribution of the data rather than assuming that the distribution is circular. We call our complete algorithm for continually learning GMMs of multiple object classes as GMM based continual learning (GBCL). Once the k feature vectors (Dt) selected by acquisition function (Section 2.2) as most informative samples are labeled by the oracle in increment t, GBCL is applied to learn GMMs for the classes Y t. For each ith feature vector xt i in Dt labeled as yt i, if yt i is a new class never seen by the model before, we initialize a new Gaussian distribution N(xt i, O) for class y with xt i as the mean (centroid) and a zero matrix (O) as the covariance matrix2. Otherwise, if yt i is a known class, we find the probabilities N(xt i|cy 1, σy 1), ..., N(xt i|cy j, σy j ), ..., N(xt i|cy ny, σy ny) for xy i to belong to all the previously learned Gaussian distributions for class y, where ny is the total number of mixture components in the GMM for class y, and cy j and σy j represent the centroid and covariance matrix for the jth mixture component of class y, respectively. If the maximum probability among the calculated probabilities for all the distributions is higher than a pre-defined probability threshold P, xt i is used to update the parameters (centroid and covariance matrix) of the most probable distribution (N(cy j, σy j )) in class y. The updated centroid ˆcy j is calculated as a weighted mean between the previous centroid cy j and xt i: ˆcy j = wy j cy j + xt i wy j + 1 (1) where, wy j is the number of images already clustered in the jth (most probable) Gaussian distribution. The updated covariance matrix ˆσy j is calculated based on the procedure described in [23]): ˆσy j = wy j 1 wy j σy j + wy j 1 wy j 2 (xt i ˆcy j)T (xt i ˆcy j) (2) where, σy j is the previous covariance matrix and (xt i ˆ(c)y j)T (xt i ˆcy j) represents the covariance between xt i and ˆcy j . If, on the other hand, the maximum probability among the calculated probabilities 2We do not describe mixing coefficients here, as they will always be 1/n for a uniform GMM, where n is the number of mixture components. for all the distributions is lower than P, a new Gaussian distribution N(xt i, O) is created for class y with xt i as the centroid and O as the covariance matrix. The result of this process is a set of N t uniform GMMs with parameters ϕ1, ϕ2, ..., ϕNt for N t classes learned up till increment t. Note that instead of using the number of mixture components as a hyperparameter, we use the probability threshold. This way we can start with a simple distribution model for each class (a single mixture component) and add more mixture components only when the new images of the class are too dissimilar from the previous mixture components, and thus cannot be modeled by the GMM. Therefore, the total number of mixture components for each class can be different based on the similarity between the images of the class. In section 2.2, we use the same idea of dissimilarity between an unlabeled image and a GMM to predict most informative samples. 2.1.1 Pseudo-rehearsal and Classifier Training To avoid catastrophic forgetting, we use pseudo-rehearsal [20] to replay the old classes when learning from new data in increment t. For pseudo-rehearsal, we sample the Gaussian distributions in the GMMs of all the previously learned classes to generate a set of pseudo-feature vectors. Note that we also store the total number of images clustered in each Gaussian distribution of the classes (wy j ) during the GMM learning phase (Section 2.1). Therefore, we generate the same number of pseudofeature vectors as the original number of images for each class. After generating the pseudo-feature vectors, the classifier model C(.; W) is trained using the labeled dataset Dt in increment t, and the pseudo-feature vectors of the previous classes. For classification of a test image x, it is first passed through the feature extractor f(x, θ) and then through the classifier C(f(x, θ), W). Softmax function (σ) is then applied on the output to generate class probabilities, and the class y with the maximum probability is predicted as the label for the test image: y = argmaxy σ(W T f(x, θ)). 2.2 Active Learning using GMMs We quantify the novelty of an object in terms of how much the model is uncertain about the object. Unlike most active learning setups [18, 14], in Fo CAL the model does not have access to a training set in each increment for training the model to predict uncertain object classes. Further, even if the model does get trained to predict unknown object classes in the first increment, it will catastrophically forget the criterion of novelty as it continually learns new object classes in the subsequent increments (unknown classes in the first increment become known to the model in the subsequent increments). Therefore, we do not train our model for active learning, and instead use the GMM representations of the previously learned object classes to predict the most uncertain objects. Considering the Fo CAL setup (as described in Section 2), in an increment t, the AI agent gets an unlabeled dataset Dt pool and it must find k < |Dt pool| most informative object samples from the dataset to be labeled. To develop an acquisition function for this, we use a combination of two active learning techniques applied to the GMM representations of the previously learned object classes. First, we use the prediction entropy H[y |xt i] of an object as the acquisition function [21]: H[y |xt i] = y=1 p(y = y|xt i)logp(y = y|xt i) (3) For an unlabeled data point xt i Dt pool, we find the predictive probability of xt i using the GMM representation of the object classes learned in the previous increments. The predictive probability p(xt i|ϕy) for of xt i to belong to the GMM of a class y can be defined as: p(xt i|ϕy) = 1 j=1 N(xt i|cy j, σy j ) (4) Intuitively, if a sample xt i is already learned by the AI agent, then its probability to belong to one of the previously learned class GMMs must be high, and thus entropy for xt i must be low. Therefore, top k samples with the highest entropy can be chosen as the most informative samples. For the second technique, we use the idea of viewpoint consistency used in active learning [14, 16]. The main idea is that if an object is already learned by the agent before, then consistent predictions must be produced for the object under different viewpoints (see Figure 3 in the supplementary file for an example). We again use the GMM representations of the previously learned classes for this acquisition function. Let s consider that there are multiple (l) viewpoints xt 1, ..., xt i, ..., xt l of an unlabeled object j available to the AI agent in increment t. We find the predictive probability of each xt i to belong to the GMM representations of the previously learned classes (equation (6)). Using the maximum predictive probability over all the classes we find the class prediction yt i for each xt i as yt i = argmaxy=1,...,N tp(xt i|ϕy). Next, we count the total number of times each of the N t 1 classes is predicted among the different viewpoints of the object j. Let Sj = {s1 j, ..., sy j, ..., s Nt 1 j } represent the total number of times each class is predicted among the different viewpoints of the object. We normalize set Sj by dividing each sy j by l such that PNt 1 y=1 sy j = 1. We then take the inverse of the maximum value in Sj (1/max Sj), which represents the inverse of the highest percentage of viewpoints of object j that are predicted consistently. Thus, we use this term as the inconsistency score of the object j. We use a combination of predictive entropy and viewpoint consistency to generate the final acquisition function for our framework. Before combining the two metrics, we transform the predictive entropy for multiple viewpoints of the same objects as it is originally designed for individual data samples. For multiple viewpoints of an object j, we find the predictive entropy of all the viewpoints individually and then take an average to get the overall predictive entropy of object j. Therefore, we maximize the following combined function to get the most informative samples in increment t: m=1 H[y |xt m] + (1 δ) 1 max Sj (5) where, δ is a hyperparameter that controls the contribution of predictive entropy and viewpoint consistency towards the overall uncertainty score of an object. Note that the viewpoint consistency can only be applied when the agent has access to multiple viewpoints of the same object. In cases when multiple viewpoints are unavailable, we can use the predictive entropy alone. In our experiments (Section 3), however, we use multiple viewpoints of all the objects. 3 Experiments We first evaluate our approach for Fo CAL on the Core-50 dataset [22] and then using the Pepper robot. We begin by presenting the implementation details and then compare our approach against SOTA continual learning approaches on the CORe-50 dataset. Finally, we present the evaluations of our approach on Pepper. Details about the CORe-50 dataset are in the supplementary file. 3.1 Implementation Details We used the Pytorch deep learning framework [24] for implementation and training of all neural network models. We used Res Net-18 [25] pre-trained on the Image Net dataset [26] as a feature extractor for GBCL. The same pre-trained network was also used in all the other continual learning approaches for a fair comparison. For GBCL, we only used the diagonal entries of the covariance matrix to keep the memory budget from growing drastically. For Fo CAL experiments on the CORe-50 dataset, we develop a new experimental setup, as the standard active learning and continual learning experimental setups are not sufficient to test Fo CAL (see Supplementary File for more details). In this experimental setup for Fo CAL on CORe-50, we combined all 8 training sessions in CORe-50 to generate 400 training object instances. In each increment, we randomly sampled m = 5 object instances (from 400 instances) and allowed the model to learn the label of k = 1 out of 5 object instances, making it a challenging problem to learn from a single object instance in each increment. Once an object was learned by the model, it was removed from the complete set of objects and, thus, was not available to the model in later increments. Hence, we allowed the model to learn all objects in 400 increments with one object learned in each increment. We used the test set accuracy at each increment as the evaluation metric. After training Figure 2: Comparison of our method (red curve) to SOTA approaches in terms of classification accuracy on the Core-50 dataset. Average incremental accuracy is reported in parenthesis. The curves show average and standard deviation of 5 runs with random seeds. Methods Increments FT 23 LWF 23 EWC 24 i Ca RL 18 NCM 17 CWR 21 FLB 19 CBCL 17 GBCL-rand 22 GBCL 15 Table 1: Total number of increments needed by each method to learn 10 classes in CORe-50. in each increment, the model was tested on the complete test set of the CORe-50 dataset. We only report the accuracy for 70 increments for this experiment because all the approaches learn all 10 classes by 70 increments and the final accuracy starts to saturate. For robustness, all the experiments were performed 5 times with random seeds. We report the average and standard deviation of the accuracies. Hyperparameters P and δ were chosen using cross-validation and were set to 0.2 and 0.7, respectively for all increments. For the shallow neural network used for classification in GBCL, we used a linear layer of the same size as the feature vectors generated by the Res Net (512 1). The shallow network was trained for 25 epochs using the cross-entropy loss optimized with stochastic gradient descent (with 0.9 as momentum). A fixed learning rate of 0.01 and minibatches of size 64 were used for training. 3.2 Experiments on the CORe-50 Dataset We compare our approach against 6 continual learning approaches (LWF [27], EWC [4], CWR [22], i Ca RL [3], NCM [11], CBCL [10]), finetuning (FT) and a few-shot batch learning baseline (FLB) [28]. FLB uses a pre-trained CNN to extract feature vectors for images and trains a linear classifier using cross entropy loss. FLB is trained on the complete training data of the previous increments and the current increment. In other words, FLB does not learn continually and therefore should have an advantage over other continual learning approaches (including ours). FT uses the same architecture as FLB but FT is trained only on the data of the current increment. FT suffers from catastrophic forgetting and therefore should produce lower accuracy than other continual learning approaches. To the best of our knowledge, none of the eight approaches can be directly applied to Fo CAL. Hence, for FLB, FT, LWF, EWC and CWR we used softmax-based uncertainty sampling to find most uncertain objects. For the softmax-based uncertainty scores, we found the softmax output of each image for an object instance. Then, we took the average of the maximum probability in the softmax output of each image of the object. The object with the highest average probability score was chosen to be labeled. Because i Ca RL, NCM, and CBCL use centroids for classification, we used the average distance of the images of the new object from the centroids as the uncertainty score. A brief description of all six continual learning approaches is provided in Supplementary File. 3.2.1 Comparison with SOTA Approaches Figure 2 compares our GBCL approach against random sampling and other SOTA approaches in terms of classification accuracy on the fixed test set for 10 classes. Our approach (GBCL) and FLB produce similar results for all 70 increments and GBCL produces slightly better accuracy than FLB in some increments. FLB, however, uses the data in all the previous increments at each new increment, while our approach only uses the data in the current increment and pseudo-feature vectors of the previous classes. We believe the reason for similar results for the two approaches is that GBCL generates and uses pseudo-feature vectors for previous classes to mitigate catastrophic forgetting, Increments 40 80 120 160 200 240 Accuracy (%) 56.7 76.5 84.1 87.5 87.9 88.3 No. of Classes 18 20 20 20 20 20 Table 2: Test set accuracy and number of classes learned by Pepper over 240 increments. and it also learns the most informative samples in each increment. As expected, FT suffers from catastrophic forgetting since it finetunes the network on new data in each increment. Regularization techniques (LWF and EWC) also suffer from catastrophic forgetting and produce the same accuracy as FT. CWR and i Ca RL produce slightly better accuracy, however they also suffer from catastrophic forgetting. NCM produces much better accuracy than the other approaches with the help of centroids to remember past classes and prioritize objects to learn. However, NCM s accuracy is significantly inferior to FLB and GBCL. CBCL produces the best accuracy among the SOTA methods because it learns multiple centroids per class. However, CBCL s accuracy is still lower than GBCL. Note that after 400 increments, FLB produces 89% final accuracy and GBCL produces 88% (1% lower) accuracy. The reason is that after 400 increments both models learned all the training samples in the dataset, and therefore the advantage of informative sampling in GBCL fades away. Further, unlike FLB, GBCL does not have access to all the data in the previous increments, and it suffers from slight forgetting compared to batch learning (FLB). In terms of memory storage, GBCL stores 239 clusters/Gaussian distributions (centroids and covariance matrices) for all the classes. To avoid huge memory storage, we only use the diagonal entries of the covariance matrix for all the experiments. Therefore, GBCL stores 239 2=478 vectors of size 512 in memory after learning over 70 increments on the CORe-50 dataset, which requires only 0.97 MB. In contrast, FLB (batch learning approach) stores all 21000 features vectors for all the classes, which requires 43.08 MB (44 times more than GBCL). For other approaches, FT does not store any data, while LWF, EWC store minimal information, such as the predicted labels for the previous classes and fisher matrix to capture weight importance. CWR stores an extra weight matrix for the classification layer which requires 0.02 MB of storage space. i Ca RL stores 2000 raw images for the previous classes, which requires 393 MB. NCM stores only 10 centroids for the previous classes, which requires 0.02 MB. Finally, CBCL stores 315 centroids which requires 0.64 MB. Note that all of the approaches also store a Res Net-18 model trained on the previous classes which requires 83 MB of space. This analysis shows that GBCL provides the best trade-off between memory storage and the overall performance in comparison with the other approaches. 3.2.2 Comparison with Random Sampling To provide more insight into our approach, we tested our GBCL approach without active learning (GBCL-random). In this case, GBCL gets a random object to be labeled in each increment. GBCLrandom produced significantly lower accuracy than FLB and GBCL for all the increments, especially in earlier increments. These results depict the contribution of our active learning approach to efficiently choose which objects to learn. In comparison with other SOTA approaches, GBCL-random beats all the approaches (except CBCL) in later increments but it is below NCM and CBCL in earlier increments. These results also demonstrate the effectiveness of GBCL (Section 2.1) of our approach to mitigate catastrophic forgetting. 3.2.3 Comparison of Learning Efficiency To further evaluate the efficiency of of our approach for Fo CAL, we report the total number of increments taken by all the approaches to learn all 10 classes (see Table 1). GBCL requires the least number of increments (15) to learn the 10 classes. FLB takes 19 increments to learn all the 10 classes because it uses the softmax based sampling. NCM and i Ca RL also take similar number of increments because they use centroids to calculate the uncertainty scores but do not have access to all the data of the previous increments. CBCL is the closest to our approach and learns all the classes in 17 increments. FT, LWF and EWC all use softmax-based sampling without using any previous data, so they take a large number of increments to learn all 10 classes. Lastly, random sampling takes the most number of increments (except for FT, LWF, and EWC) to learn all the 10 classes. These results clearly show that our approach is the most efficient at finding the unknown objects and learning them in a small number of increments. 3.3 Experiment on the Pepper Robot In this experiment we used the Pepper robot to learn 240 household objects belonging to 20 classes (12 objects per class). We performed the experiment in an indoor lab environment with objects placed at four different locations in the lab with different backgrounds. At each location we placed 5 different objects on a table (Figure 2 (a, b) in the supplementary file). For this setup, the robot first localized the 5 individual objects in the image. We use Retina Net [29] pre-trained on MS COCO dataset [30] for object detection and localization. The robot then captures images of the individual objects from multiple viewpoints (see Figure 2 (c) in the supplementary file for an example). These images are used by the active learning module of GBCL to find the most informative object. The user (experimenter) then provides the class label of the chosen object to the robot using a keyboard. The newly labeled object is then used to update GBCL. Details about our complete system for the Pepper robot are in the Supplementary File. This experiment tried to reproduce some of the challenges that might arise in realistic household environments. For example, a large number of objects were learned at different times, over multiple days (8 weeks), depicting a true lifelong learning robot in real-world environments. Different objects were used in the training and test sets. Moreover, the background of the objects was non-ideal since many parts of the background could be viewed as objects by the object detector. For example, consider Figure 2 in the supplementary file where the air-conditioner and the window could be viewed as objects by the object detector. Objects also varied in size and some were almost transparent. Further, the lighting conditions were not ideal. Some images were taken at night, while some during the day with different variations of sun light coming through the windows in the environment (see Figure 2 and Figure 6 for examples. More examples in the Supplementary File). To evaluate this experiment, we created a test set of 60 objects (3 objects per class) that were different from objects in the training set. We tested Pepper s ability to classify the objects in the test set after 40, 80, 120, 160, 200, and 240 increments of learning new objects. We also report the number of object classes learned by Pepper after each of these increments (Table 2). After only 80 increments, Pepper learned all 20 object classes and achieved 76.5% accuracy on the fixed test set. For the rest of the increments the increase in accuracy was minimal, especially in later increments (120-240), similar to the results on the CORe-50 dataset. The reason is that in the earlier increments Pepper learned objects belonging to a large number of classes quickly which increased the overall accuracy. However, in the later increments Pepper only learned about more objects of the previously learned classes, thus the increase in overall accuracy was minimal. We should also note that even after 240 increments (same as the total number of training objects) Pepper only learned 197 objects. The reason is that the object localization module was not perfect, and it failed to detect many objects (43 objects). 4 Ablation Studies We performed two ablation studies to examine the contribution of different components of our approach and the effect of hyperparameter values on GBCL s performance. This set of experiments were performed on the dataset collected by the Pepper robot (we call it Pepper dataset in this paper). As GBCL s accuracy saturates by 160 increments (Table 2), we performed all of these experiments for 160 increments, with a single object learned in each increment. We report accuracy for increments 0, 40, 60, 80, 120, 140, 160, and the average incremental accuracy. The rest of the experimental setup was the same as in the experiments for the CORe-50 dataset (Section 3.2). For the first ablation study, we investigate the effect of hyperparameter δ on choosing the two acquisition functions (entropy and viewpoint consistency) for active learning. Figure 3 (a) depicts the impact of choosing different values ranging from 0 to 1, for the hyperparameter δ. For the extreme values i.e. 0 and 1, our model chooses only one of the acquisition functions for active learning. Therefore, the model produces the lowest accuracy for δ values 0 and 1. However, there is no significant difference in accuracy when choosing either of the acquisition functions. For other delta values close to 0 or 1, the accuracy increases slightly, but the best accuracy is achieved with δ = 0.7. However, for other values of δ close to 0.7 (such as 0.6), there is not a significant difference in accuracy (only 0.07%), which shows that our approach is not highly sensitive to a large range of possible values of δ. These results also confirm that choosing a combination of the two acquisition functions produces better performance than choosing either of them alone. Figure 3: The effects of varying δ, and the probability threshold P on the classification accuracy on the test set of Pepper dataset. While changing one of the two parameters, the other parameter s values was set to be δ = 0.7, P = 0.2 for the best results For the second ablation study, we investigate the effect of hyperparameter P that determines the number of clusters (or Gaussian distributions) generated for each object class. Figure 3 (b) shows the impact of P on GBCL s performance. For P = 0, GBCL simply stores a single cluster for each class which might not be enough to capture the complex distributions of the object classes. Therefore, for P values close to 0, there is a significant drop in GBCL s accuracy ( 18%). As the P value increases, the model starts to assign more clusters for each object class, and thus it is able to capture the complex distributions of the object data. The average incremental accuracy of the model increase significantly when P = 0.15 with 41 total clusters for all 20 object classes. The best accuracy is achieved for P = 0.2 (83 total clusters), however, for a large range of P values the average incremental accuracy stays within a range of 3%. As the P value continues to increase, the model starts to assign more clusters per class, therefore for higher values of P the model uses more storage space. Further, for P values close to 1, the model starts to recruit more clusters per class with only a few features (or a single feature) assigned to each cluster. Thus, the model starts to resemble more with the batch learning case when all the features of the previously learned classes can be stored. Because of this, we start to see an increase in accuracy again when P = 0.7 (379 clusters, 4.5 times more than the number of clusters stored for P = 0.2). Overall, these results show that our model is not highly sensitive to probability threshold values within a large range, and values close to P = 0.2 provide a nice trade-off between memory storage and model s performance. Finally, note that we found both hyperparameters using cross-validation on the CORe-50 dataset and not on the Pepper dataset. However, even on the Pepper dataset the same hyperparameter values produce the best results. This shows that the chosen hyperparameter values are not dependent on a single dataset. 5 Related Work Most continual learning approaches are designed for class-incremental setting in which a model (often a neural network) learns from training data of different classes in different increments and then evaluated on the test data of all the classes learned so far [3]. Most existing class-incremental learning (CIL) methods avoid catastrophic forgetting by storing a portion of the training samples from previous classes and retraining the model on a mixture of the stored data and new data [3, 31]. However, this approach does not scale as additional data exhausts memory capacity limiting performance in real-world applications. To avoid this problem, some CL approaches use regularization techniques [27, 4]. Although these approaches solve the memory storage issues, their performance is significantly inferior to approaches that store old class data. A few works have been proposed for FSCIL, to continually learn from a few-examples per class using limited memory [10, 9, 32, 33]. However, as mentioned in [17, 34], one significant limitation of all CL approaches is that they require the complete training data of each class to be labeled and available in a single increment, however in real-world robotic applications data is available in a streaming manner (online learning [19]) and it might be mostly unlabeled. For object learning, many researchers have presented active learning techniques using uncertainty sampling [35, 18, 14, 15, 36, 16]. Most of these approaches train deep neural networks with special loss terms such that the networks can predict the most uncertain samples. All of these approaches, however, are trained for batch learning setting and will thus suffer from catastrophic forgetting when attempting to learn continually. Further, active learning approaches can predict unknown classes in the first increment after batch training. However, in a continual learning setting these approaches lose their ability to recognize unknown classes in subsequent increments because the model will assign the unknown classes to the newly learned classes (learned in the previous increments) that were unknown in the first increment. Another related field to active learning is open-set recognition (or out of the distribution detection (OOD)) [37, 38, 38, 39], in which a model that is trained on a set of tasks/classes might face samples belonging previously unknown classes in the test set. Therefore, during the testing/deployment phase, the model must be able to detect which of the samples belong to unknown classes. Although methods for OOD [40, 41] have been developed to detect unknown classes, these models do not train on the newly encountered unknown samples. Mundt et al. [17] present a thorough analysis of previous continual and active learning, and OOD works, and suggest a common viewpoint with open-set recognition acting as a natural interface between continual and active learning. Towards this goal, we present an approach that allows an AI agent to actively select the most informative (unknown) samples from unlabeled data, and continually learn from the actively selected object samples. 6 Conclusion This paper presents and evaluates a novel method for a challenging problem: Few-Shot Continual Active Learning (Fo CAL) for object classification. Experimental results demonstrate that our approach for Fo CAL is highly efficient and helps the model learn the most uncertain objects continually without forgetting earlier classes. Our approach not only outperforms the SOTA approaches on a benchmark dataset but also allows a robot to actively learn objects in a real environment over a long period of time. Finally, we have also released the dataset collected by the Pepper robot as a benchmark for future evaluations of Fo CAL models. The work presented here has a few limitations. 1) The experimental setup is still quite simple compared to a real household environment. 2) We assume that correct object labels are provided by the human assistant to the robot. 3) We did not recruit real participants to teach the robot, but the experimenter acted as a user for the robot. 4) We used a fixed feature extractor and object detector. In the future, we hope to work on these limitations and test our system with real human participants. Particularly, we hope to develop techniques that can allow feature representations to be learned from a few samples over a large number of increments. There are a number of positive and negative long-term outcomes of this work. With respect to positive outcomes, the improved performance for continual learning by our approach may one day lead to robots that can tailor their behavior and learning to the needs of a person. We envision a home service robot [42] that uses continual learning to learn a person s food preferences before serving them breakfast, or a domestic service robot that learns how a person prefers their home to be cleaned and adjusts its behavior to their preferences. In terms of negative outcomes, one potential downside is that an active learning robot might learn unnecessary objects which are never going to be used by the robot. Learning such objects can potentially decrease the recognition performance on necessary objects that are required by the robot. Using Fo CAL to learn unnecessary objects can also take valuable time that might be better used by the robot to perform necessary actions. For example, consider a dish washing robot which only needs to know kitchen utensils. The robot does not need to learn mechanical tools or children s toys but uncertainty-based sampling will cause the robot to learn these unnecessary objects. In such cases, some constraints might be added to a Fo CAL system in order to avoid the robot from learning unnecessary objects. Acknowledgements The authors acknowledge helpful comments of Alan R. Wagner for a preliminary version of this work. [1] Ali Ayub and Alan R. Wagner. Learning novel objects continually through curiosity. In IEEE ICRA 2021 (Workshop titled, Towards Curious Robots: Modern Approaches for Intrinsically-Motivated Intelligent Behavior)), 2021. [2] Ali Ayub and Alan R. Wagner. Online learning of objects through curiosity-driven active learning. In IEEE Ro Man (Workshop on Lifelong Learning for Long-term Human-Robot Interaction), 2020. [3] Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H. Lampert. i Ca RL: Incremental classifier and representation learning. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017. [4] James Kirkpatrick, Razvan Pascanu, Neil C. Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A. Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, Demis Hassabis, Claudia Clopath, Dharshan Kumaran, and Raia Hadsell. Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences of the United States of America, 114(13):3521 3526, 2017. [5] Ali Ayub and Alan R. Wagner. Tell me what this is: Few-shot incremental object learning by a robot. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020. [6] Ali Ayub and Alan Wagner. EEC: Learning to encode and regenerate images for continual learning. In International Conference on Learning Representations, 2021. [7] Saihui Hou, Xinyu Pan, Chen Change Loy, Zilei Wang, and Dahua Lin. Learning a unified classifier incrementally via rebalancing. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. [8] Yue Wu, Yinpeng Chen, Lijuan Wang, Yuancheng Ye, Zicheng Liu, Yandong Guo, and Yun Fu. Large scale incremental learning. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. [9] Xiaoyu Tao, Xiaopeng Hong, Xinyuan Chang, Songlin Dong, Xing Wei, and Yihong Gong. Few-shot class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020. [10] Ali Ayub and Alan R. Wagner. Cognitively-inspired model for incremental learning using a few examples. In The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2020. [11] M. Dehghan, Z. Zhang, M. Siam, J. Jin, L. Petrich, and M. Jagersand. Online object and task learning via human robot interaction. In 2019 International Conference on Robotics and Automation (ICRA), pages 2132 2138, May 2019. [12] Ali Ayub and Alan R. Wagner. What am i allowed to do here?: Online learning of context-specific norms by pepper. In Alan R. Wagner, David Feil-Seifer, Kerstin S. Haring, Silvia Rossi, Thomas Williams, Hongsheng He, and Shuzhi Sam Ge, editors, Social Robotics, pages 220 231, Cham, 2020. Springer International Publishing. [13] Robert M. French. Dynamically constraining connectionist networks to produce distributed, orthogonal representations to reduce catastrophic interference. Proceedings of the Sixteenth Annual Conference of the Cognitive Science Society, pages 335 340, 2019. [14] Yawar Siddiqui, Julien Valentin, and Matthias Niessner. Viewal: Active learning with viewpoint entropy for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020. [15] Donggeun Yoo and In So Kweon. Learning loss for active learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. [16] Devendra Singh Chaplot, Helen Jiang, Saurabh Gupta, and Abhinav Gupta. Semantic curiosity for active visual learning. In Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm, editors, Computer Vision ECCV 2020, pages 309 326, Cham, 2020. Springer International Publishing. [17] Martin Mundt, Yong Won Hong, Iuliia Pliushch, and Visvanathan Ramesh. A wholistic view of continual learning with deep neural networks: Forgotten lessons and the bridge to active and open world learning. Co RR, abs/2009.01797, 2020. [18] Yarin Gal, Riashat Islam, and Zoubin Ghahramani. Deep bayesian active learning with image data. In Proceedings of the 34th International Conference on Machine Learning - Volume 70, page 1183 1192. JMLR.org, 2017. [19] Rahaf Aljundi, Klaas Kelchtermans, and Tinne Tuytelaars. Task-free continual learning. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. [20] Anthony Robins. Catastrophic forgetting, rehearsal and pseudorehearsal. Connection Science, 7(2):123 146, 1995. [21] Claude Elwood Shannon. A mathematical theory of communication. University of Illinois Press, 1949. [22] Vincenzo Lomonaco and Davide Maltoni. Core50: a new dataset and benchmark for continuous object recognition. In Proceedings of the 1st Annual Conference on Robot Learning, volume 78, pages 17 26, 2017. [23] Sanjoy Dasgupta and Daniel Hsu. On-line estimation with the multivariate gaussian distribution. In Twentieth Annual Conference on Learning Theory, volume 4539, page 278 292, June 2007. [24] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pages 8024 8035. Curran Associates, Inc., 2019. [25] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016. [26] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, pages 1097 1105. 2012. [27] Z. Li and D. Hoiem. Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(12):2935 2947, Dec 2018. [28] Wei-Yu Chen, Yen-Cheng Liu, Zsolt Kira, Yu-Chiang Frank Wang, and Jia-Bin Huang. A closer look at few-shot classification. In International Conference on Learning Representations, 2019. [29] Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Dollar. Focal loss for dense object detection. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017. [30] Tsung-Yi Lin, Michael Maire, Serge J. Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C. Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, 2014. [31] Francisco M. Castro, Manuel J. Marin-Jimenez, Nicolas Guil, Cordelia Schmid, and Karteek Alahari. End-to-end incremental learning. In The European Conference on Computer Vision (ECCV), September 2018. [32] Xiaoyu Tao, Xinyuan Chang, Xiaopeng Hong, Xing Wei, and Yihong Gong. Topology-preserving classincremental learning. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XIX, page 254 270. Springer-Verlag, 2020. [33] Ali Ayub and Alan R. Wagner. F-siol-310: A robotic dataset and benchmark for few-shot incremental object learning. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 13496 13502, 2021. [34] Martin Mundt, Steven Lang, Quentin Delfosse, and Kristian Kersting. CLEVA-compass: A continual learning evaluation assessment compass to promote research transparency and comparability. In International Conference on Learning Representations, 2022. [35] William H. Beluch, Tim Genewein, Andreas Nürnberger, and Jan M. Köhler. The power of ensembles for active learning in image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018. [36] Tingke Shen, Amlan Kar, and Sanja Fidler. Learning to caption images through a lifetime by asking questions. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019. [37] Terrance E Boult, Steve Cruz, Akshay Raj Dhamija, Manuel Gunther, James Henrydoss, and Walter J Scheirer. Learning and the unknown: Surveying steps toward open world recognition. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 9801 9807, 2019. [38] Abhijit Bendale and Terrance Boult. Towards open world recognition. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2015. [39] Kimin Lee, Kibok Lee, Honglak Lee, and Jinwoo Shin. A simple unified framework for detecting outof-distribution samples and adversarial attacks. Advances in neural information processing systems, 31, 2018. [40] Walter J Scheirer, Lalit P Jain, and Terrance E Boult. Probability models for open set recognition. IEEE transactions on pattern analysis and machine intelligence, 36(11):2317 2324, 2014. [41] Ryota Yoshihashi, Wen Shao, Rei Kawakami, Shaodi You, Makoto Iida, and Takeshi Naemura. Classification-reconstruction learning for open-set recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4016 4025, 2019. [42] Ali Ayub, Chrystopher L. Nehaniv, and Kerstin Dautenhahn. Don t forget to buy milk: Contextually aware grocery reminder household robot. ar Xiv:2207.09050, 2022. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [Yes] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]