# humanintheloop_slam__ab928a14.pdf Human-in-the-Loop SLAM Samer B. Nashed, Joydeep Biswas College of Information and Computer Sciences 140 Governors Drive Amherst MA, USA 01003 Building large-scale, globally consistent maps is a challenging problem, made more difficult in environments with limited access, sparse features, or when using data collected by novice users. For such scenarios, where state-of-the-art mapping algorithms produce globally inconsistent maps, we introduce a systematic approach to incorporating sparse human corrections, which we term Human-in-the-Loop Simultaneous Localization and Mapping (Hit L-SLAM). Given an initial factor graph for pose graph SLAM, Hit L-SLAM accepts approximate, potentially erroneous, and rank-deficient human input, infers the intended correction via expectation maximization (EM), back-propagates the extracted corrections over the pose graph, and finally jointly optimizes the factor graph including the human inputs as human correction factor terms, to yield globally consistent large-scale maps. We thus contribute an EM formulation for inferring potentially rank-deficient human corrections to mapping, and human correction factor extensions to the factor graphs for pose graph SLAM that result in a principled approach to joint optimization of the pose graph while simultaneously accounting for multiple forms of human correction. We present empirical results showing the effectiveness of Hit L-SLAM at generating globally accurate and consistent maps even when given poor initial estimates of the map. Introduction Building large-scale globally consistent metric maps requires accurate relative location information between poses with large spatial separation. However, due to sensor noise and range limitations, such correlations across distant poses are difficult to extract from real robot sensors. Even when such observations are made, extracting such correlations autonomously is a computationally intensive problem. Furthermore, the order of exploration, and the speed of the robot during the exploration affect the numerical stability, and consequently the global consistency of large-scale maps. Due to these factors, even state of the art mapping algorithms often yield inaccurate or inconsistent large-scale maps, especially when processing data collected by novice users in challenging environments. To address these challenges and limitations of large-scale mapping, we propose Human-in-the-Loop SLAM (Hit L- Copyright c 2018, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: Hit L SLAM example, showing a) the input initial map with global consistency errors, and b) the resulting final map produced by Hit L-SLAM by incorporating human corrections (blue lines) along with the input. SLAM), a principled approach to incorporate approximate human corrections in the process of solving for metric maps1. Fig. 1 presents an example of Hit L-SLAM in practice. Hit L-SLAM operates on a pose graph estimate of a map along with the corresponding observations from each pose, either from an existing state-of-the-art SLAM solver, or aligned only by odometry. In an interactive, iterative process, Hit L-SLAM accepts human corrections, re-solves the pose graph problem, and presents the updated map estimate. This iterative procedure is repeated until the user is satisfied by the mapping result, and provides no further corrections. We thus present three primary contributions: 1) an EMbased algorithm (Dempster, Laird, and Rubin 1977) to interpret several types of approximate human correction for mapping ( 4), 2) a human factor formulation to incorporate a variety of types of human corrections in a factor graph for SLAM ( 5), and 3) a two-stage solver for the resultant, hybrid factor graph composed of human and robot factors, which minimally distorts trajectories in the presence of rank deficient human corrections ( 5). We show how Hit L-SLAM introduces numerical stability in the mapping problem by using human corrections to introduce off-diagonal blocks in the information matrix. Finally, we present several examples of Hit L-SLAM operating on maps that intrinsically included erroneous observations and poor initial map estimates, and 1Code and sample data is available at: https://github.com/umass-amrl/hitl-slam The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18) producing accurate, globally consistent maps. Related Work Solutions to robotic mapping and SLAM have improved dramatically in recent years, but state-of-the-art algorithms still fall short at being able to repeatable and robustly produce globally consistent maps, particularly when deployed over large areas and by non-expert users. This is in part due to the difficulty of the data association problem (Dissanayake et al. 2011; Bailey and Durrant-Whyte 2006; Aulinas et al. 2008). The idea of humans and robots collaborating in the process of map building to overcome such limitations is not new, and is known as Human-Augmented Mapping (HAM). Work within HAM belongs primarily to one of two groups, depending on whether the human and robot collaborate in-person during data collection (C-HAM), or whether the human provides input remotely or after the data collection (R-HAM). Many C-HAM techniques exist to address semantic (Nieto-Granda et al. 2010; Christensen and Topp 2010) and topological (Topp and Christensen 2006) mapping. A number of approaches have also been proposed for integrating semantic and topological information, along with human trackers (Milella et al. 2007), interaction models (Topp et al. 2006), and appearance information (Pronobis and Jensfelt 2012), into conceptual spatial maps (Zender et al. 2007), which are organized in a hierarchical manner. There are two limitations in these C-HAM approaches. First, a human must be present with the robot during data collection. This places physical constraints on the type of environments which can be mapped, as they must be accessible and traversable by a human. Second, these methods are inefficient with respect to the human s attention, since most of the time the human s presence is not critical to the robot s function, for instance during navigation between waypoints. These approaches, which focus mostly on semantic and topological mapping, also typically assume that the robot is able to construct a nearly perfect metric map entirely autonomously. While this is reasonable for small environments, globally consistent metric mapping of large, dynamic spaces is still a hard problem. In contrast, most of the effort in R-HAM has been concentrated on either incorporating human input remotely via tele-operation such as in the Urban Search and Rescue (USAR) problem (Murphy 2004; Nourbakhsh et al. 2005), or in high level decision making such as goal assignment or coordination of multiple agents (Olson et al. 2013; Parasuraman et al. 2007; Doroodgar et al. 2010). Some RHAM techniques for metric mapping and pose estimation have also been explored, but these involve either having the robot retrace its steps to fill in parts missed by the human (Kim et al. 2009) or by having additional agents and sensors in the environment (Kleiner, Dornhege, and Dali 2007). A number of other approaches have dealt with interpreting graphical or textual human input within the contexts of localization (Behzadian et al. 2015; Boniardi et al. 2016) and semantic mapping (Hemachandra, Walter, and Teller 2014). While these approaches solve similar signal interpretation problems, this paper specifically focuses on metric mapping. Ideally, a robot could explore an area only once with no need for human guidance or input during deployment, and later with minimal effort, a human could make any corrections necessary to achieve a near-perfect metric map. This is precisely what Hit L-SLAM does, and additionally Hit LSLAM does not require in-person interactions between the human and robot during the data collection. Human-in-the-Loop SLAM Hit L-SLAM operates on a factor graph G = {X, F}, where X is the set of estimated poses along the robot s trajectory, and F = {R, H} is the set of factors which encode information about both relative pose constraints arising from odometry and observations, R, and constraints supplied by the human, H. The initial factor graph G0 may be provided by any pose graph SLAM algorithm, and Hit L-SLAM is capable of handling constraints in G0 with or without loop closure. In our experiments, we used Episodic non-Markov Localization (En ML) (Biswas and Veloso 2017) without any explicit loop closures beyond the length of each episode. Hit L-SLAM runs iteratively, with the human specifying constraints on observations in the map, and the robot then enforcing those constraints along with all previous constraints to produce a revised estimate of the map. To account for inaccuracies in human-provided corrections, interpretation of the such input is necessary before human correction factors can be computed and added to the factor graph. Each iteration, the robot first proposes an initial graph Gi = {Xi, Fi}, then the human supplies a set of correction factors Hi, and finally the robot re-optimizes the poses in the factor graph, producing G i = {X i, Fi Hi}. Definition 1. A human correction factor, h, is defined by the tuple h = Pa, Pb, Sa, Sb, Xa, Xb, m with: Pa, Pb R2 : Sets of end-points of the two line segments a, b drawn by the human, Sa, Sb S : Sets of observations selected by the two line segments a, b respectively, Xa, Xb X1:t : Sets of poses from which the observations Sa, Sb were made, m M : The mode of correction. Sa, Sb are subsets of all observations S, and poses xi are added to the sets Xa, Xb if there are observations in Sa, Sb that arising from pose xi. M is an enumeration of the modes of human correction, a subset of which are shown in Fig. 2. The modes M of correction are defined as follows: 1. Colocation: A full rank constraint specifying that two sets of observations are at the same location, and with the same orientation. 2. Collinearity: A rank deficient constraint specifying that two sets of observations are on the same line, with an unspecified translation along the line. 3. Perpendicularity: A rank deficient constraint specifying that the two sets of observations are perpendicular, with an unspecified translation along either of their lines. 4. Parallelism: A rank deficient constraint specifying that the two sets of observations are parallel, with an unspecified translation along the parallel lines. Figure 2: Result of transforming observation point clouds based on different human constraints, showing (a) Original map, (b) Colocation constraint, (c) Collinear constraint, (d) Co-orientation constraint. In all sub-figures the red and blue lines denote Pa and Pb, respectively, and red and blue points denote Sa and Sb. S \ (Sa Sb) appear in orange. Each iteration of Hit L-SLAM proceeds in two steps, shown in Fig. 3. First, the human input is gathered, interpreted, and a set of human correction factors are instantiated (Block 1). Second, a combination of analytical and numerical techniques is used to jointly optimize the factor graph using both the human correction factors and the relative pose factors (Block 2). The resulting final map may be further revised and compressed by algorithms such as Long-Term Vector Mapping (Nashed and Biswas 2016). Figure 3: Flow of information during processing of the ith human input. Block 1 (yellow) outlines the evaluation of human input, and block 2 (purple) outlines the factor graph construction and optimization processes. Note that the joint optimization process optimizes both pose parameters and human constraint parameters. We model the problem of interpreting human input as finding the observation subsets Sa, Sb and human input sets Pa, Pb which maximize the joint correction input likelihood, p(Sa, Sb, Pa, Pb|P 0 a , P 0 b , m), which is the likelihood of selecting observation sets Sa, Sb and point sets Pa, Pb, given initial human input P 0 a , P 0 b and correction mode m. To find Sa, Sb and Pa, Pb we use the sets P 0 a , P 0 b and observations in a neighborhood around P 0 a , P 0 b as initial estimates in an Expectation Maximization approach. As the pose parameters are adjusted during optimization in later iterations of Hit L-SLAM, the locations of points in Pa, Pb may change, but once an observation is established as a member of Sa or Sb its status is not changed. Once Pa, Pb and Sa, Sb are determined for a new constraint, then given m we can find the set of poses X 1:t which best satisfy all given constraints. We first compute an initial estimate X0 1:t by analytic back-propagation of the most recent human correction factor, considering sequential constraints in the pose-graph. Next, we construct and solve a joint optimization problem over the relative pose factors r and the human correction factors h. This amounts to finding the set of poses X 1:t which minimize the sum of the cost of all factors, X 1:t = argmin X1:t i=1 cr(ri) + where cr : R R computes the cost from relative posegraph factor ri, and cm : H R computes the cost from human correction factor hj with correction mode m. Later sections cover the construction of the human correction factors and the formulation of the optimization problem. Interpreting Human Input Human Input Interface Hit L-SLAM users provide input by first entering the provide correction state by pressing the p key. Once in the provide correction state, they enter points for Pa, Pb by clicking and dragging along the feature (line segment) they wish to specify. The mode m is determined by which key is held down during the click and drag. For instance, CTRL signifies a colocation constraint, while SHIFT signifies a collinear constraint. To finalize their entry, the user exits the provide correction state by again pressing the p key. Exit from the provide correction state triggers the algorithm in full, and the user may specify additional corrections once a revised version of the map is presented. Human Input Interpretation Due to a number of factors including imprecise input devices, screen resolution, and human error, what the human actually enters and what they intend to enter may differ slightly. Given the raw human input line segment end-points P 0 a , P 0 b and the mode of correction m, we frame the interpretation of human input as the problem of identifying the observation sets Sa, Sb and the effective line segment endpoints Pa, Pb most likely to be captured by the potentially noisy points P 0 a , P 0 b . To do this we use the EM algorithm, which maximizes the log-likelihood ℓ, zi p(zi|si, θold) log(p(zi, si|θ)), where the parameters θ = {Pa, Pb} are the interpreted human input (initially assumed to be P 0 a , P 0 b ), the si S are the observations, and the latent variables zi are indicator variables denoting the inclusion or exclusion of si from Sa or Sb. The expressions for p(zi|si, θold) and p(zi, si|θ) come from a generative model of human error based on the normal distribution, N(μ(θ), σ2). Here, σ is the standard deviation of the human s accuracy when manually specifying points, and is determined empirically; μ(θ) is the center or surface of the feature. Let δ(si, θ) be the squared Euclidean distance between a given observation si and the feature (in this case a line segment) parameterized by θ. Note that p(zi|si, θ) is convex due to our Gaussian model of human error. Thus, the EM formulation reduces to iterative least-squares over changing subsets of S within the neighborhoods of Pa, Pb. The raw human inputs P 0 a , P 0 b are taken as the initial guess to the solution θ, and are successively refined of iterations of the EM algorithm to compute the final interpreted human input Pa, Pb. Once Pa, Pb have been determined, along with observations Sa, Sb, we can find the poses responsible for those observations Xa, Xb, thus fully defining the human correction factor h. To make this process more robust to human error when providing corrections, a given pose is only allowed in Xa or Xb if there exist a minimum of Tp elements in Sa or Sb corresponding to that pose. The threshold Tp is used for outlier rejection of provided human corrections. It is empirically determined by evaluating a human s ability to accurately select points corresponding to map features, and is the minimum number of points a feature must have for it to be capable of being repeatedly and accurately selected by a human. Solving Hit L-SLAM After interpreting human input, new pose estimates are computed in three steps. First, all explicit corrections indicated by the human are made by applying the appropriate transformation to Xb and subsequent poses. Next, any resultant discontinuities are addressed using Closed-Form Online Pose-Chain SLAM (COP-SLAM) (Dubbelman and Browning 2015). And last, final pose parameters are calculated via non-linear least-squares optimization of a factor graph. The three-step approach is necessary in order to avoid local minima. Applying Explicit Human Corrections Although the user may select sets of observations in any order, we define all poses xi Xa to occur before all poses xj Xb. That is, Pa is the input which selects observations Sa arising from poses Xa such that xi Xa and xj Xb, i < j, where Xb is defined analogously by observations Sb specified by input Pb. Given Pa and Pb, we find the affine correction transformation A which transforms the set of points defined by Pb to the correct location relative to the set of points defined by Pa, as specified by mode m. If the correction mode is rank deficient, we force the motion of the observations as a whole to be zero along the null space dimensions. For co-orientation, this means that the translation correction components of A are zero, and for collinearity the translation along the axis of collinearity is zero. Fig. 2 shows the effect of applying different types of constraints to a set of point clouds. After finding A we then consider the poses in Xb to constitute points on a rigid body, and transform that body by A. The poses xk such that xj Xb, k > j, are treated similarly, such that the relative transformations between all poses occurring during or after Xb remain unchanged. Error Backpropagation If Xa Xb does not form a contiguous sequence of poses, then this explicit change creates at least one discontinuity between the earliest pose in Xb, xb 0 and its predecessor, xc. We define affine transformation C such that xb 0 = Acb Cxc, where Acb was the original relative transformation between xc and xb 0. Given C, and the pose and covariance estimates for poses between Xa and Xb, we use COP-SLAM over these intermediate poses to transform xc without inducing further discontinuities. The idea behind COP-SLAM is a covariance-aware distribution of translation and rotation across many poses, such that the final pose in the pose-chain ends up at the correct location and orientation. The goal is to find a set of updates U to the relative transformations between poses in the posechain such that C = n i=1 Ui. COP-SLAM has two primary weaknesses as a solution to applying human corrections in Hit L-SLAM. First, it requires translation uncertainty estimates to be isotropic, which is not true in general. Second, COP-SLAM deals poorly with nested loops, where it initially produces good pose estimates but during later adjustments may produce inconsistencies between observations. This is because COP-SLAM is not able to simultaneously satisfy both current and previous constraints. Due to these issues, we use COP-SLAM as an initial estimate to a non-linear least-squares optimization problem, which produces a more robust, globally consistent map. Hit L-SLAM Optimization Without loop closure, a pose-chain of N poses has O(N) factors. With most loop closure schemes, each loop can be closed by adding one additional factor per loop. In Hit LSLAM, the data provided by the human is richer than most front-end systems, and reflecting this in the factor graph could potentially lead to a prohibitively large number of factors. If |Xa| = n and |Xb| = m, then a na ıve algorithm that adds a factor between all pairs (xa i , xa j ), (xa i , xb j), and (xb i, xb j), where xa Xa and xb Xb, would add (m + n)2 factors for every loop. This is a poor approach for two reasons. One, the large number of factors can slow down the optimizer and potentially prevent it from reaching the global optimum. And two, this formulation implies that every factor is independent of every other factor, which is incorrect. Thus, we propose a method for reasoning about human correction factors jointly, in a manner which creates a constant number of factors per loop while also preserving the structure and information of the input. Given a human cor- rection factor h = Pa, Pb, Sa, Sb, Xa, Xb, m , we define cm as the sum of three residuals, Ra, Rb, and Rp. The definitions of Ra and Rb are the same regardless of the correction mode m: |Sa| i=1 δ(sa i , Pa) |Sa| |Sb| i=1 δ(sb i, Pb) |Sb| As before, δ(s, P) denotes the squared Euclidean distance from observation s to the closest point on the feature defined by the set of points P. All features used in this study are line segments, but depending on m, more complicated features with different definitions for δ(s, P) may be used. Ra implicitly enforces the interdependence of different xa Xa, since moving a pose away from its desired relative location to other poses in Xa will incur cost due to misaligned observations. The effect on Xb by Rb is analogous. Figure 4: Subset of a factor graph containing a human factor h. Factors Ra and Rb drive observations in Sa and Sb toward features Pa and Pb, respectively. Factor Rp enforces the geometric relationship between Pa and Pb. Note that parameters in Xa (blue poses) and Xb (red poses) as well as Pa and Pb are jointly optimized. The relative constraints between poses in Xa and poses in Xb are enforced indirectly by the third residual, Rp. Depending on the mode, colocation (+), collinearity ( ), coorientation parallel ( ), co-orientation perpendicular ( ), the definition changes: R+ p = K1||cmb cma|| + K2(1 (ˆna ˆnb)), R p = K1||(cmb cma) ˆna|| + K2(1 (ˆna ˆnb)), R p = K2(1 (ˆna ˆnb)), R p = K2(ˆna ˆnb). Here, cma and cmb are the centers of mass of Pa and Pb, respectively, and ˆna and ˆnb are the unit normal vectors for the feature (line) defined by Pa and Pb, respectively. K1 and K2 are constants that determine the relative costs of translational error (K1) and rotational error (K2). The various forms of Rp all drive the points in Pb to the correct location and orientation relative to Pa. During optimization the solver is allowed to vary pose locations and orientations, and by doing so the associated observation locations, as well as points in Pa and Pb. Fig. 4 illustrates the topology of the human correction factors in our factor graph. Note that Hit L-SLAM allows human correction factors to be added to the factor graph in a larger set of situations compared to autonomous loop closure. Hit L-SLAM introduces information loop closure by adding correlations between distant poses without the poses being at the same location as in conventional loop closure. The off-diagonal elements in the information matrix thus introduced by Hit L-SLAM assist in enforcing global consistency just as the off-diagonal elements introduced by loop closure. Fig. 5 further illustrates this point note that the information matrix is still symmetric and sparse, but with the addition of off-diagonal elements from the human corrections. Figure 5: Example map (a) with corrections and resulting information matrix (b). The white band diagonal represents the correlations from the initial factor graph G0. The colored lines on the map show the human correction input: colocation (blue) and collinear (magenta).. The constraints correspond to the blue and magenta off-diagonal entries in the information matrix. Results Evaluation of Hit L-SLAM is carried out through two sets of experiments. The first set is designed to test the accuracy of Hit L-SLAM, and the second set is designed to test the scalability of Hit L-SLAM to large environments. To test the accuracy of Hit L-SLAM, we construct a data set in a large room during which no two parallel walls are simultaneously visible to the robot. We do this by limiting the range of our robot s laser to 1.5m so that it sees a wall only Figure 6: Initial and final maps from Hit L-SLAM. Each map is of the same floor, and consists of between 600 and 700 poses. Maps in the left column (a) are initial maps, and maps in the right column (b) are final maps. Observations are shown in orange and poses are shown as arrows. Poses which are part of a human constraint are blue, while those which are not are in black. Figure 7: A large map a) corrected by Hit L-SLAM b) using human correction, some of which are highlighted c). d) shows an approximate overlay of the map onto an aerial image of the complex from google earth. The map contains over 3000 poses. when very close. We then drive it around the room for which we have ground truth dimensions. This creates sequences of lost poses throughout the pose-chain which rely purely on odometry to localize, thus accruing error over time. We then impose human constraints on the resultant map and compare to ground truth, shown in Fig. 8. Note that the human corrections do not directly enforce any of the measured dimensions. The initial map shows a room width of 5.97m, and an angle between opposite walls of 4.1 . Hit L-SLAM finds a room width of 6.31m, while the ground truth width is 6.33m, and produces opposite walls which are within 1 of parallel. Note also that due to the limited sensor range, the global correctness must come from proper application of human constraints to the factor graph including the lost poses between wall observations. To quantitatively evaluate accuracy on larger maps, where exact ground truth is sparse, we measured an inter-corridor spacing (Fig. 6 2b), which is constant along the length of the building. We also measured angles between walls we know to be parallel or perpendicular. The results for ground truth comparisons before and after Hit L-SLAM, displayed in Table 1, show that Hit L-SLAM is able to drastically reduce map errors even when given poor quality initial maps. We introduce an additional metric for quantitative map evaluation. We define the pair-wise inconsistency Ii,j between poses xi and xj to be the area which observations from pose xi show as free space and observations from pose xj show as occupied space. We define the total inconsistency I over the map as the pair-wise sum of inconsistencies between all pairs of poses, I = N 1 i=1 N j=i+1 Ii,j. The inconsistency metric thus serves as a quantitative metric of global registration error between observations in the map, and allows us to track the effectiveness of global registration using Hit L-SLAM over multiple iterations. The initial inconsistency values for maps LGRC 3A and 3B were 297.5m2 and 184.3m2, respectively. The final inconsistency values were 47.6m2 and 3.7m2, respectively, for an average Figure 8: Initial a) and final b) maps for the lost poses experiment. Observations are shown in orange, poses are black arrows, and ground truth (walls) is represented by the black lines. Poses involved in human constraints are colored blue. Map Samples Input Err. Hit L-SLAM Err. A T A( ) T(m) A( ) T(m) Lost Poses 10 4 3.1 0.07 1.0 0.02 LGRC 3A 14 10 9.8 3.3 1.5 0.06 LGRC 3B 14 10 7.6 3.1 1.1 0.02 BIG MAP 22 10 5.9 2.8 1.6 0.03 Mean 60 34 6.74 2.71 1.4 0.04 Table 1: Quantitative mapping errors using Hit L-SLAM compared to ground truth, in the input maps, and after Hit LSLAM. The Samples column denotes how many pairwise feature comparisons were made on the map and then compared to hand-measured ground truth. Angular (A) errors are in degrees, translation (T) errors in meters. inconsistency reduction of 91% relative to the initial map, thus demonstrating the improved global consistency of the map generated using Hit L-SLAM. Fig. 6 and Fig. 9 offer some qualitative examples of Hit L-SLAM s performance. To test the scalability of Hit L-SLAM, we gathered several datasets with between 600 and 700 poses, and one with over 3000 poses and nearly 1km of indoor travel between three large buildings. Fig. 6 shows some of the moderately sized maps, and Fig. 7 details the largest map. 16 constraints were required to fix the largest map, and computation time never exceeded the time required to re-display the map, or for the human to move to a new map location. All maps shown in Fig. 6 were corrected interactively by the human using Hit L-SLAM in under 15 minutes. Furthermore, Hit L-SLAM solves two common problems which are difficult or impossible to solve via re-deployment: 1) a severely bent hallway, in Fig. 6 1a), and 2) a sensor failure, in Fig. 6 2a) which caused the robot to incorrectly estimate its heading by roughly 30 degrees at one point. Combined, these results show that incorporating human input into metric mapping can be done in a principled, computationally tractable manner, which allows us to fix metric mapping consistency errors in less time and with higher accuracy than previously possible, given a small amount of human input. Figure 9: Example of how rank-deficient (collinear) constraints can be combined to effect a full rank constraint. Often, it is easier to tell if two sets of observations are collinear rather than colocated. Supporting rank-deficient constraints allows Hit L-SLAM to operate on a larger set of maps. Conclusion We present Human-in-the-Loop SLAM (Hit L-SLAM), an algorithm designed to leverage human ability and metaknowledge as they relate to the data association problem for robotic mapping. Hit L-SLAM contributes a generalized framework for interpreting human input using the EM algorithm, as well as a factor graph based algorithm for incorporating human input into pose-graph SLAM. Future work in this area could proceed towards further reducing the human requirements, and extending this method for higher dimensional SLAM and for different sensor types. Acknowledgments The authors would like to thank Manuela Veloso from Carnegie Mellon University for providing the Co Bot4 robot used to collect the AMRL datasets, and to perform experiments at University of Massachusetts Amherst. This work is supported in part by AFRL and DARPA under agreement #FA8750-16-2-0042. References Aulinas, J.; Petillot, Y. R.; Salvi, J.; and Llad o, X. 2008. The slam problem: a survey. In CCIA, 363 371. Citeseer. Bailey, T., and Durrant-Whyte, H. 2006. Simultaneous localization and mapping (slam): Part ii. IEEE Robotics & Automation Magazine 13(3):108 117. Behzadian, B.; Agarwal, P.; Burgard, W.; and Tipaldi, G. D. 2015. Monte carlo localization in hand-drawn maps. In Intelligent Robots and Systems (IROS), 2015 IEEE/RSJ International Conference on, 4291 4296. IEEE. Biswas, J., and Veloso, M. M. 2017. Episodic non-markov localization. Robotics and Autonomous Systems 87:162 176. Boniardi, F.; Valada, A.; Burgard, W.; and Tipaldi, G. D. 2016. Autonomous indoor robot navigation using a sketch interface for drawing maps and routes. In Robotics and Automation (ICRA), 2016 IEEE International Conference on, 2896 2901. IEEE. Christensen, H. I., and Topp, E. A. 2010. Detecting region transitions for human-augmented mapping. Dempster, A. P.; Laird, N. M.; and Rubin, D. B. 1977. Maximum likelihood from incomplete data via the em algorithm. Journal of the Royal Statistical Society. Series B (Methodological) 39(1):1 38. Dissanayake, G.; Huang, S.; Wang, Z.; and Ranasinghe, R. 2011. A review of recent developments in simultaneous localization and mapping. In 2011 6th International Conference on Industrial and Information Systems, 477 482. IEEE. Doroodgar, B.; Ficocelli, M.; Mobedi, B.; and Nejat, G. 2010. The search for survivors: Cooperative human-robot interaction in search and rescue environments using semiautonomous robots. In Robotics and Automation (ICRA), 2010 IEEE International Conference on, 2858 2863. IEEE. Dubbelman, G., and Browning, B. 2015. Cop-slam: Closedform online pose-chain optimization for visual slam. IEEE Transactions on Robotics 31(5):1194 1213. Hemachandra, S.; Walter, M. R.; and Teller, S. 2014. Information theoretic question asking to improve spatial semantic representations. In 2014 AAAI Fall Symposium Series. Kim, S.; Cheong, H.; Park, J.-H.; and Park, S.-K. 2009. Human augmented mapping for indoor environments using a stereo camera. In 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, 5609 5614. IEEE. Kleiner, A.; Dornhege, C.; and Dali, S. 2007. Mapping disaster areas jointly: Rfid-coordinated slam by hurnans and robots. In 2007 IEEE International Workshop on Safety, Security and Rescue Robotics, 1 6. IEEE. Milella, A.; Dimiccoli, C.; Cicirelli, G.; and Distante, A. 2007. Laser-based people-following for human-augmented mapping of indoor environments. In Artificial Intelligence and Applications, 169 175. Murphy, R. R. 2004. Human-robot interaction in rescue robotics. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 34(2):138 153. Nashed, S., and Biswas, J. 2016. Curating long-term vector maps. In Intelligent Robots and Systems (IROS), 2016 IEEE/RSJ International Conference on, 4643 4648. IEEE. Nieto-Granda, C.; Rogers, J. G.; Trevor, A. J.; and Christensen, H. I. 2010. Semantic map partitioning in indoor environments using regional analysis. In Intelligent Robots and Systems (IROS), 2010 IEEE/RSJ International Conference on, 1451 1456. IEEE. Nourbakhsh, I. R.; Sycara, K.; Koes, M.; Yong, M.; Lewis, M.; and Burion, S. 2005. Human-robot teaming for search and rescue. IEEE Pervasive Computing 4(1):72 79. Olson, E.; Strom, J.; Goeddel, R.; Morton, R.; Ranganathan, P.; and Richardson, A. 2013. Exploration and mapping with autonomous robot teams. Communications of the ACM 56(3):62 70. Parasuraman, R.; Barnes, M.; Cosenzo, K.; and Mulgund, S. 2007. Adaptive automation for human-robot teaming in future command and control systems. Technical report, DTIC Document. Pronobis, A., and Jensfelt, P. 2012. Large-scale semantic mapping and reasoning with heterogeneous modalities. In Robotics and Automation (ICRA), 2012 IEEE International Conference on, 3515 3522. IEEE. Topp, E. A., and Christensen, H. I. 2006. Topological modelling for human augmented mapping. In 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2257 2263. IEEE. Topp, E. A.; Huettenrauch, H.; Christensen, H. I.; and Eklundh, K. S. 2006. Bringing together human and robotic environment representations-a pilot study. In 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, 4946 4952. IEEE. Zender, H.; Jensfelt, P.; Mozos, O. M.; Kruijff, G.-J. M.; and Burgard, W. 2007. An integrated robotic system for spatial understanding and situated interaction in indoor environments. In AAAI, volume 7, 1584 1589.