# lingospace_languageconditioned_incremental_grounding_for_space__387245c1.pdf LINGO-Space: Language-Conditioned Incremental Grounding for Space Dohyun Kim, Nayoung Oh, Deokmin Hwang, Daehyung Park* Korea Advanced Institute of Science and Technology, Republic of Korea {dohyun141, lightsalt, gsh04089, daehyung}@kaist.ac.kr We aim to solve the problem of spatially localizing composite instructions referring to space: space grounding. Compared to current instance grounding, space grounding is challenging due to the ill-posedness of identifying locations referred to by discrete expressions and the compositional ambiguity of referring expressions. Therefore, we propose a novel probabilistic space-grounding methodology (LINGO-Space) that accurately identifies a probabilistic distribution of space being referred to and incrementally updates it, given subsequent referring expressions leveraging configurable polar distributions. Our evaluations show that the estimation using polar distributions enables a robot to ground locations successfully through 20 table-top manipulation benchmark tests. We also show that updating the distribution helps the grounding method accurately narrow the referring space. We finally demonstrate the robustness of the space grounding with simulated manipulation and real quadruped robot navigation tasks. Code and videos are available at https://lingo-space.github.io. Introduction Robotic natural-language grounding methods have primarily focused on identifying objects, actions, or events (Brohan et al. 2023; Mees, Borja-Diaz, and Burgard 2023). However, to robustly carry out tasks following human instructions in physical space, robots need to identify the space of operations and interpret spatial relations within the instructions. For example, given a directional expression (e.g., place a cup on the table and close to the plate ), a robot should determine the most suitable location for placement based on the description and environmental observations. We aim to solve the problem of localizing spatial references within instructions. Referred to as space grounding, this problem involves identifying potential locations for reaching or placing objects. Unlike conventional instance grounding problems, such as visual object grounding (Shridhar, Mittal, and Hsu 2020) and scene-graph grounding (Kim et al. 2023), space grounding presents complexities due to inherent ambiguity in identifying referred locations (as illustrated in Fig. 1). Given the ambiguity and the compositional nature of referring expressions, a grounding solution *D. Park is the corresponding author. Copyright 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: An illustration of incremental space grounding in the navigation task. Our method, LINGO-Space, identifies the distribution of the target location indicated by a natural language instruction with referring expressions. should be capable of reasoning about potential space candidates with uncertainty and adapting to new references. Conventional space grounding approaches often map spatially relational expressions (e.g., to the right side of a box ) to specific directional and distance-based coordinates, learning patterns from training dataset (Jain et al. 2023; Namasivayam et al. 2023). However, positional ambiguity (e.g., missing distance information) in spatial expressions and referential ambiguity (e.g., a plurality of similar objects) in the scene often lower effective space grounding (Zhao, Lee, and Hsu 2023). Furthermore, representational ambiguity restricts the scalability of space grounding when dealing with complex expressions and scenes. On the other hand, understanding composite expressions is crucial for accurate grounding in space. Most instructions entail sequences of spatiotemporal descriptions (e.g., Enter the room, then place the cup on the table ). However, most approaches often encode multiple expressions simultaneously without explicit separation (Roy et al. 2019). Recently, PARAGON (Zhao, Lee, and Hsu 2023) decomposes The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) a composite instruction into object-centric relation expressions and jointly encodes them using graph neural networks. Given the variability in results due to the expression order and incoming expressions, we need an incremental grounding method with composite expressions. Therefore, we propose LINGO-Space, a languageconditioned incremental grounding method for space.1 This method identifies a probabilistic distribution of the referenced space by leveraging configurable polar distributions. Our method incrementally updates the distribution given subsequent referring expressions, resolving compositional ambiguity via a large language model (LLM)-guided semantic parser. We also mitigate referential ambiguity by leveraging scene-graph-based representations in grounding. Our evaluation shows that estimating polar distributions effectively grounds space as described by referring expressions, while conventional methods have difficulty capturing uncertainty. We also show the capability to refine the distribution accurately and narrow down the referenced space as humans encounter space navigation in complex domains. Our contributions are as follows: We propose a novel space representation using a mixture of configurable polar distributions, offering a probability distribution of referred locations. We introduce an incremental grounding network integrated with an LLM-based semantic parser, enabling robust and precise grounding of incoming expressions. We conduct 20 benchmark evaluations, comparing with state-of-the-art baselines, and demonstrate the real-world applicability of our method through space-reaching experiments involving a quadruped robot, Spot. Related Work Language grounding: The problem of language instruction delivery has received increasing attention in robotics. Early works have focused on understanding entities or supplementary visual concepts (Matuszek et al. 2012). Recent works have incorporated spatial relations to enhance the identification of instances referred to in expressions (Howard, Tellex, and Roy 2014; Paul et al. 2018; Hatori et al. 2018; Shridhar, Mittal, and Hsu 2020). Space grounding: There are efforts mapping spatial relations to the region of actions using various representations: potential fields (Stopp et al. 1994), discrete regions with fuzzy memberships (Tan, Ju, and Liu 2014), and points from a multi-class logistic classifier (Guadarrama et al. 2013). Early approaches have employed predetermined distances or directions, or randomly sampled locations to represent spatial relations. Neural representations have emerged predicting pixel positions (Venkatesh et al. 2021) or pixel-wise probabilities (Mees et al. 2020) for placement tasks. To overcome limitations associated with pixel-based distributions, researchers have used parametric probability distributions, such as a polar distribution (Kartmann et al. 2020), a mixture of Gaussian distributions (Zhao, Lee, and Hsu 2023), 1LINGO-Space is an abbreviation of Language-conditioned INcremental Gr Ounding method for Space and a Boltzmann distribution (Gkanatsios et al. 2023). Our proposed method adopts the polar distribution as a basis for modeling spatial concepts, avoiding the need to predefine the number of components as required by Gaussian mixture models (Kartmann et al. 2020; Paxton et al. 2022). Further, our method considers the order of expressions and the semantic and geometric relations among objects, allowing for handling semantically identical objects. Composite instructions: Composite linguistic instructions often introduce structural ambiguity. Researchers often use parsing as a solution, breaking down composite expressions using hand-crafted or grammatical rules (Tellex et al. 2011; Howard et al. 2022). Recently, Zhao, Lee, and Hsu (2023) have introduced the grounding method, PARAGON, in which its neural parsing module extracts object-centric relations. Similarly, Gkanatsios et al. (2023) decompose expressions into spatial predicates using a neural semantic parser, i.e., a sequence-to-tree model (Dong and Lapata 2016). While these works often deal with one or two referring expressions, our method incrementally manages an arbitrary number of referring expressions by introducing an LLM-based parser. Large language models: LLMs have brought increasing attention offering benefits in the areas of understanding highlevel commands (Brohan et al. 2023), extracting common manipulation-related knowledge (Ren et al. 2023), planning with natural language commands (Huang et al. 2023; Song et al. 2023; Driess et al. 2023; Mees, Borja-Diaz, and Burgard 2023), and programming (Singh et al. 2023; Liang et al. 2023). While these approaches generally focus on the LLM s capability to leverage semantic knowledge for understanding natural language instructions, we focus on another capability: decomposing linguistically complex commands into sub-commands or problems. Shah et al. (2022) employ an LLM to generate a list of landmarks within composite commands. Similarly, Liu et al. (2022) use an LLM to identify referring expressions and translate natural language commands into linear temporal logics. Our method also uses an LLM to parse composite referring instructions and transform them into a structured format, enhancing the grounding process. Problem Formulation Consider the problem of determining a desired location x R2 based on a natural language instruction Λ, while taking into account a set of objects O = {o1, ...o N} in the environment, where N denotes the number of objects. To enhance the accuracy of indicating the location x , the instruction Λ might include object references that leverage their geometric relationships with the intended location. To robustly represent potential locations, referred to as space, we model the target location as a probability distribution parameterized by θ. Therefore, we formulate an optimization problem in which we marginalize out θ and a scene graph Υsg that encodes O with their relationships (i.e., edge features): x = argmax x P(x|Λ, O), (1) θ,Υsg P(x|θ)P(θ|Λ, Υsg)P(Υsg|O), (2) The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 2: The overall architecture of LINGO-Space on a tabletop manipulation task. Given a composite instruction, a graph generator provides a scene graph. A semantic parser decomposes the instruction into a structured form of relation-embedding tuples r(i), where i {1, ..., M}. Finally, a spatial-distribution estimator incrementally updates a probabilistic distribution of locations satisfying spatial constraints encoded in the embedding tuples. where we assume conditional independence between the current distribution model θ, given the scene graph Υsg, and the object set O. This work assumes a scene-graph generator produces an optimal graph Υ sg. Another problem is the use of composite instructions with multiple relations (i.e., referring expressions) in sequence. We assume a plurality of similar reference objects to be present in the environment, potentially leading to instances where semantically identical labels appear on the graph Υ sg. To mitigate compositional ambiguity in composite instructions, we decompose Λ into multiple phrases, Λ = [Λ(1), ..., Λ(M)], where M is the number of constituent phrases. Each phrase Λ(i) includes a single spatial relation about a referenced object (e.g., left of the block ). We then reformulate the objective function in Eq. (2) using an iterative update form with Υ sg: P(x|θM) | {z } Location selector i=1 [P(θi|θi 1, Λ(i), Υ sg) | {z } Spatial-distribution estimator P(Λ(i)|Λ, Υ sg) | {z } Semantic parser where P(θ1|θ0, Λ(1), Υsg) = P(θ1|Λ(1), Υsg). We describe each process and the graph generator below. Methodology: LINGO-Space We present LINGO-Space, an incremental probabilistic grounding approach that predicts the spatial distribution of the target space referenced in a composite instruction. The architecture of our method, as depicted in Fig. 2, consists of 1) a scene-graph generator, 2) a semantic parser, and 3) a spatial-distribution estimator. Below, we describe each module and the incremental process of estimating the spatial distribution to ground the desired location effectively. A Scene-Graph Generator A scene graph Υsg = (V, E) is a graphical representation of a scene consisting of detected objects as nodes V and their pairwise relationships as directed edges E. Each node u V entails node features f u including its Cartesian coordinate f u coord R2, bounding box f u box R4, and visual feature f u viz RDviz; f u = (f u coord, f u box, f u viz), where Dviz is a fixed size. In detail, a bounding box detector (e.g., Grounding DINO (Liu et al. 2023) for manipulation) finds f u coord and f u box. Then, we encode its cropped object image as f u viz using the CLIP image encoder (Radford et al. 2021). Each edge euv E represents a spatial relationship, as a textual predicate, from u to v V. We determine predicates (i.e., near , in ) using the box coordinate and size, (fcoord, fbox). We encode each predicate as an edge feature f uv edge RDtxt using the CLIP text encoder (Radford et al. 2021), where Dtxt is a fixed size. We design a scene-graph generator that returns a graph Υsg in a dictionary form; V is a dictionary with node identification numbers (ID) as keys and node features as values, e.g., {23: [f 23 coord, f 23 box, f 23 viz] }, where 23 is an ID number. Note that we use node IDs as an interchangeable concept with nodes. We also represent an edge set E as a list of relationship triplets, (u ID, f u IDv ID edge , v ID), where u ID and v ID are the subject and object node IDs. This work assumes that each edge contains only one relation in a pre-defined set. A Semantic Parser We introduce an LLM-based semantic parser, using Chat GPT (Open AI 2023), which 1) breaks down a composite instruction Λ with M referring expressions into its constituent phrases [Λ(main), Λ(1), ..., Λ(M)] and 2) transforms these phrases into a structured format, leveraging the LLM s proficiency in in-context learning through prompts. We design the prompt to consist of a task description and parsing demonstrations. The task description explains our parsing task as well as its reasoning steps: 1) identification of an action from the instruction, 2) identification of a source instance associated with the action, and 3) identification of target information from the referring expressions, character- The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 3: Architecture of the spatial-distribution estimator. Given the graph representation of the problem description, the network predicts instance-wise polar distributions, updating the internal model with the previous state. ized by a relational predicate and a referenced object. Then, the demonstrations provide three input-output examples to regularize the output format. In detail, our parser represents the main phrase Λ(main) as an action with an associated source object and the other referring phrases, [Λ(1), ..., Λ(M)], as relational predicates with referenced objects. To incrementally ground the phrases, our parser converts the referenced object-predicate pairs as a list of relation tuples, [r(1), ..., r(M)]. For instance, Input: put the cyan bowl above the chocolate and left of the silver spoon. Output: {action: move , source: cyan bowl , target: [( chocolate , above ), ( silver spoon , left )] }. We then post-process the output form to have better representations for grounding. For it, we replace the action into a robot skill with the highest word similarity in a skill set. We also replace the text-based r(i) into an embedding-based tuple r(i) = (f (i) ref , f (i) pred), where f (i) ref RDtxt and f (i) pred RDtxt are encoded from the CLIP text encoder (Radford et al. 2021). A Spatial-Distribution Estimator Our spatial distribution estimator predicts a probability distribution of destinations given a referring phrase and a scene graph. Given a sequence of phrases, the estimator incrementally updates the distribution using a graph-based incremental grounding network. Spatial distribution We represent the probability distribution as a mixture of instance-wise polar distributions, where a polar distribution is a joint probability density function of two random variables, distance d R 0 and angle ϕ [ π, π], in the polar coordinate system. As Kartmann et al. (2020), we assume that d and ϕ follow a Gaussian distribution N and a Von Mises (i.e., circular) distribution M, respectively; (d, ϕ) (N(µd, σ2 d), M(µϕ, κϕ)), (4) where µd, σ2 d, µϕ, and κϕ indicate mean, variance, location, and concentration, respectively. Then, the mixture of instance-wise polar distribution given a tuple r(i) is, P(x|r(i), Υsg) = j=1 wj P(d; µdj, σ2 dj) P(ϕ; µϕj, κϕj). where wj is a weight that represents the relevance between the j-th instance and the relation tuple r(i). We represent the entire distribution as the mixture model parameters θ: θ = [(w1, µd1, σ2 d1, µϕ1, κϕ1), ..., (w N, µd N , σ2 d N , µϕN , κϕN )]. Pre-processing Given a node feature f j and a relation tuple r(i), we pre-process them to have better representations by projecting into another space via positional encoding and matrix projection processes. The positional encoding embeds the Cartesian coordinate f j coord using sinusoidal positional encoding γ( ): f j coord = γ(f j coord) R2(2K+1) where K is a predefined maximum frequency K N (Mildenhall et al. 2021). The matrix projection process projects each feature (i.e., f j viz, f (i) ref , and f (i) pred) into a new feature space with dimension DH by multiplying a learnable projection matrix: f j viz = Mvizf j viz, f (i) ref = Mreff (i) ref , f (i) pred = Mpredf (i) pred, where Mviz RDH Dviz, Mref RDH DCLIP, and Mpred RDH DCLIP. In addition, to use the last estimation model state f (i 1) state RN DH , we also compress f (i 1) state into f (i 1) state RDH by applying a max-pooling operation and a linear projection, where DH = 4DH + 2(2K + 1). Therefore, we obtain a new feature vector X(i,j) 0 RDH that is a concatenation of the projected feature vectors, X(i,j) 0 = concat( f j coord, f j viz, f (i) ref , f (i) pred, f (i 1) state ). (6) For i = 1, we use a zero vector as the last model state. Estimation network We design the estimation network to predict instant-wise polar distributions given a relation tuple r(i), taking new node features X0 = (X(i,1) 0 , ..., X(i,N) 0 ) and edge features E0 = (..., f uv edge, ...). Fig. 3 shows the network architecture, which is a stack of GPS layers (Ramp aˇsek et al. 2022), where each GPS layer is a hybrid layer of a messagepassing neural network (MPNN) and a global attention network. This work uses GINE (Hu et al. 2020) as the MPNN layer and Transformer (Vaswani et al. 2017) as the global attention layer. The GPS layer allows the network to update node and edge features, Xl+1, El+1 = GPSl(Xl, El, A), (7) where A RN N is the adjacency matrix of the scene graph Υsg, l {1, ..., L}, and L is the number of layers. In the estimation network, the l-th GPS layer returns instant-wise hidden states X(i,j) l DH to predict the The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Instruction Input Image to the lower right corner of a grey cup and to the lower left of a spoon far from the cyan cylinder and to the left above of the cyan bowl Prediction Ground-truth Reference object LINGO-Space CLIPort Para Gon Figure 4: Qualitative evaluation with LINGO-Space, CLIPORT, and PARAGON. Grey boxes represent the object each i-th phrase refers to, while red dots and blue dots represent the ground-truth and the prediction, respectively. We plot 100 particles for the PARAGON s prediction result. The results demonstrate that LINGO-Space is capable of accurately identifying the space referred to by a composite instruction by narrowing down the space. distribution parameters θ[j] for the j-th node as well as the current prediction state f (l) state RN DH . In detail, on the last L-th layer, the network outputs instant-wise hidden state X(M,j) L RDH and predicts instant-wise polar distribution parameters θ[j] applying a two-layer multilayer perceptron (MLP) per parameter. When the distribution parameters are non-negatives (e.g., wj, σ2 dj, and κϕj), we use softplus activation functions. In addition, for the concentration κϕj, we enable the MLP to produce 1 κϕ instead of κϕ since the inverse of concentration is analogous to the variance. For the parameter µϕj, to avoid the discontinuity on angles (e.g., 0 and 360 ), we map XM,j L to [xϕj, yϕj] R2 via a two-layer MLP and then apply an atan2 activation function µϕj = atan2(yϕj, xϕj) R. Through iterations, the estimation network returns spatial distributions conditioned on all the possible subsequences of the given relations r(1), ..., r(M) in sequence. We then generate a final spatial distribution using Eq. (3) and Eq. (5): p(θM|r(1), ..., r(M), Υ sg) = QM i=1 pi(θi|θi 1, r(i), Υ sg). Objective function To train the estimation network, we introduce a composite loss function L that is a linear combination of two loss functions, L1 and L2. L1 is the negative log-likelihood of the spatial distribution given the groundtruth locations xdes and ground-truth weight wdes j : j=1 wdes j P(xdes; µdj, σ2 dj, µϕj, κϕj) L2 is the cross-entropy loss between the predicted weight wj and the ground-truth weight wdes j : j=1 wdes j log(wj). (9) Then, the combined loss is L = λL1+(1 λ)L2, where λ is a hyperparameter. Given a composite instruction, we com- pute the combined loss and incrementally update the network given each relation r(i) in sequence. Experimental Setup Our experiments aim to answer the following questions: Does the proposed method improve the performance of space grounding given 1) an instruction with a single referring expression and 2) a composite instruction with multiple referring expressions? Further, can the proposed method apply to real-world tasks? Grounding with a Referring Expression We evaluate the grounding capability of inferring a location for successfully placing objects within a tabletop domain, guided by instructions containing a referring expression. We use three baseline methods with their benchmarks within the Py Bullet simulator (Coumans and Bai 2016). Each benchmark provides a top-down view of RGB-D images with synthesized structured instructions. Below are the training and test procedures per benchmark. CLIPORT s benchmark (Shridhar, Manuelli, and Fox 2022): We use three tasks designed to pack an object inside a referenced object. Task scenes contain between four to ten objects from the Google Scanned Objects (Downs et al. 2022) or primitive shapes, denoted as google and shape, respectively. Task instructions include objects seen during training or not, denoted as seen and unseen, respectively. We use instructions with referential expressions such as pack the bull figure in the brown box. The assessment metric is a success score ( [0, 1]) reflecting the extent of relationship satisfaction between the located object volume and the desired container. PARAGON s benchmark (Zhao, Lee, and Hsu 2023): This benchmark generates a dataset for the task of placing an object in the presence of semantically identical objects following one of nine directional relations: center, left, right, above, below, left above, left The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) below, right above, and right below. The evaluation metric is a binary success score ( {0, 1}), denoting whether all predicates have been satisfied after placements. SREM s benchmark (Gkanatsios et al. 2023): We use eight tasks designed to rearrange a colored object inside a referenced object following spatial instructions featuring one of four directional relations: left, right, behind, and front. Each scene contains between four to seven objects, as in CLIPORT benchmark. The evaluation metric uses a success score as CLIPORT with the most conservative threshold. LINGO-Space s benchmark: We introduce close-seencolors, close-unseen-colors, far-seen-colors, far-unseencolors tasks with new predicates, close and far. Other setups are similar to the SREM benchmark. For PARAGON benchmark, for each task, training and testing employ 400 and 200 scenes, respectively. Otherwise, we train models on 100 samples, with subsequent testing performed on 200 randomized samples. Grounding with Multiple Referring Expressions We also assess the performance of incremental grounding in the presence of multiple relations in sequence. However, benchmarks such as CLIPORT and PARAGON focus on instructions with simple relations or structures. Instead, we introduce a new task labeled as composite to illustrate better the challenge of grounding subsequent relations in environments with multiple semantically identical objects. This task assumes tabletop grounding scenarios akin to CLIPORT. Each sample consists of 640 320 RGB-D images and synthesized referring instructions with 10-direction relations: left, right, above, below, left above, right above, left below, right below, close, and far . For example, we use put the green ring to the left of the gray cube, the above of the gray cube, and the right of the red bowl. In the dataset, we randomly place two-to-seven objects, considering one-to-three independent relation phrases for training and one-to-six phrases for testing. Our benchmarks use SREM s scoring criteria below. In this task, only one region strictly satisfies all the relations. To verify it, we use manually programmed boundingbased relation checkers. Across all tasks, we train models on 200 samples without having semantically identical objects and repeated directional predicates. We then test on 300 samples. In addition, we perform evaluations with SREM s tasks designed for multiple referential instructions (i.e., compone-step-(un)seen-colors ). After grounding a target location, we compute a score reflecting the extent of relationship satisfaction; score = #satisfaction/|relations|. Baseline Methods CLIPORT is a language-conditioned imitation learning model that predicts pixel locations using CLIP (Radford et al. 2021) and Transporter Networks (Zeng et al. 2021). Note that we disabled its rotational augmentation except for the CLIPORT benchmark. PARAGON is a parsing-based visual grounding model for object placement. This model generates particles as location candidates (i.e., pixels). We use the pixel point with maximum probability as a placement position. For visual feature extraction, we use the provided Mask RCNN (He et al. 2017) for PARAGON benchmark and Grounding Dino (Liu et al. 2023) for the other benchmarks. SREM is an instruction-guided rearrangement framework. Parsing an instruction into multiple spatial predicates, SREM s open-vocabulary visual detector (Jain et al. 2022) grounds them to objects in input images. For LINGO-Space benchmark, we train energy-based models for new predicates. We then train the grounder with each task dataset while training the parser with each benchmark dataset following the paper. We exclude the Transporter Networks (Zeng et al. 2021) since we provide the ground-truth location given a bounding box. We also disable the closed-loop execution for fairness in comparison. Grounding with a referring expression. We analyze the grounding performance of our proposed method and baselines via 12 benchmarks. Table 1 shows our method outperforms three baseline methods, demonstrating superior performance in 11 out of 12 tasks. Our method consistently exhibits the highest success scores even when faced with unseen objects, owing to its ability to incorporate embedded visual and linguistic features. However, our method exhibits a 1.0 lower performance in the behind-seen-colors task since our method does not account for volumes causing placement failures. In addition, SREM fails to attain scores in the simple task due to its dependence on a pre-defined instruction structure, in contrast to our LLM-based parser. We extend the evaluation using novel predicates (see Table 2). Our method achieves consistently high scores due to its ability to represent diverse distributions corresponding to not only close but also far, while the performance of the other methods degraded significantly. Although SREM is able to accommodate close with energy-based representation, the far predicates cannot be seamlessly accommodated by conventional rules or location representations. Grounding with multiple referring expressions. Our method significantly improves the performance of grounding given a composite instruction. Table 3 shows our method results in superior success scores in all four tasks. Our method is a maximum of 62 higher than the second-best approach in each benchmark since our LLM-based parser extracts relation tuples from diverse structures of instructions resolving compositional ambiguity. PARAGON shows a performance degradation in their benchmark since we trained PARAGON with the compositional task data only, unlike the paper setup. Although PARAGON s performance record in the paper is 67.9, the record is still 22.6 lower than our method result. SREM shows good performance in its benchmark, training its grounder and parser with the task and benchmark datasets, respectively, following the literature. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Benchmark CLIPORT SREM PARAGON Task packing -seen -google packing -unseen -google packing -unseen -shapes left -seen -colors left -unseen -colors right -seen -colors right -unseen -colors behind -seen -colors behind -unseen -colors front -seen -colors front -unseen -colors simple CLIPORT 98.0 97.7 95.5 88.0 86.5 90.5 98.5 99.5 96.0 98.5 99.0 38.5 PARAGON 98.1 98.0 99.5 75.5 61.5 87.5 86.0 99.0 97.5 99.0 96.5 41.5 (85.1) SREM 98.7 97.3 100 93.0 94.5 82.0 81.5 99.0 98.0 99.0 100 42.5 LINGO-Space 99.2 98.0 100 99.5 97.5 99.5 100 98.5 98.5 100 100 80.0 Table 1: Evaluation (success score) on 12 benchmark tasks with a single referring expression. The 12 tasks are from CLIPORT, PARAGON, and SREM, where methods are trained and tested within each task s dataset. The score indicates how successfully each method identified a location satisfying relations in [0, 100]. The number in the parentheses is the result of the literature. Task close-seen -colors close-unseen -colors far-seen -colors far-unseen -colors CLIPORT 38.5 18.5 59.5 60.0 PARAGON 38.5 41.5 31.5 42.0 SREM 91.0 90.5 45.0 44.5 LINGO-Space 86.0 81.0 95.5 95.0 Table 2: Evaluation (success score) on our 4 benchmark tasks with new predicates: close and far. Benchmark PARAGON SREM LINGO -Space Task composi tional comp-one -step -seen -colors comp-one -step -unseen -colors CLIPORT 26.0 87.5 84.8 54.4 PARAGON 28.5 (67.9) 76.4 77.3 56.0 SREM 1.5 93.4 92.1 42.2 LINGO-Space 90.5 97.5 96.5 79.1 Table 3: Evaluation (success score) on 4 benchmark tasks with multiple referring expressions. The number in the parentheses is the result of the literature. However, SREM often fails to parse the composite instructions in other datasets since it requires a specific format of composite instructions containing clause instead of phrase as a referring expression. Further, as shown in Fig. 5, our method shows the capability of handling multiple expressions with the consistently highest scores given an increasing number of referring expressions, while other approaches are going to fail to ground four to six referring expressions. This is because 1) the composite task includes multiple similar objects causing referential ambiguity and 2) the instruction includes more predicates without using clauses. However, our method resolves them by leveraging the semantic and geometric relationships encoded in the scene graph. Real-world demonstrations. We finally demonstrated the space-grounding capability of our LINGO-Space, integrating it on a navigation framework for a quadruped robot, Figure 5: Grounding performance on the increasing number of expressions. Each graph uses a distinct score metric: (left) the binary success score as PARAGON benchmark and (right) the success score. Spot, from Boston Dynamics. Fig. 1 illustrates the languageguided navigation experiment. A human operator delivers a command, such as move to the front of the red box and close to the tree. Our robot generates a scene graph using a Li DAR-based bounding box detector and parses the instruction using a Chat GPT (Open AI 2023). The robot then successfully identified the goal distribution, and it reached the best location we wanted.2 We introduced LINGO-Space, a language-conditioned incremental space-grounding method for composite instructions with multiple referring expressions. LINGO-space is a probabilistic grounding network leveraging a mixture of learnable polar distributions to predict probable location distributions. Our evaluation shows the effectiveness of the proposed probabilistic approach in representing desired space. Further, we demonstrate that, when coupled with the proposed LLM-guided semantic parser, our network enhances reasoning complex composite instructions with diverse referring expressions. Compared with state-of-the-art baselines, our model outperforms in grounding success score, generalizability, and scalability. We finally validate its practical usability by deploying LINGO-Space to a real-world navigation task running a quadruped robot. 2See details on our website: https://lingo-space.github.io The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Acknowledgements This work was partly supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No.2022-0-00311), National Research Foundation of Korea (NRF) grants funded by the Korea government (MSIT) (No.2021R1A4A3032834 and 2021R1C1C1004368). References Brohan, A.; Chebotar, Y.; Finn, C.; Hausman, K.; Herzog, A.; Ho, D.; Ibarz, J.; Irpan, A.; Jang, E.; Julian, R.; et al. 2023. Do as i can, not as i say: Grounding language in robotic affordances. In Proceedings of the Conference on Robot Learning (Co RL), 287 318. PMLR. Coumans, E.; and Bai, Y. 2016. Py Bullet, a Python module for physics simulation for games, robotics and machine learning. http://pybullet.org. Dong, L.; and Lapata, M. 2016. Language to logical form with neural attention. In Proceedings of the Association for Computational Linguistics (ACL), 33 43. Downs, L.; Francis, A.; Koenig, N.; Kinman, B.; Hickman, R.; Reymann, K.; Mc Hugh, T. B.; and Vanhoucke, V. 2022. Google scanned objects: A high-quality dataset of 3D scanned household items. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 2553 2560. IEEE. Driess, D.; Xia, F.; Sajjadi, M. S. M.; Lynch, C.; Chowdhery, A.; Ichter, B.; Wahid, A.; Tompson, J.; Vuong, Q.; Yu, T.; Huang, W.; Chebotar, Y.; Sermanet, P.; Duckworth, D.; Levine, S.; Vanhoucke, V.; Hausman, K.; Toussaint, M.; Greff, K.; Zeng, A.; Mordatch, I.; and Florence, P. 2023. Pa LM-E: An embodied multimodal language model. In Proceedings of the International Conference on Machine Learning (ICML). PMLR. Gkanatsios, N.; Jain, A.; Xian, Z.; Zhang, Y.; Atkeson, C. G.; and Fragkiadaki, K. 2023. Energy-based models are zero-shot planners for compositional scene rearrangement. In Proceedings of Robotics: Science and Systems (RSS). Guadarrama, S.; Riano, L.; Golland, D.; Go, D.; Jia, Y.; Klein, D.; Abbeel, P.; Darrell, T.; et al. 2013. Grounding spatial relations for human-robot interaction. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 1640 1647. IEEE. Hatori, J.; Kikuchi, Y.; Kobayashi, S.; Takahashi, K.; Tsuboi, Y.; Unno, Y.; Ko, W.; and Tan, J. 2018. Interactively picking real-world objects with unconstrained spoken language instructions. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 3774 3781. IEEE. He, K.; Gkioxari, G.; Doll ar, P.; and Girshick, R. 2017. Mask r-cnn. In Proceedings of the International Conference on Computer Vision (ICCV), 2961 2969. Howard, T.; Stump, E.; Fink, J.; Arkin, J.; Paul, R.; Park, D.; Roy, S.; et al. 2022. An intelligence architecture for grounded language communication with field robots. Field Robotics, 468 512. Howard, T. M.; Tellex, S.; and Roy, N. 2014. A natural language planner interface for mobile manipulators. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 6652 6659. IEEE. Hu, W.; Liu, B.; Gomes, J.; Zitnik, M.; Liang, P.; Pande, V.; and Leskovec, J. 2020. Strategies for pre-training graph neural networks. In Proceedings of the International Conference on Learning Representation (ICLR). Huang, W.; Xia, F.; Xiao, T.; Chan, H.; Liang, J.; Florence, P.; Zeng, A.; Tompson, J.; Mordatch, I.; Chebotar, Y.; et al. 2023. Inner monologue: Embodied reasoning through planning with language models. In Proceedings of the Conference on Robot Learning (Co RL), 1769 1782. PMLR. Jain, A.; Gkanatsios, N.; Mediratta, I.; and Fragkiadaki, K. 2022. Bottom up top down detection transformers for language grounding in images and point clouds. In Proceedings of the European Conference on Computer Vision (ECCV), 417 433. Springer. Jain, K.; Chhangani, V.; Tiwari, A.; Krishna, K. M.; and Gandhi, V. 2023. Ground then navigate: Language-guided navigation in dynamic scenes. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 4113 4120. IEEE. Kartmann, R.; Zhou, Y.; Liu, D.; Paus, F.; and Asfour, T. 2020. Representing spatial object relations as parametric polar distribution for scene manipulation based on verbal commands. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 8373 8380. IEEE. Kim, D.; Kim, Y.; Jang, J.; Song, M.; Choi, W.; and Park, D. 2023. SGGNet2: Speech-scene graph grounding network for speech-guided navigation. In Proceedings of the IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), 1648 1654. IEEE. Liang, J.; Huang, W.; Xia, F.; Xu, P.; Hausman, K.; Ichter, B.; Florence, P.; and Zeng, A. 2023. Code as policies: Language model programs for embodied control. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 9493 9500. IEEE. Liu, J. X.; Yang, Z.; Idrees, I.; Liang, S.; Schornstein, B.; Tellex, S.; and Shah, A. 2022. Lang2LTL: Translating natural language commands to temporal robot task specification. In The Workshop on Language and Robotics at Conference on robot learning. Liu, S.; Zeng, Z.; Ren, T.; Li, F.; Zhang, H.; Yang, J.; Li, C.; Yang, J.; Su, H.; Zhu, J.; et al. 2023. Grounding dino: Marrying dino with grounded pre-training for open-set object detection. ar Xiv preprint ar Xiv:2303.05499. Matuszek, C.; Fitz Gerald, N.; Zettlemoyer, L.; Bo, L.; and Fox, D. 2012. A joint model of language and perception for grounded attribute learning. In Proceedings of the International Conference on Machine Learning (ICML), 1435 1442. Mees, O.; Borja-Diaz, J.; and Burgard, W. 2023. Grounding language with visual affordances over unstructured data. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 11576 11582. IEEE. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Mees, O.; Emek, A.; Vertens, J.; and Burgard, W. 2020. Learning object placements for relational instructions by hallucinating scene representations. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 94 100. IEEE. Mildenhall, B.; Srinivasan, P. P.; Tancik, M.; Barron, J. T.; Ramamoorthi, R.; and Ng, R. 2021. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65(1): 99 106. Namasivayam, K.; Singh, H.; Bindal, V.; Tuli, A.; Agrawal, V.; Jain, R.; Singla, P.; and Paul, R. 2023. Learning neurosymbolic programs for language guided robot manipulation. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 7973 7980. IEEE. Open AI. 2023. Chat GPT (Aug 14 version). https://chat. openai.com/chat. Large language model. Paul, R.; Arkin, J.; Aksaray, D.; Roy, N.; and Howard, T. M. 2018. Efficient grounding of abstract spatial concepts for natural language interaction with robot platforms. International Journal of Robotics Research, 37(10): 1269 1299. Paxton, C.; Xie, C.; Hermans, T.; and Fox, D. 2022. Predicting stable configurations for semantic placement of novel objects. In Proceedings of the Conference on Robot Learning (Co RL), 806 815. PMLR. Radford, A.; Kim, J. W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. 2021. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning (ICML), 8748 8763. PMLR. Ramp aˇsek, L.; Galkin, M.; Dwivedi, V. P.; Luu, A. T.; Wolf, G.; and Beaini, D. 2022. Recipe for a general, powerful, scalable graph transformer. Conference on Neural Information Processing Systems (Neur IPS), 35: 14501 14515. Ren, A. Z.; Govil, B.; Yang, T.-Y.; Narasimhan, K. R.; and Majumdar, A. 2023. Leveraging language for accelerated learning of tool manipulation. In Proceedings of the Conference on Robot Learning (Co RL), 1531 1541. PMLR. Roy, S.; Noseworthy, M.; Paul, R.; Park, D.; and Roy, N. 2019. Leveraging past references for robust language grounding. In Proceedings of the Association for Computational Linguistics (ACL), 430 440. Shah, D.; Osi nski, B.; Levine, S.; et al. 2022. LM-Nav: Robotic navigation with large pre-trained models of language, vision, and action. In Proceedings of the Conference on Robot Learning (Co RL), 492 504. PMLR. Shridhar, M.; Manuelli, L.; and Fox, D. 2022. Cliport: What and where pathways for robotic manipulation. In Proceedings of the Conference on Robot Learning (Co RL), 894 906. PMLR. Shridhar, M.; Mittal, D.; and Hsu, D. 2020. INGRESS: Interactive visual grounding of referring expressions. International Journal of Robotics Research, 39(2-3): 217 232. Singh, I.; Blukis, V.; Mousavian, A.; Goyal, A.; Xu, D.; Tremblay, J.; Fox, D.; Thomason, J.; and Garg, A. 2023. Progprompt: Generating situated robot task plans using large language models. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 11523 11530. IEEE. Song, C. H.; Wu, J.; Washington, C.; Sadler, B. M.; Chao, W.-L.; and Su, Y. 2023. Llm-planner: Few-shot grounded planning for embodied agents with large language models. In Proceedings of the International Conference on Computer Vision (ICCV), 2998 3009. Stopp, E.; Gapp, K.-P.; Herzog, G.; Laengle, T.; and Lueth, T. C. 1994. Utilizing spatial relations for natural language access to an autonomous mobile robot. In Proceedings of the German Annual Conference on Artificial Intelligence, 39 50. Springer. Tan, J.; Ju, Z.; and Liu, H. 2014. Grounding spatial relations in natural language by fuzzy representation for human-robot interaction. In Proceedings of the IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), 1743 1750. IEEE. Tellex, S.; Kollar, T.; Dickerson, S.; Walter, M.; Banerjee, A.; Teller, S.; and Roy, N. 2011. Understanding natural language commands for robotic navigation and mobile manipulation. In Proceedings of the National Conference on Artificial Intelligence (AAAI), volume 25, 1507 1514. AAAI Press. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; and Polosukhin, I. 2017. Attention is all you need. Conference on Neural Information Processing Systems (Neur IPS), 30. Venkatesh, S. G.; Biswas, A.; Upadrashta, R.; Srinivasan, V.; Talukdar, P.; and Amrutur, B. 2021. Spatial reasoning from natural language instructions for robot manipulation. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 11196 11202. IEEE. Zeng, A.; Florence, P.; Tompson, J.; Welker, S.; Chien, J.; Attarian, M.; Armstrong, T.; Krasin, I.; Duong, D.; Sindhwani, V.; et al. 2021. Transporter networks: Rearranging the visual world for robotic manipulation. In Proceedings of the Conference on Robot Learning (Co RL), 726 747. PMLR. Zhao, Z.; Lee, W. S.; and Hsu, D. 2023. Differentiable parsing and visual grounding of natural language instructions for object placement. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), 11546 11553. IEEE. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)