# segment_any_3d_object_with_language__95ac4ede.pdf Published as a conference paper at ICLR 2025 SEGMENT ANY 3D OBJECT WITH LANGUAGE Seungjun Lee , Yuyang Zhao , Gim Hee Lee Department of Computer Science, National University of Singapore {seungjun.lee, yuyang.zhao, gimhee.lee}@comp.nus.edu.sg https://cvrp-sole.github.io (a) Can I wash my hands? (b) Brown Furnitures. (c) Device to play game. Figure 1: Qualitative results of SOLE with various language instructions. SOLE is highly generalizable and can segment corresponding instances with various language instructions, including but not limited to (a) visual questions, (b) attributes description, and (c) functional description. In this paper, we investigate Open-Vocabulary 3D Instance Segmentation (OV3DIS) with free-form language instructions. Earlier works mainly rely on annotated base categories for training which leads to limited generalization to unseen novel categories. To mitigate the poor generalizability to novel categories, recent works generate class-agnostic masks or projecting generalized masks from 2D to 3D, subsequently classifying them with the assistance of 2D foundation model. However, these works often disregard semantic information in the mask generation, leading to sub-optimal performance. Instead, generating generalizable but semantic-aware masks directly from 3D point clouds would result in superior outcomes. To the end, we introduce Segment any 3D Object with Languag E (SOLE), which is a semantic and geometric-aware visual-language learning framework with strong generalizability by generating semantic-related masks directly from 3D point clouds. Specifically, we propose a multimodal fusion network to incorporate multimodal semantics in both backbone and decoder. In addition, to align the 3D segmentation model with various language instructions and enhance the mask quality, we introduce three types of multimodal associations as supervision. Our SOLE outperforms previous methods by a large margin on Scan Netv2, Scan Net200, and Replica benchmarks, and the results are even closed to the fullysupervised counterpart despite the absence of class annotations in the training. Furthermore, extensive qualitative results demonstrate the versatility of our SOLE to language instructions. 1 INTRODUCTION 3D instance segmentation which aims at detecting, segmenting and recognizing object instances in 3D scenes is one of the crucial tasks for 3D scene understanding. Effective and generalizable 3D instance segmentation has great potential in real-world applications, including but not limited to autonomous driving, augmented reality (AR), and virtual reality (VR). Owing to its significance, 3D instance segmentation has achieved remarkable success in the computer vision community (Schult et al., 2022; Vu et al., 2022; He et al., 2022). Previous 3D instance segmentation models mainly focus on the closed-set setting, where the training and testing stages share the same categories. However, novel and unseen categories with various shapes and semantic meaning are inevitable in real-world applications. Failure to segment such instances drastically narrows the scope of application. Equal Contribution Published as a conference paper at ICLR 2025 (a) Class agnostic 3D masks and classification (b) Semantic-aware 3D masks (ours) Figure 2: Left (a) : Previous works train class-agnostic mask proposal module with only using mask annotations. In the inference time, generated 3D masks are projected to 2D images and subsequently classified with the help of 2D foundation model. Right (b) : In contrast, we train semantic-aware mask proposal module with giving Multi-Modal Associations (MMA) and mask annotations as supervision. With regarding both geometry and semantic information in the mask generation, our SOLE can produce highly generalizable segments. In view of the strong limitations of closed-set setting, open-set 3D instance segmentation (OS-3DIS) that aims at detecting and segmenting unseen classes based on instructions is introduced and investigated in the community. Most of the works (Huang et al., 2023b; Ding et al., 2022b; Nguyen et al., 2024) leverage category names or descriptions as segmentation instructions, which is also termed as open-vocabulary 3D instance segmentation (OV-3DIS). The early approaches (Ding et al., 2022b; Yang et al., 2023; Ding et al., 2023) split categories in each dataset into base and novel set. Only base categories are available in the training stage, but the model is expected to segment novel categories during inference. Due to the lack of novel classes during training, these methods easily overfit to the base categories and thus yielding sub-optimal performance on novel categories. In addition, they suffer from severe performance degradation when they are evaluated on the data with different distributions. In this regards, recent works (Takmaz et al., 2023; Huang et al., 2023b; Yan et al., 2024; Lu et al., 2023) explore more generalizable OV-3DIS with the help of 2D foundation models (Radford et al., 2021; Oquab et al., 2023; Zhou et al., 2022b). Specifically, Nguyen et al. (2024); Takmaz et al. (2023); Huang et al. (2023b) learn class-agnostic 3D masks from mask annotation and then project the point clouds to 2D images to obtain class labels from foundation models (Fig. 2a). Yan et al. (2024); Lu et al. (2023); Yin et al. (2024) predict 2D instances with 2D open-vocabulary instance segmentation model (Zhou et al., 2022b) and fuse them to obtain 3D predictions. However, class-agnostic masks and 2D projected masks ignore the semantic and geometry information in the mask generation, respectively, leading to the sub-optimal performance (details in Appendix. E.1). To this end, we investigate the OV-3DIS problem from the perspective of mask generator in this paper (Fig. 2b). We propose the semantic-aware mask generator to obtain semantic-related masks from 3D point clouds, yielding better and more generalizable 3D masks. In this paper, we propose SOLE: Segment any 3D Object with Languag E to circumvent the abovementioned issues for OV-3DIS. To realize generalizable open-set 3D instance segmentation, our SOLE requires two main attributes: generating and classifying 3D masks directly from 3D point clouds, and responsive to free-form language instructions. The 3D segmentation network is required to be aligned with language instructions to directly segment and classify instances from point clouds. To this end, we build a multimodal fusion network with two main techniques: 1) Point-wise CLIP features obtained from pre-trained multimodal 2D segmentation model (Ghiasi et al., 2022) are incorporated to the backbone to enhance the generality of the model; 2) Cross-modality decoder is introduced to integrate information from language-domain features, facilitating the effective fusion of multimodal knowledge. Furthermore, we improve the generalization ability across various scene and language instructions with a novel visual-language learning framework, training the 3D segmentation network with three types of multimodal associations: 1) Mask-visual association aligns the 3D visual information with 2D visual information in the foundation model; 2) mask-caption association aims at enabling the mask generation and classification from various language forms; and 3) mask-entity association introduces fine-grained textual information into the framework, enhancing the ability of instance segmentation. These associations improve the language instruction alignment and enhance the 3D mask prediction with more abundant semantic information. Equipped with a multimodal fusion network and three types of multimodal associations, our visuallanguage learning framework (SOLE) outperforms previous works by a large margin on Scan Netv2 (Dai et al., 2017), Scan Net200 (Rozenberszki et al., 2022) and Replica (Straub et al., 2019) benchmarks. Furthermore, SOLE can respond to free-form queries, including but not limited to Published as a conference paper at ICLR 2025 questions, attributes description, and functional description (Fig. 1 and Fig. 7). In summary, our contributions are as follows: We propose a visual-language learning framework for OV-3DIS, SOLE. A multimodal fusion network is designed for SOLE, which can directly predict semantic-related masks from 3D point clouds with multimodal information, leading to high-quality and generalizable segments. We propose three types of multimodal associations to improve the alignment between 3D segmentation model with the language. The associations improve the mask quality and the response ability to language instructions. SOLE achieves state-of-the-art results on Scan Netv2, Scannet200 and Replica benchmarks, and the results are even close to the fully-supervised counterpart. In addition, extensive qualitative results demonstrate that SOLE can respond to various language questions and instructions. 2 RELATED WORK Closed-Set 3D Instance Segmentation. 3D instance segmentation aims at detecting, segmenting and recognizing the object instances in 3D scenes. Previous works (He et al., 2021; 2022; Ngo et al., 2023; Schult et al., 2022; Sun et al., 2023; Vu et al., 2022; Yi et al., 2019; Zhang et al., 2021; Hou et al., 2019; Yang et al., 2019a; Chen et al., 2021; Dong et al., 2022; Jiang et al., 2020; Liu et al., 2022; Wu et al., 2022) mainly consider the closed-set setting, where the training and testing categories are the same. These methods vary in feature extraction and decoding process. With the development of transformer models, mask prediction becomes a more efficient and effective way than traditional box detection decoding approaches. Mask3D (Schult et al., 2022) samples a fixed number of points across the scene as queries, and then directly predicts the final masks with attention mechanism, achieving better results. However, closed-set methods lack the capability to handle the unseen categories and thus hindering their application in the real world. Open-Vocabulary 2D Segmentation. Owing to the recent success of large-scale vision-language models (Alayrac et al., 2022; Cherti et al., 2023; Girdhar et al., 2023; Jia et al., 2021; Radford et al., 2021; Yu et al., 2022; Yuan et al., 2021), notable achievements have been made in open-vocabulary or zero-shot 2D segmentation (Ding et al., 2022a; Ghiasi et al., 2022; Gu et al., 2021; He et al., 2023; Kuo et al., 2022; Li et al., 2022; Liang et al., 2023; Ma et al., 2022; Rao et al., 2022; Xu et al., 2022; 2023; Zabari & Hoshen, 2021; Zhou et al., 2022a; Cho et al., 2023; Ma et al., 2023; Yuan et al., 2024). The common key idea is to leverage 2D mulitmodal foundation models (Radford et al., 2021; Jia et al., 2021) for the transfer of image-level embeddings to the pixel-level downstream tasks. LSeg (Li et al., 2022), Open Seg (Ghiasi et al., 2022), and OVSeg (Liang et al., 2023) align pixel-level or mask-level visual features to text features from foundation model for open-vocabulary semantic segmentation. Recently, Open-Vocabulary SAM (Yuan et al., 2024) integrates CLIP and SAM, further exploring interactive open-world tasks. Other works such as X-Decoder (Zou et al., 2023), Free Seg (Qin et al., 2023) and SEEM (Zou et al., 2024) suggest more unified-framework for open-vocabulary segmentation, include instance, panoptic, and referring segmentation. Open-Vocabulary 3D Scene Understanding. The remarkable success achieved in open-vocabulary 2D segmentation (OV-2DS) has spurred several endeavors in open-vocabulary 3D segmentation. However, the techniques in OV-2DS cannot be directly transferred to the 3D domain due to the lack of 3D multimodal foundation model. Consequently, researchers propose to align 2D images and 3D point clouds and thus lifting 2D foundation models to 3D. For open-vocabulary 3D semantic segmentation, (Chen et al., 2023; Ding et al., 2022b; Ha & Song, 2022; Huang et al., 2023a; Jatavallabhula et al., 2023; Peng et al., 2023; Shafiullah et al., 2022; Shah et al., 2023) construct task-agnostic point-wise feature representations from 2D foundation models (Radford et al., 2021), and then use these features to query the open-vocabulary concepts within 3D scene. These works focus purely on transferring semantic information from 2D to 3D, limiting the application for instance-level recognition tasks. In this regard, open-vocabulary 3D instance segmentation (OV-3DIS) (Ding et al., 2023; Huang et al., 2023b; Takmaz et al., 2023; Lu et al., 2023; Nguyen et al., 2024; Yan et al., 2024) is introduced to detect and segment instances of various categories in 3D scenes. PLA (Ding et al., 2022b) and its variants (Yang et al., 2023; Ding et al., 2023) split the training categories into base and novel classes, and train the model only with base class annotation. Open Mask3D (Takmaz et al., 2023) and Open Ins3D (Huang et al., 2023b) learn class-agnostic 3D masks from mask annotations and then use the corresponding 2D images to obtain class labels from foundation models. With Published as a conference paper at ICLR 2025 Feature Backbone Per-Point CLIP Features Pooling Transformer Instance Queries Cross Modality Decoder Cross Attention Mask Features Self Attention Cross Attention Instance Queries Caption Features Point Features Mask Proposals 3D Point Cloud Feature Backbone Per-Point CLIP Features Transformer Class Features Mask Features Mask Proposals Pooled Features Masked Pooling Soft Geometric Mean Training Inference 3D Point Cloud Multimodal Associations Figure 3: Overall framework of SOLE. SOLE is built on transformer-based instance segmentation model with multimodal adaptations. For model architecture, backbone features are integrated with per-point CLIP features and subsequently fed into the cross-modality decoder (CMD). CMD aggregates the point-wise features and textual features into the instance queries, finally segmenting the instances, which are supervised by multimodal associations. During inference, predicted mask features are combined with the per-point CLIP features, enhancing the open-vocabulary performance. following similar paradigm, Open3DIS (Nguyen et al., 2024) further improves the quality of masks by fusing 3D segments and 2D masks. Recently, researchers also investigate direct lifting of 2D predictions from 2D instance segmentation model (Zhou et al., 2022b) to 3D without training (Yan et al., 2024; Lu et al., 2023; Yin et al., 2024). However, previous works often disregard semantic information in the mask generation, leading to the poor semantic generalization ability. Considering the limitations of previous works, we significantly improve OV-3DIS by designing a visual-language learning framework with a multimodal network and various multimodal associations. Objective. The goal of open-vocabulary 3D instance segmentation (OV-3DIS) with free-form language instructions is defined as follows: Given a 3D point cloud P RM C, the corresponding 2D images I and the instance-level 3D masks m, we aim to train a 3D instance segmentation network without ground-truth class annotations. During inference, given a text prompt q, the trained 3D instance segmentation network must detect and segment corresponding instances. Mask-Prediction Baseline. We build our framework on the transformer-based 3D instance segmentation model Mask3D (Schult et al., 2022), which treats the instance segmentation task as the mask prediction paradigm. Specifically, the transformer decoders with mask queries are used to segment instances. Given Nq queries selected from the scene, cross attention is used to aggregate information from the point clouds to instance queries. After several decoder layers, Nq queries become Nq masks with corresponding semantic prediction. During training, Hungarian matching (Kuhn, 1955) is adopted to match and train the model with ground truth labels and masks. At the inference stage, Nq masks with correct semantic classification results are taken as the final outputs. Our SOLE leverages the mask prediction paradigm with transformer-based architecture, where the model is only trained with masks without ground truth labels to achieve generalizable OV-3DIS. Overview. The overall architecture of SOLE is illustrated in Fig. 3. To realize open-vocabulary instance segmentation with free-form language instructions, we improve the transformer-based instance segmentation model with multimodal information: point-wise CLIP features in the backbone (Sec. 3.1) and textual information in the decoder (Sec. 3.2). Furthermore, to achieve better generalization ability without ground truth class labels, we construct three types of multimodal associations on target instances: mask-visual association, mask-caption association and mask-entity association Published as a conference paper at ICLR 2025 3D Point Cloud Per-Point CLIP Features Masked Pooling Image Captioning "A wooden desk with a chair in a room." "A blue chair sitting next to a desk in a wooden office" Mask-Caption Association Mask-Visual Association "a wooden desk" Noun Phrases "a blue chair" "a wooden office" Noun Phrases "A wooden desk with a chair in a room." "A blue chair sitting next to a desk in a wooden office" Mask-Entity Association "a wooden desk" x 0.7 "a chair" x 0.2 "a room" x 0.1 "a blue chair" x 0.8 "a desk" x 0.1 "a wooden office" x 0.1 Figure 4: Three types of multimodal association instance. For each ground truth instance mask, we first pool the per-point CLIP features to obtain Mask-Visual Association f MVA. Subsequently, f MVA is fed into CLIP space captioning model to generate caption and corresponding textual feature f MCA for each mask, termed as Mask-Caption Association. Finally, noun phrases are extracted from mask caption and the embeddings of them are aggregated via multimodal attention to get Mask-Entity Association f MEA. The three multimodal associations are used for supervising SOLE to acquire the ability to segment 3D objects with free-form language instructions. to train SOLE. Equipped with the multimodal framework and associations, our SOLE can effectively segment instances given various language prompts. 3.1 BACKBONE FEATURE ENSEMBLE Initializing the backbone with pre-trained model (Jia et al., 2022; Zhao et al., 2023b;a) is an efficient and effective way to improve the performance on the downstream tasks, especially when the downstream data is not in abundance. For 3D open-set setting, leveraging 2D foundation model is crucial due to the limited 3D data. We thus follow (Peng et al., 2023) to project pre-trained visual features of 2D images to 3D point clouds based on the camera pose. To maintain the fine-grained and generalizable features, we leverage Open Seg (Ghiasi et al., 2022) as the 2D backbone. These features contain visual information in the CLIP (Radford et al., 2021) feature space, which is aligned with textual information. Since CLIP feature space mainly focuses on semantic information due to the image-level contrastive training, leveraging the projected features solely cannot achieve optimal performance on instance segmentation. To this end, we train a 3D instance segmentation backbone and combine its features f b RM D with the projected 2D CLIP features f p RM C: \ t ilde {\m at h bf {f}} {{b}} = \mathrm {concat}(\mathbf {f} {{p}}, \mathbf {f} {{b}}) \in \mathbb {R} {M \times (C+D)}, (1) where M denotes the number of points while D and C denote the feature dimension of 3D instance segmentation backbone and the projected 2D features, respectively. Note that features of different resolutions are extracted from the 3D backbone and respectively incorporated with the 2D CLIP features. As illustrated in Fig. 3, the same pooling strategy with 3D backbone is adopted to CLIP features, aligning the resolution. Finally, incorporated point-wise features with multiple resolutions are fed into cross modality decoder. 3.2 CROSS MODALITY DECODER (CMD) Projected 2D CLIP features provide generalizable visual information but the language information is not explicitly integrated, limiting the responsive ability to language instructions. To circumvent this issue, we introduce Cross Modality Decoder (CMD) to incorporate textual information in the decoding process of our framework. Specifically, each CMD module contains three attention layers. Instance queries first extract visual information from the CLIP-combined backbone features f b. Published as a conference paper at ICLR 2025 CLIP textual features are then projected to key and value in the second attention layer, incorporating the text domain knowledge. During the training, CLIP textual features are obtained from the caption features of each target mask, f MCA RNc C (See Sec. 3.3 for details), whereas, during the inference, it can be the description of the query instance or other form of language instructions, such as visual questions or functional attributes. Finally, self-attention is applied to the instance queries to further improve the representation. By fusing the multimodal knowledge from CLIP with multilevel CMD as the decoder, SOLE can respond to various language instructions with high-quality results. 3.3 VISION-LANGUAGE LEARNING We do vision-language learning to enable our SOLE towards generalizable OV-3DIS. To respond effectively to various language instructions, we leverage multimodal information stemming from target mask annotations, to supervise the segmentation network. Specifically, three types of supervision in hierarchical granularity are proposed: 1) mask-visual association, 2) mask-caption association and 3) mask-entity association. Mask-Visual Association (MVA). Using the correspondence between 2D images and 3D point clouds, we can get the instance-level CLIP visual features f MVA RNm C by averaging the perpoint CLIP features within the Nm target instance masks m = [m1, m2, . . . , m Nm]. The instancelevel CLIP visual features can serve as the supervision to indirectly align the 3D segmentation model to CLIP textual space. In addition, as the intermediate representation between 3D point cloud and language, the mask-visual association is also the basis for the following two fine-grained associations. Mask-Caption Association (MCA). Despite being in the CLIP feature space, mask-visual association is not an accurate and precision language supervision. Instead, directly supervising the model with language instructions would yield better results. Due to the strong generalization ability of CLIP (Radford et al., 2021), text generation from CLIP space is widely investigated in the community (Tewel et al., 2022; Mokady et al., 2021; Li et al., 2023). Since the instance-level CLIP visual features f MVA in the mask-visual association is in CLIP visual space, we can feed them to the CLIP space caption generation model (De Cap (Li et al., 2023)) to obtain the mask captions c = [c1, c2, ..., c Nm]. The mask captions are then fed into CLIP textual model to extract the maskcaption association f MCA. This association represents the language information for the instance masks, used in CMD to fuse textual information during the training. Mask-Entity Association (MEA). Although mask-caption association can provide detailed language descriptions for both semantics and geometry, it may be ambiguous for specific categories. As shown in the example of Fig. 4. The mask caption for a desk is A wooden desk with a chair in a room . Such caption can lead to the confusion of the model between the chair and the desk, or misinterpretation of the two instances as a single one. It is therefore important to introduce a more fine-grained visual-language association for better semantic learning. Since the objects are commonly the nouns in the caption, we can extract the entity-level descriptions for the nouns and match them with the instances. Specifically, as illustrated in Fig. 4, we first extract all the noun phrases ei for each mask caption ci and obtain the text feature of each noun phrase from CLIP text encoder T as below: \ma t hc a l {E }(c _ i ) = \mat h b f { e}_i = [ e_1, e_2, \dots , e_{N_e i}], \quad \mathbf {f} {\mathbf {e}}_i = \mathcal {T}(\mathbf {e}_i) \in \mathbb {R} {{N_e i} \times C}, (2) where E( ) denotes the NLP tool to extract noun phrases and N i e denotes the number of nouns obtained from mask caption ci. The entities can be matched to the mask in either a hard or soft manner. Intuitively, the most similar entity can be viewed as the mask label. However, there are two main issues with such a hard matching. First, the generated caption and the similarity results may not be accurate, leading to wrong supervision. Second, although the entity is correct, hard matching ignores the geometry information in the context and thus impairing the responsive ability to language instructions. To this end, we propose a soft matching to get mask-entity association by multimodal attention. Specifically, the aggregated entity feature for the i-th mask f MEA i is obtained based on the attention map Ac,e constructed by cosine similarity cos( , ) between mask feature and Published as a conference paper at ICLR 2025 entity features: \m a t hbf { f } \ }_i = A_ { c , e} \c dot \ mat h bf { f } { \ m a th b f { e }} _ i = \sum _k {N_e i} \frac {\exp {\left (\sj {\mathrm {cos}\left (\mathbf {f} \text {MVA}_i, \mathbf {f} {e_k}_i\right )}\right )}}{\sum _j {N_e i} \exp {\left (\sj {\mathrm {cos}\left (\mathbf {f} \text {MVA}_i, \mathbf {f} {e_j}_i\right )}\right ) } } \cdot \mathbf {f} {e_k}_i, (3) where f MVA i denotes the mask-visual association feature for i-th mask, and f ek i is the CLIP textual feature for k-th entity in the i-th mask caption. With the aggregated entity feature, the 3D mask can be aligned with a specific instance category. 3.4 TRAINING AND INFERENCE Training. The three types of multimodal associations are effective supervision to learn a generalizable 3D instance segmentation model. We follow the mask prediction paradigm to train the segmentation model, which matches the ground truth instances with the predicted masks via Hungarian matching (Kuhn, 1955). Specifically, the matching cost between i-th predicted mask and j-th ground truth instance is calculated as: \b eg i n {al i g n ed} \ m a t h cal {C } ( i , j ) = & - \la m bd a _ {\m a t h r m {MM A }}\ l eft ( \sj {p\ le f t (\mathrm {cos}\left (\mathbf {f} {\mathbf {m}}_i, \mathbf {f} \text {MVA}_j \right )\right ) + p\left (\mathrm {cos}\left (\mathbf {f} {\mathbf {m}}_i, \mathbf {f} \text {MCA}_j \right )\right ) + p\left (\mathrm {cos}\left (\mathbf {f} {\mathbf {m}}_i, \mathbf {f} \text {MEA}_j \right ) \right )}\right ) \\ &+ \lambda _{\mathrm {dice}}\mathcal {L}_{\mathrm {dice}}(i, j) + \lambda _{\mathrm {BCE}}\mathcal {L}_{\mathrm {BCE}}(i, j), \end {aligned} where p( ) denotes the softmax probability following the cosine similarity between the predicted instance and the ground truth. After matching the masks and ground truth instances, the model is trained with the combination of mask and semantic loss. Specifically, all three types of associations are used to semantically supervise the model. For each association, we follow (Zhou et al., 2023) to use the combination of focal loss (Lin et al., 2017) and dice loss, which can ensure the segmentation result for each class is independently generated. The semantic multimodal association loss Lj MMA for j-th ground truth mask is: a l {L}_{\ma thrm {M M A } } j = \su m _a \l e f t ( \mathcal {L}_{\mathrm {focal}}(\hat {p}_{\sigma (j)} a, y_j a) + \mathcal {L}_{\mathrm {dice}} (\hat {p}_{\sigma (j)} a, y_j a) \right ), (5) where a {MVA, MCA, MEA} denotes three types of associations and ya j is the binary label for matching. ˆpa σ(j) = sigmoid(f m σ(j) f a j ) is the semantic probability between the prediction with the association a. The overall training loss is the combination of mask loss and semantic loss: l {L} = \f r ac {1}{N_m} \sum _ j { N _m} \left ( \lam bda _{\mathrm {MMA}} \mathcal {L}_{\mathrm {MMA}} j + \lambda _{\mathrm {dice}}\mathcal {L}_{\mathrm {dice}}(\hat {m}_{\sigma (j)}, m_j) + \lambda _{\mathrm {BCE}}\mathcal {L}_{\mathrm {BCE}}(\hat {m}_{\sigma (j)}, m_j) \right ), (6) where ˆmσ(j) denotes matched predicted mask with j-th target mask. Inference. During inference, we combine the visual feature from CLIP with the predicted mask feature to achieve better generalization ability. Specifically, after obtaining the 3D masks, per-point CLIP features are pooled within the mask. The pooled CLIP feature and mask feature are then given to the text classifier f t to obtain the respective classification probability Pm = p(cos(f m, f t)) and Pp = p(cos(f p, f t)), and the final probability is yielded by soft geometric mean between them: \ lab el { eq:s o ft_ geom etric_ mean} \sj {\mathbf {P} = \max \left (\mathbf {P} m,\mathbf {P} p\right ) \tau \cdot \min \left (\mathbf {P} m,\mathbf {P} p\right ) {1-\tau },} (7) where τ is the exponent to increase confidence, which we set to 0.667 in this paper. For benchmark evaluation, we use CLIP textual features of all category names as the classifier. For responding to other language instructions, we use the CLIP textual feature of corresponding language instruction as binary classifier. 4 EXPERIMENTS 4.1 EXPERIMENTAL SETTING Datasets. We evaluate SOLE on the popular scene understanding datasets: Scan Netv2 (Dai et al., 2017), Scan Net200 (Rozenberszki et al., 2022) and Replica (Straub et al., 2019) in both closedset and open-set 3D instance segmentation tasks. Scan Netv2 (Dai et al., 2017) is a popular indoor Published as a conference paper at ICLR 2025 Table 1: The comparison of closed-set 3D instance segmentation setting on Scan Netv2. SOLE is compared with mask-training methods and the fully-supervised counterpart (upper bound). SOLE outperforms all the OV-3DIS methods and achieves competitive results with the fully-supervised model. The best results are in bold while the second best results are underscored. Method AP AP50 AP25 voxel size Open Ins3D (Huang et al., 2023b) - 28.7 38.9 2cm Open Mask3D (Takmaz et al., 2023) 31.0 39.5 44.0 2cm Open3DIS (only 3D) (Nguyen et al., 2024) 31.3 42.4 47.8 2cm SOLE w 4cm voxel size 30.8 52.5 70.9 4cm SOLE w/o text sup 35.0 50.2 60.2 2cm SOLE (ours) 44.4 (+13.1) 62.2 (+19.8) 71.4 (+23.6) 2cm Mask3D (Schult et al., 2022) (fully sup) 55.2 73.7 83.5 2cm Table 2: The comparison of closed-set 3D instance segmentation setting on Scan Net200. SOLE is compared with mask training methods on the overall segmentation performance and on each subset. SOLE significantly outperforms state-of-the-art methods on five out of the six evaluation metrics under the same conditions using only proposals from a 3D Network. Method AP AP50 AP25 APhead APcom APtail Open Ins3D (Huang et al., 2023b) 8.8 10.3 14.4 16.0 6.5 4.2 Open Mask3D (Takmaz et al., 2023) 15.4 19.9 23.1 17.1 14.1 14.9 Open3DIS (only 3D) (Nguyen et al., 2024) 18.6 23.1 27.3 24.7 16.9 13.3 SOLE (ours) 20.1 Mask3D (fully sup) (Schult et al., 2022) 26.9 36.2 41.4 39.8 21.7 17.9 point cloud dataset with 18 instance classes, where other furniture class is disregarded due to its ambigity. Scan Net200 (Rozenberszki et al., 2022) is a fine-grained annotated version of Scan Netv2 that contains 200 classes of head (66 categories), common (68 categories) and tail (66 categories) subsets. For Scan Netv2 and Scan Net200, we evaluate the closed-set setting and the hierarchical open-set setting. Replica (Straub et al., 2019) is a high-quality synthetic dataset annotated with 48 instance categories. Following (Takmaz et al., 2023), we evaluate on eight scenes in Replica for open-set instance segmentation, including {office0, office1, office2, office3, office4, room0, room1 and room2.} Implementation Details. Following the Mask3D (Schult et al., 2022), we adopt Minkowski UNet (Choy et al., 2019) as backbone. The feature backbone extracts point features in 5 scales, while 4 layers of transformer decoder iteratively refine the instance queries. Our model is trained for 600 epochs with Adam W (Loshchilov & Hutter, 2017) optimizer. The learning rate is set to 1 10 4 with cyclical decay. In training, we set λMMA = 20.0, λdice = 2.0 and λBCE = 5.0 as the loss weight. Baselines. We compare SOLE mainly with recent mask-training methods (Takmaz et al., 2023; Huang et al., 2023b; Nguyen et al., 2024). Mask-training methods (Takmaz et al., 2023; Huang et al., 2023b; Nguyen et al., 2024) train class-agnostic mask generator with mask annotations and get the semantic prediction with 2D foundation models. Evaluation is conducted under the same conditions using only proposals from a 3D Network on all the categories. The proposed semanticaware mask generator can be combined with the classification techniques proposed by previous works, further improving the performance (Appendix. D.2). Evaluation Metric. Average precision (AP) of different Io U thresholds is adopted as the evaluation metric, including AP under 25%, 50% Io U and the average AP from 50% to 95% Io U. 4.2 COMPARISON WITH PREVIOUS METHODS Closed-Set 3D Instance Segmentation. We compare our SOLE with mask-training methods (Takmaz et al., 2023; Huang et al., 2023b; Nguyen et al., 2024) on the closed-set 3D instance segmenta- Published as a conference paper at ICLR 2025 Table 3: The comparison of hierarchical open-set 3D instance segmentation setting on Scan Netv2 (Dai et al., 2017) Scan Net200 (Rozenberszki et al., 2022). SOLE is compared with Open Mask3D and Open3DIS on both base and novel classes and achieves the best results. Method Novel Classes Base Classes All Classes AP AP50 AP25 AP AP50 AP25 AP APtail Open Mask3D (Takmaz et al., 2023) 11.9 15.2 17.8 14.3 18.3 21.2 12.6 11.5 Open3DIS (only 3D) (Nguyen et al., 2024) 14.9 19.2 22.1 16.5 22.4 26.2 15.3 10.9 SOLE (ours) 19.1 Table 4: The comparison of open-set 3D instance segmentation setting on Scan Net200 (Rozenberszki et al., 2022) Replica (Straub et al., 2019). SOLE outperforms state-of-the-art methods (Takmaz et al., 2023; Nguyen et al., 2024) on all the evaluation metrics. Method Mask Training AP AP50 AP25 Open Ins3D (Huang et al., 2023b) Scan Net200 13.6 18.0 19.7 Open Mask3D (Takmaz et al., 2023) Scan Net200 13.1 18.4 24.2 Open3DIS (only 3D) (Nguyen et al., 2024) Scan Net200 14.9 18.8 23.6 SOLE (ours) Scan Net200 24.7 (+9.8) 31.8 (+13.0) 40.3 (+16.1) tion setting. From the comparison results in Tab. 1, we can make the following observations. First, with the same mask supervision, SOLE significantly surpasses state-of-the-art method (Nguyen et al., 2024) by 13.1%, 19.8%, and 23.6% on AP, AP50 and AP25, respectively. Second, our SOLE can even achieve competitive performance with the fully-supervised counterpart (44.4% v.s. 55.2% in AP) despite not using the class labels. Finally, we provide two variants of SOLE to further verify our effectiveness. SOLE w 4cm voxel size leverages 4cm voxel size instead of 2cm as in previous works. Larger voxel size can save the memory requirements and speed up the model with the loss of precision. Despite using a small voxel size, SOLE w 4cm voxel size is still on par with Open Mask3D and Open3DIS in AP, and largely outperforms them in AP50 and AP25. Furthermore, we verify that the effectiveness of our framework is not limited to the caption model and NLP tools by conducting experiments without any additional textual information, i.e. SOLE w/o text sup. In this experiment, mask-caption association and mask-entity association is removed since the caption is not available. The model can still achieve the state-of-the-art performance despite only trained with mask-visual association. Additionally, we compare SOLE with mask-training methods on Scan Net200 (Rozenberszki et al., 2022) in Tab. 2. All of the methods are evaluated on the overall segmentation performance and the performance on each of the three subsets. SOLE outperforms state-of-the-art methods (Nguyen et al., 2024; Takmaz et al., 2023) on five out of the six metrics and achieves comparable performance on the tail classes. The results on Scan Net200 (Rozenberszki et al., 2022) further demonstrate the effectiveness of our framework. Hierarchical and Cross-Domain Open-Set 3DIS. To evaluate the generalization capability of our work, we compare our SOLE with mask training methods (Huang et al., 2023b; Takmaz et al., 2023; Nguyen et al., 2024) in open-set setting, using Scannet200 (Rozenberszki et al., 2022) and Replica (Straub et al., 2019) datasets. For Scan Net200, both models are trained with mask annotations in Scan Netv2 (Dai et al., 2017). Following (Takmaz et al., 2023), 53 classes that are semantically close to the Scan Net, are grouped as Base . The remaining 147 classes are grouped as Novel . Both in-distribution ( base ) and out-of-distribution ( novel ) classes are reported in Tab. 3. Our SOLE outperforms Open Mask3D (Takmaz et al., 2023) and Open3DIS (Nguyen et al., 2024) by a large margin on both base and novel classes. Furthermore, to verify the generalization ability of SOLE when both domain shift and category shift exist, we compare our framework with all of the mask training methods on the synthetic Replica benchmark (Straub et al., 2019). Models are trained on the annotated masks on Scan Net200. As shown in Tab. 4, our method further shows superior robustness on more out-of-distribution data from Replica, achieving +9.8% improvement in AP score compared to Open3DIS. 4.3 ABLATION STUDIES AND ANALYSIS In this section, we conduct several ablation studies to validate our design choices. All of the studies are evaluated on Scan Netv2 (Dai et al., 2017) dataset. Published as a conference paper at ICLR 2025 Table 5: Multimodal fusion network. No. f p f b CMD AP AP50 AP25 voxel size 1 18.7 36.4 58.1 4cm 2 25.4 47.0 66.0 4cm 3 30.8 52.5 70.9 4cm 4 42.8 60.5 68.9 2cm 5 44.4 62.2 71.4 2cm Table 6: Multimodal associations. No. f MVA f MCA f MEA AP AP50 AP25 voxel size 1 24.5 42.0 56.0 4cm 2 30.4 53.0 68.7 4cm 3 32.1 53.8 70.0 4cm 4 29.1 50.9 66.8 4cm 5 30.3 53.7 70.4 4cm 6 30.8 52.5 70.9 4cm (b) f MV A, f MCA (c) f MV A, f MCA, f MEA Figure 5: Qualitative analysis on multimodal associations. Given the free-form language instruction, I wanna see outside. , SOLE trained only with f MEA captures the wrong object ((a)), whereas it segments the related object when f MVA and f MCA are given as the supervision ((b) and (c)). Multimodel Fusion Network. In Tab. 5, we conduct component analysis on multimodal fusion network, validating the effectiveness of backbone feature ensemble and Cross-Modality Decoder (CMD). As for the backbone feature ensemble, leveraging projected 2D CLIP features f p (first row) as only backbone can have better semantic information but lack the 3D geometry detection ability, leading to poor semantic recognition ability. In contrast, solely using 3D instance backbone feature f b (second row) cannot inherit the generalizable semantic information, resulting in sub-optimal performance. Combining the two features (third row) can make full use of generalized semantic information while learning good geometry detection ability from 3D masks, yielding optimal results. Additionally, Cross Modality Decoder (CMD) can further enhance the ability to understand language instructions, improving AP by 1.6%. Multimodal Associations. We analyze the components of multimodal associations (f MVA, f MCA, and f MEA) in Tab. 6, reporting the scores of various combinations on Scan Netv2 (Dai et al., 2017) with 4cm voxel size. We have the following observations. First, using any of multimodal associations can already achieve significant performance, outperforming previous state-of-the-art method (Open Ins3D (Huang et al., 2023b)) with larger voxel size (lower resolution). Second, among the three types of associations, mask-entity association f MEA is the most effective one on evaluation metrics since it can align the masks with specific categories. Third, when combining f MEA with the other two associations, the model suffers from performance degradation on AP and AP50 while the performance improves on AP25. This observation shows that mask-visual association and maskcaption association can help semantic learning but impair mask accuracy. To this end, we further illustrate qualitative results in Fig. 5. Given a free-form language instruction instead of category name, e.g., I wanna see outside , the model only using mask-entity association cannot segment the correct instance (Fig. 5a) while the model incorporating the other associations (Fig. 5b and Fig. 5c) can. Therefore, despite slightly impairing the performance on benchmark, mask-visual association and mask-caption association are crucial to recognizing free-form language instructions, benefiting the applications in real-world scenarios. 5 CONCLUSION In this paper, we propose a novel framework, SOLE, for open-vocabulary 3D instance segmentation with free-form language instructions. SOLE contains a multimodal fusion network and is supervised with three types of multimodal associations, aiming at aligning the model with various free-form language instructions. Our framework outperforms previous methods by a large margin on three benchmarks while achieving competitive performance with the fully-supervised counterpart. Moreover, extensive qualitative results demonstrate the versatility of our SOLE to language instructions. Published as a conference paper at ICLR 2025 ACKNOWLEDGEMENTS This research work is supported by the Agency for Science, Technology and Research (A*STAR) under its MTC Programmatic Funds (Grant No. M23L7b0021), and the Tier 2 grant MOET2EP20124-0015 from the Singapore Ministry of Education. ETHICS STATEMENT This research adheres to the ethical guidelines outlined in the ICLR Code of Ethics. In this section, we take following aspects into account for the ethical considerations. Data and Privacy : The datasets used in this research are either publicly available (Straub et al., 2019) or accessible under the agreement on term of use (Dai et al., 2017; Rozenberszki et al., 2022). Potential Harmful Applications : The proposed method for open-vocabulary 3D instance segmentation could be used in sensitive domains such as surveillance. We encourage future users to apply the model responsibly, with appropriate safeguards to mitigate potential misuse. Research Integrity : We confirm that all experiments were conducted in accordance with best practices in research ethics, and results were reported with complete transparency and accuracy. REPRODUCIBILITY STATEMENT To ensure the reproducibility of our work, we state specific implementation details in Sec. 4.1 and Appendix. Sec. A. The code will be made publicly available. Jean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, Arthur Mensch, Katherine Millican, Malcolm Reynolds, et al. Flamingo: a visual language model for few-shot learning. Advances in Neural Information Processing Systems, 35:23716 23736, 2022. 3 Boyuan Chen, Fei Xia, Brian Ichter, Kanishka Rao, Keerthana Gopalakrishnan, Michael S Ryoo, Austin Stone, and Daniel Kappler. Open-vocabulary queryable scene representations for real world planning. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 11509 11522. IEEE, 2023. 3 Dave Zhenyu Chen, Angel X Chang, and Matthias Nießner. Scanrefer: 3d object localization in rgb-d scans using natural language. In European conference on computer vision, pp. 202 221. Springer, 2020. 17, 18 Shaoyu Chen, Jiemin Fang, Qian Zhang, Wenyu Liu, and Xinggang Wang. Hierarchical aggregation for 3d instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 15467 15476, 2021. 3 Mehdi Cherti, Romain Beaumont, Ross Wightman, Mitchell Wortsman, Gabriel Ilharco, Cade Gordon, Christoph Schuhmann, Ludwig Schmidt, and Jenia Jitsev. Reproducible scaling laws for contrastive language-image learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2818 2829, 2023. 3 Seokju Cho, Heeseong Shin, Sunghwan Hong, Seungjun An, Seungjun Lee, Anurag Arnab, Paul Hongsuck Seo, and Seungryong Kim. Cat-seg: Cost aggregation for open-vocabulary semantic segmentation. ar Xiv preprint ar Xiv:2303.11797, 2023. 3 Christopher Choy, Jun Young Gwak, and Silvio Savarese. 4d spatio-temporal convnets: Minkowski convolutional neural networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 3075 3084, 2019. 8 Published as a conference paper at ICLR 2025 Angela Dai, Angel X Chang, Manolis Savva, Maciej Halber, Thomas Funkhouser, and Matthias Nießner. Scannet: Richly-annotated 3d reconstructions of indoor scenes. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5828 5839, 2017. 2, 7, 9, 10, 11, 18, 19 Jian Ding, Nan Xue, Gui-Song Xia, and Dengxin Dai. Decoupling zero-shot semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11583 11592, 2022a. 3 Runyu Ding, Jihan Yang, Chuhui Xue, Wenqing Zhang, Song Bai, and Xiaojuan Qi. Pla: Languagedriven open-vocabulary 3d scene understanding. ar Xiv preprint ar Xiv:2211.16312, 2022b. 2, 3 Runyu Ding, Jihan Yang, Chuhui Xue, Wenqing Zhang, Song Bai, and Xiaojuan Qi. Lowis3d: Language-driven open-world instance-level 3d scene understanding. ar Xiv preprint ar Xiv:2308.00353, 2023. 2, 3 Shichao Dong, Guosheng Lin, and Tzu-Yi Hung. Learning regional purity for instance segmentation on 3d point clouds. In European Conference on Computer Vision, pp. 56 72. Springer, 2022. 3 Martin Ester, Hans-Peter Kriegel, J org Sander, Xiaowei Xu, et al. A density-based algorithm for discovering clusters in large spatial databases with noise. In KDD, 1996. 17 Golnaz Ghiasi, Xiuye Gu, Yin Cui, and Tsung-Yi Lin. Scaling open-vocabulary image segmentation with image-level labels. In European Conference on Computer Vision, pp. 540 557. Springer, 2022. 2, 3, 5 Rohit Girdhar, Alaaeldin El-Nouby, Zhuang Liu, Mannat Singh, Kalyan Vasudev Alwala, Armand Joulin, and Ishan Misra. Imagebind: One embedding space to bind them all. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15180 15190, 2023. 3 Xiuye Gu, Tsung-Yi Lin, Weicheng Kuo, and Yin Cui. Open-vocabulary object detection via vision and language knowledge distillation. ar Xiv preprint ar Xiv:2104.13921, 2021. 3 Huy Ha and Shuran Song. Semantic abstraction: Open-world 3d scene understanding from 2d vision-language models. ar Xiv preprint ar Xiv:2207.11514, 2022. 3 Sunan He, Taian Guo, Tao Dai, Ruizhi Qiao, Xiujun Shu, Bo Ren, and Shu-Tao Xia. Openvocabulary multi-label classification via multi-modal knowledge transfer. In Proceedings of the AAAI Conference on Artificial Intelligence, 2023. 3 Tong He, Chunhua Shen, and Anton Van Den Hengel. Dyco3d: Robust instance segmentation of 3d point clouds through dynamic convolution. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 354 363, 2021. 3 Tong He, Wei Yin, Chunhua Shen, and Anton van den Hengel. Pointinst3d: Segmenting 3d instances by points. In European Conference on Computer Vision, pp. 286 302. Springer, 2022. 1, 3 Yining Hong, Haoyu Zhen, Peihao Chen, Shuhong Zheng, Yilun Du, Zhenfang Chen, and Chuang Gan. 3d-llm: Injecting the 3d world into large language models. Advances in Neural Information Processing Systems, 36:20482 20494, 2023. 18 Matthew Honnibal and Ines Montani. spa Cy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing. To appear, 2017. 17 Ji Hou, Angela Dai, and Matthias Nießner. 3d-sis: 3d semantic instance segmentation of rgb-d scans. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 4421 4430, 2019. 3 Ronghang Hu, Huazhe Xu, Marcus Rohrbach, Jiashi Feng, Kate Saenko, and Trevor Darrell. Natural language object retrieval. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4555 4564, 2016. 17, 18 Published as a conference paper at ICLR 2025 Chenguang Huang, Oier Mees, Andy Zeng, and Wolfram Burgard. Visual language maps for robot navigation. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 10608 10615. IEEE, 2023a. 3 Zhening Huang, Xiaoyang Wu, Xi Chen, Hengshuang Zhao, Lei Zhu, and Joan Lasenby. Openins3d: Snap and lookup for 3d open-vocabulary instance segmentation. ar Xiv preprint ar Xiv:2309.00616, 2023b. 2, 3, 8, 9, 10, 19, 20 Krishna Murthy Jatavallabhula, Alihusein Kuwajerwala, Qiao Gu, Mohd Omama, Tao Chen, Alaa Maalouf, Shuang Li, Ganesh Iyer, Soroush Saryazdi, Nikhil Keetha, et al. Conceptfusion: Openset multimodal 3d mapping. ar Xiv preprint ar Xiv:2302.07241, 2023. 3 Baoxiong Jia, Yixin Chen, Huangyue Yu, Yan Wang, Xuesong Niu, Tengyu Liu, Qing Li, and Siyuan Huang. Sceneverse: Scaling 3d vision-language learning for grounded scene understanding. ar Xiv preprint ar Xiv:2401.09340, 2024. 21 Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun-Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. In International conference on machine learning, pp. 4904 4916. PMLR, 2021. 3 Menglin Jia, Luming Tang, Bor-Chun Chen, Claire Cardie, Serge Belongie, Bharath Hariharan, and Ser-Nam Lim. Visual prompt tuning. In European Conference on Computer Vision, pp. 709 727. Springer, 2022. 5 Li Jiang, Hengshuang Zhao, Shaoshuai Shi, Shu Liu, Chi-Wing Fu, and Jiaya Jia. Pointgroup: Dualset point grouping for 3d instance segmentation. In Proceedings of the IEEE/CVF conference on computer vision and Pattern recognition, pp. 4867 4876, 2020. 3 Harold W Kuhn. The hungarian method for the assignment problem. Naval research logistics quarterly, 2(1-2):83 97, 1955. 4, 7 Weicheng Kuo, Yin Cui, Xiuye Gu, AJ Piergiovanni, and Anelia Angelova. F-vlm: Open-vocabulary object detection upon frozen vision and language models. ar Xiv preprint ar Xiv:2209.15639, 2022. 3 Boyi Li, Kilian Q. Weinberger, Serge Belongie, Vladlen Koltun, and Ren e Ranftl. Language-driven semantic segmentation, 2022. 3 Wei Li, Linchao Zhu, Longyin Wen, and Yi Yang. Decap: Decoding clip latents for zero-shot captioning via text-only training, 2023. 6, 17 Feng Liang, Bichen Wu, Xiaoliang Dai, Kunpeng Li, Yinan Zhao, Hang Zhang, Peizhao Zhang, Peter Vajda, and Diana Marculescu. Open-vocabulary semantic segmentation with mask-adapted clip. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7061 7070, 2023. 3 Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Doll ar. Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision, pp. 2980 2988, 2017. 7 Jiaheng Liu, Tong He, Honghui Yang, Rui Su, Jiayi Tian, Junran Wu, Hongcheng Guo, Ke Xu, and Wanli Ouyang. 3d-queryis: A query-based framework for 3d instance segmentation. ar Xiv preprint ar Xiv:2211.09375, 2022. 3 Steven Loria et al. textblob documentation. Release 0.15, 2(8):269, 2018. 17 Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. 8 Shiyang Lu, Haonan Chang, Eric Pu Jing, Abdeslam Boularias, and Kostas Bekris. Ovir-3d: Openvocabulary 3d instance retrieval without training on 3d data. In Conference on Robot Learning, pp. 1610 1620. PMLR, 2023. 2, 3, 4 Published as a conference paper at ICLR 2025 Chaofan Ma, Yuhuan Yang, Yanfeng Wang, Ya Zhang, and Weidi Xie. Open-vocabulary semantic segmentation with frozen vision-language models. ar Xiv preprint ar Xiv:2210.15138, 2022. 3 Chaofan Ma, Yuhuan Yang, Chen Ju, Fei Zhang, Ya Zhang, and Yanfeng Wang. Attrseg: openvocabulary semantic segmentation via attribute decomposition-aggregation. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. 3 Ron Mokady, Amir Hertz, and Amit H Bermano. Clipcap: Clip prefix for image captioning. ar Xiv preprint ar Xiv:2111.09734, 2021. 6 Tuan Duc Ngo, Binh-Son Hua, and Khoi Nguyen. Isbnet: a 3d point cloud instance segmentation network with instance-aware sampling and box-aware dynamic convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13550 13559, 2023. 3 Phuc Nguyen, Tuan Duc Ngo, Evangelos Kalogerakis, Chuang Gan, Anh Tran, Cuong Pham, and Khoi Nguyen. Open3dis: Open-vocabulary 3d instance segmentation with 2d mask guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4018 4028, 2024. 2, 3, 4, 8, 9, 19, 20 Maxime Oquab, Timoth ee Darcet, Th eo Moutakanni, Huy Vo, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, et al. Dinov2: Learning robust visual features without supervision. ar Xiv preprint ar Xiv:2304.07193, 2023. 2 Songyou Peng, Kyle Genova, Chiyu Jiang, Andrea Tagliasacchi, Marc Pollefeys, Thomas Funkhouser, et al. Openscene: 3d scene understanding with open vocabularies. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 815 824, 2023. 3, 5 Charles R Qi, Or Litany, Kaiming He, and Leonidas J Guibas. Deep hough voting for 3d object detection in point clouds. In proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9277 9286, 2019. 17, 18 Charles Ruizhongtai Qi, Li Yi, Hao Su, and Leonidas J Guibas. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Neur IPS, 2017. 17 Jie Qin, Jie Wu, Pengxiang Yan, Ming Li, Ren Yuxi, Xuefeng Xiao, Yitong Wang, Rui Wang, Shilei Wen, Xin Pan, et al. Freeseg: Unified, universal and open-vocabulary image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 19446 19455, 2023. 3 Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. PMLR, 2021. 2, 3, 5, 6 Yongming Rao, Wenliang Zhao, Guangyi Chen, Yansong Tang, Zheng Zhu, Guan Huang, Jie Zhou, and Jiwen Lu. Denseclip: Language-guided dense prediction with context-aware prompting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18082 18091, 2022. 3 David Rozenberszki, Or Litany, and Angela Dai. Language-grounded indoor 3d semantic segmentation in the wild. In European Conference on Computer Vision, pp. 125 141. Springer, 2022. 2, 7, 8, 9, 11, 20 Jonas Schult, Francis Engelmann, Alexander Hermans, Or Litany, Siyu Tang, and Bastian Leibe. Mask3d for 3d semantic instance segmentation. ar Xiv preprint ar Xiv:2210.03105, 2022. 1, 3, 4, 8, 17, 20 Nur Muhammad Mahi Shafiullah, Chris Paxton, Lerrel Pinto, Soumith Chintala, and Arthur Szlam. Clip-fields: Weakly supervised semantic fields for robotic memory. ar Xiv preprint ar Xiv:2210.05663, 2022. 3 Published as a conference paper at ICLR 2025 Dhruv Shah, Bła zej Osi nski, Sergey Levine, et al. Lm-nav: Robotic navigation with large pretrained models of language, vision, and action. In Conference on Robot Learning, pp. 492 504. PMLR, 2023. 3 Julian Straub, Thomas Whelan, Lingni Ma, Yufan Chen, Erik Wijmans, Simon Green, Jakob J Engel, Raul Mur-Artal, Carl Ren, Shobhit Verma, et al. The replica dataset: A digital replica of indoor spaces. ar Xiv preprint ar Xiv:1906.05797, 2019. 2, 7, 8, 9, 11 Jiahao Sun, Chunmei Qing, Junpeng Tan, and Xiangmin Xu. Superpoint transformer for 3d scene instance segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, 2023. 3 Ayc a Takmaz, Elisabetta Fedele, Robert W Sumner, Marc Pollefeys, Federico Tombari, and Francis Engelmann. Openmask3d: Open-vocabulary 3d instance segmentation. ar Xiv preprint ar Xiv:2306.13631, 2023. 2, 3, 8, 9, 19, 20 Yoad Tewel, Yoav Shalev, Idan Schwartz, and Lior Wolf. Zerocap: Zero-shot image-to-text generation for visual-semantic arithmetic. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 17918 17928, 2022. 6 Thang Vu, Kookhoi Kim, Tung M Luu, Thanh Nguyen, and Chang D Yoo. Softgroup for 3d instance segmentation on point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2708 2717, 2022. 1, 3 Yizheng Wu, Min Shi, Shuaiyuan Du, Hao Lu, Zhiguo Cao, and Weicai Zhong. 3d instances as 1d kernels. In European Conference on Computer Vision, pp. 235 252. Springer, 2022. 3 Jiarui Xu, Shalini De Mello, Sifei Liu, Wonmin Byeon, Thomas Breuel, Jan Kautz, and Xiaolong Wang. Groupvit: Semantic segmentation emerges from text supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18134 18144, 2022. 3 Jiarui Xu, Sifei Liu, Arash Vahdat, Wonmin Byeon, Xiaolong Wang, and Shalini De Mello. Openvocabulary panoptic segmentation with text-to-image diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2955 2966, 2023. 3 Mi Yan, Jiazhao Zhang, Yan Zhu, and He Wang. Maskclustering: View consensus based mask graph clustering for open-vocabulary 3d instance segmentation. ar Xiv preprint ar Xiv:2401.07745, 2024. 2, 3, 4 Bo Yang, Jianan Wang, Ronald Clark, Qingyong Hu, Sen Wang, Andrew Markham, and Niki Trigoni. Learning object bounding boxes for 3d instance segmentation on point clouds. Advances in neural information processing systems, 32, 2019a. 3 Jihan Yang, Runyu Ding, Zhe Wang, and Xiaojuan Qi. Regionplc: Regional point-language contrastive learning for open-world 3d scene understanding. ar Xiv preprint ar Xiv:2304.00962, 2023. 2, 3 Zhengyuan Yang, Boqing Gong, Liwei Wang, Wenbing Huang, Dong Yu, and Jiebo Luo. A fast and accurate one-stage approach to visual grounding. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 4683 4693, 2019b. 17, 18 Li Yi, Wang Zhao, He Wang, Minhyuk Sung, and Leonidas J Guibas. Gspn: Generative shape proposal network for 3d instance segmentation in point cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3947 3956, 2019. 3 Yingda Yin, Yuzheng Liu, Yang Xiao, Daniel Cohen-Or, Jingwei Huang, and Baoquan Chen. Sai3d: Segment any instance in 3d scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3292 3302, 2024. 2, 4 Jiahui Yu, Zirui Wang, Vijay Vasudevan, Legg Yeung, Mojtaba Seyedhosseini, and Yonghui Wu. Coca: Contrastive captioners are image-text foundation models. ar Xiv preprint ar Xiv:2205.01917, 2022. 3 Published as a conference paper at ICLR 2025 Haobo Yuan, Xiangtai Li, Chong Zhou, Yining Li, Kai Chen, and Chen Change Loy. Openvocabulary sam: Segment and recognize twenty-thousand classes interactively. ar Xiv preprint ar Xiv:2401.02955, 2024. 3 Lu Yuan, Dongdong Chen, Yi-Ling Chen, Noel Codella, Xiyang Dai, Jianfeng Gao, Houdong Hu, Xuedong Huang, Boxin Li, Chunyuan Li, et al. Florence: A new foundation model for computer vision. ar Xiv preprint ar Xiv:2111.11432, 2021. 3 Nir Zabari and Yedid Hoshen. Semantic segmentation in-the-wild without seeing any segmentation examples. ar Xiv preprint ar Xiv:2112.03185, 2021. 3 C Zhang, H Wan, S Liu, X Shen, and Z Wu. Pvt: Point-voxel transformer for 3d deep learning. arxiv 2021. ar Xiv preprint ar Xiv:2108.06076, 2021. 3 Hengyuan Zhao, Hao Luo, Yuyang Zhao, Pichao Wang, Fan Wang, and Mike Zheng Shou. Revisit parameter-efficient transfer learning: A two-stage paradigm. ar Xiv preprint ar Xiv:2303.07910, 2023a. 5 Henry Hengyuan Zhao, Pichao Wang, Yuyang Zhao, Hao Luo, Fan Wang, and Mike Zheng Shou. Sct: A simple baseline for parameter-efficient fine-tuning via salient channels. International Journal of Computer Vision, pp. 1 19, 2023b. 5 Chong Zhou, Chen Change Loy, and Bo Dai. Extract free dense labels from clip. In European Conference on Computer Vision, pp. 696 712. Springer, 2022a. 3 Xingyi Zhou, Rohit Girdhar, Armand Joulin, Philipp Kr ahenb uhl, and Ishan Misra. Detecting twenty-thousand classes using image-level supervision. In European Conference on Computer Vision, pp. 350 368. Springer, 2022b. 2, 4 Ziqin Zhou, Yinjie Lei, Bowen Zhang, Lingqiao Liu, and Yifan Liu. Zegclip: Towards adapting clip for zero-shot semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11175 11185, 2023. 7 Xueyan Zou, Zi-Yi Dou, Jianwei Yang, Zhe Gan, Linjie Li, Chunyuan Li, Xiyang Dai, Harkirat Behl, Jianfeng Wang, Lu Yuan, et al. Generalized decoding for pixel, image, and language. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15116 15127, 2023. 3 Xueyan Zou, Jianwei Yang, Hao Zhang, Feng Li, Linjie Li, Jianfeng Wang, Lijuan Wang, Jianfeng Gao, and Yong Jae Lee. Segment everything everywhere all at once. Advances in Neural Information Processing Systems, 36, 2024. 3 Published as a conference paper at ICLR 2025 Segment Any 3D Object with language supplementary material In this supplementary material, we provide more implementation details, introduce additional evaluation on free-form language instructions, and conduct more qualitative and quantitative analysis. More implementation details are provided in Sec. A. SOLE is evaluated on 3D visual grounding task to verify the responsive ability to free-form language instructions in Sec. B Analysis of CLIP visual features are provided in Sec. C. Variants of SOLE are introduced in terms of scalability and compatibility in Sec. D. Analysis of semantic-aware mask generator are provided in Sec. E. More qualitative results about the mask caption and segmentation results are shown in Sec. F. A IMPLEMENTATION DETAILS Segmentation Network. Following Mask3D (Schult et al., 2022), we use the transformer-based mask-prediction paradigm to obtain instance mask and semantic features. The masks are initialized from object queries and regressed by attention layers. For each 3D point cloud scene, we use farthest point sampling (Qi et al., 2017) to get 150 points as object queries. After getting masks from the segmentation model, we use DBSCAN (Ester et al., 1996) to break down non-contiguous masks into smaller, spatially contiguous clusters to improve the mask quality. The maximum distance and neighborhood points number are set to 0.95 and 1, respectively. Text Information Generation and Extraction. To effectively generate a caption for each mask, we use a caption model in CLIP space, i.e., De Cap (Li et al., 2023). De Cap is a lightweight transformer model to generate captions from CLIP image embedding. It contains a 4-layer Transformer with 4 attention heads as the language model and the visual embedding is obtained from the pre-trained Vi T-L/32 CLIP model. We feed the mask features that are average pooled from the projected CLIP visual features into the De Cap model to obtain the mask caption. Then the caption is integrated into the text prompt a {} in a scene. to better align with our data, e.g. a blue chair in a scene. . With the mask caption, noun phrases are extracted by the NLP library, Text Blob (Loria et al., 2018) and spa Cy (Honnibal & Montani, 2017), to get the mask-entity association. B 3D VISUAL GROUNDING To further verify the effectiveness of SOLE on various language instructions, we conduct experiments on 3D visual grounding benchmark Scan Refer (Chen et al., 2020). 3D visual grounding aims at localizing 3D objects with free-form text descriptions. Therefore, we query SOLE with each text prompt in the Scan Refer validation set to get the corresponding instance, and then 3D bounding box is obtained from the instance masks. The performance is evaluated on the matching accuracy with Io U over 0.25 (ACC@25) and 0.5 (ACC@50). Baselines. We compare SOLE with five specialist baseline models and one generalist model. Specialist models mean that the models are designed and trained for 3D visual grounding only while generalist models are models that can address other tasks such as instance segmentation with class names. For the specialist models, OCRand (Chen et al., 2020) uses an oracle with ground truth bounding boxes of objects, and selects a random box that matches the object category. Vote Rand (Qi et al., 2019; Chen et al., 2020) leverages the pre-trained Vote Net (Qi et al., 2019) to predict bounding boxes and randomly select a box of the correct semantic class. SCRC (Hu et al., 2016) and One-stage (Yang et al., 2019b) are 2D approaches with 3D extension using back-projection. Scan Refer (Chen et al., 2020) uses a pre-trained Vote Net (Qi et al., 2019) with a trained GRU to select a matching bounding box. For the generalist model, 3D-LLM directly predicts the location of the bounding box corresponding to the text description via large language model. Note that except for Published as a conference paper at ICLR 2025 Table 7: Results on Scan Refer (Chen et al., 2020) for 3D visual grounding task. SOLE achieves the best performance on generalist models with weak supervision. Method Type Supervision ACC@25 ACC@50 OCRand (Chen et al., 2020) Full 30.0 29.8 Vote Rand (Qi et al., 2019; Chen et al., 2020) Full 10.0 5.3 SCRC (Hu et al., 2016) Full 18.7 6.5 One-stage (Yang et al., 2019b) Full 20.4 9.0 Scan Refer (Chen et al., 2020) Full 41.2 27.4 3D-LLM (flamingo) (Hong et al., 2023) Generalist Full 21.2 SOLE Zero-Shot 25.2 22.6 (a) K-Means clustering of f p (b) K-Means clustering of f b (c) K-Means clustering of f b Figure 6: K-means clustering of different backbone features. Different colors denote different clusters. OCRand and Vote Rand where training is not required, the other four baseline models are trained or fine-tuned on the Scan Refer training set. Differently, SOLE is only trained with instance segmentation masks of Scan Netv2 (Dai et al., 2017), without leveraging any labels in Scan Refer. Results. As shown in Tab. 7, with instance mask annotation, SOLE outperforms the generalist model 3D-LLM by 4% on ACC@25. In addition, SOLE can achieve competitive performance with the fully-supervised specialist counterpart on ACC@50 (22.6% v.s. 27.4%). Such results demonstrate the strong generalization ability and the effectiveness in responding to free-form language instructions of our framework. C ANALYSIS OF CLIP VISUAL FEATURE CLIP visual features play an important role in the generalization ability of SOLE. In this section, we further analyze the effectiveness of CLIP visual features in the backbone feature ensemble and inference ensemble. C.1 BACKBONE FEATURE ENSEMBLE As shown in Tab. 5, solely using 3D instance backbone feature f b or projected CLIP visual feature f p cannot achieve the best performance. 3D backbone features lack the generalized semantic information while projected CLIP visual features lack the location and geometry information. To further verify the effectiveness of the backbone feature ensemble, we visualize the clustering results of different features in Fig. 6. In Fig. 6a, all chairs are clustered together (the green cluster), showing that the projected CLIP features contain good semantic information but cannot detect the instances. In Fig. 6b, different instances within one category can be identified, e.g., chairs are in three clusters. However, the semantic generalization ability is degraded. As in the highlighted red circle, the trash can is detected by projected CLIP visual features (Fig. 6a) but misclassified to a chair cluster when only using 3D backbone (Fig. 6b). Compared with using the two features separately, SOLE combines the two features and thus achieves better semantic generalization ability (segmentation of the trash can) and segmentation performance (chairs are clustered into six clusters). The visualization results further demonstrate the effectiveness of the backbone feature ensemble. Published as a conference paper at ICLR 2025 Table 8: Analysis on classification probability ensemble. Results are reported on the Scan Netv2 (Dai et al., 2017) dataset in 2cm voxel size. Component AP AP50 AP25 w.o. Ensemble 42.2 58.6 66.9 hard geometric mean 43.7 61.1 70.1 soft geometric mean (ours) 44.4 62.2 71.4 Table 9: Analysis on τ. Results are reported on the Scan Netv2 (Dai et al., 2017) dataset in 2cm voxel size. 0.1 44.1 0.334 44.2 0.667 (ours) 44.4 0.9 42.4 Table 10: Analysis of the light version of SOLE on Scan Netv2 (Dai et al., 2017). Despite of performance drop, SOLE-light still outperforms mask-training methods with the fastest inference time. The best results are in bold while the second best results are underscored. Method AP AP50 AP25 Inference Time (s) voxel size Open Ins3D (Huang et al., 2023b) - 28.7 38.9 16.3 2cm Open Mask3D (Takmaz et al., 2023) 31.0 39.5 44.0 553.7 2cm Open3DIS (only 3D) (Nguyen et al., 2024) 31.3 42.4 47.8 56.6 2cm SOLE (ours) 44.4 62.2 71.4 454.2 2cm SOLE-light 37.1 50.8 59.0 13.1 2cm C.2 INFERENCE ENSEMBLE During inference, we combine the CLIP visual features with the predicted mask feature to achieve better generalization ability. Specifically, after obtaining the 3D masks, per-point CLIP features are pooled within the mask. The pooled CLIP feature and mask feature are then fed into the classifier to obtain the respective classification probability p(f m) and p(f p), and the final probability is yielded by the ensemble of them. In Tab. 8, we compare three options to ensemble the class probabilities of 3D segmentation model and CLIP model. w.o. Ensemble denotes only using the prediction of 3D segmentation model, and hard geometric mean refers to the standard geometric mean, formulated as p(f m)τ p(f p)1 τ. Our method, soft geometric mean (Eq. 7), shows the best results among the ensemble methods, demonstrating the effectiveness of dynamically fusing the prediction from two models. However, as shown in the first row, SOLE already achieves competitive performance even without utilizing the CLIP prediction, further demonstrating the strong generalization ability of our multimodal fusion network. We also analyze the parameter τ on soft geometric mean in Tab. 9. SOLE is robust to τ while achieving the best result when it is 0.667. As shown in Fig. 2, SOLE focuses more on the 3D semantic-aware mask generation than the semantic classification. Therefore, SOLE itself can be used for fast segmentation and it can also be combined with generalizable semantic classification and 2D mask refinement techniques proposed by previous and concurrent works (Takmaz et al., 2023; Nguyen et al., 2024). By default, we use the soft geometric mean to combine the CLIP visual feature with predicted mask feature for better generalization ability. In this section, we investigate two variants of SOLE: the light version without using any additional information for segmentation and the advanced version leveraging the semantic classification techniques from other methods. D.1 SOLE-LIGHT SOLE-light introduces a light inference technique to exclude 2D foundation model (CLIP) used in backbone feature ensemble and soft geometric mean. In Tab. 10, we compare performance and inference time of SOLE-light to previous mask training methods (Takmaz et al., 2023; Huang et al., 2023b; Nguyen et al., 2024) on Scan Netv2 (Dai et al., 2017) dataset. Removing the CLIP model significantly improve the inference time by 30 times (454s v.s. 13s) with the performance drop to some degree. However, SOLE-light still outperforms previous mask training methods by a large margin, further demonstrating the strong generalization ability and good mask quality of our semantic-aware mask generator. Published as a conference paper at ICLR 2025 Table 11: Analysis of compatibility with previous works on Scan Net200 (Rozenberszki et al., 2022). SOLE shows high compatibility with previous mask training methods, achieving state-ofthe-art performance when it is incorporated into Open3DIS framework. Method 2D masks Classification AP AP50 AP25 Open Mask3D (Takmaz et al., 2023) Open Mask3D 15.4 19.9 23.1 Open3DIS (Nguyen et al., 2024) Open3DIS 23.7 29.4 32.8 SOLE (ours) SOLE 20.1 28.1 33.6 SOLE + Open Mask3D SOLE + Open Mask3D 20.2 28.5 33.2 SOLE + Open3DIS SOLE + Open3DIS 24.5 30.1 34.1 Table 12: Mask quality evaluation on Scan Net200 (Rozenberszki et al., 2022). SOLE significantly outperforms Mask3D-CA in terms of mask quality and the performance is even comparable to original Mask3D despite the absence of filtering step. Method Semantic Guidance filtering AP AP50 AP25 Mask3D-CA None 23.1 29.9 33.4 SOLE (ours) MMA 43.5 56.2 62.5 Mask3D (Schult et al., 2022) GT labels 53.0 72.5 82.4 SOLE (ours) MMA 43.6 59.3 70.2 D.2 SOLE-ADVANCED Previous works (Huang et al., 2023b; Takmaz et al., 2023; Nguyen et al., 2024) commonly use the 3D masks generated by a pre-trained class-agnostic mask generator and propose techniques to refine and classify the masks with the help 2D vision large model (VLM). Since SOLE focuses more on the semantic-aware 3D mask prediction, we introduce several variants of SOLE in Tab. 11 by incorporating the previous techniques into our framework and evaluate on Scan Net200 (Rozenberszki et al., 2022). Specifically, SOLE + Open Mask3D further leverages the frame selection and mask feature aggregation in Open Mask3D to improve the masks classification on top of SOLE. Similarly, SOLE + Open3DIS uses 2D mask guidance and pointwise feature extraction in Open3DIS to improve the masks and classification. As shown in Tab. 11, our model exhibits the high compatibility with previous work and achieves the state-of-the-art performance (24.5%AP) on Scan Net200 when it is combined with Open3DIS framework. E SEMANTIC-AWARE MASK GENERATOR One of the main contributions of SOLE is that we improve the mask generator with multimodal associations and multimodal fusion network. In this section, we evaluate and analyze how semantic information improves the mask quality. E.1 QUALITY OF SEMANTIC-AWARE MASKS In contrast to previous works (Huang et al., 2023b; Takmaz et al., 2023; Nguyen et al., 2024) that disregard semantic information in the mask generation, our SOLE generates semantic-aware masks with showing superior quality and generalizability. To investigate the impact of semantic information to mask quality, Tab. 12 further examines the mask evaluation in the class-agnostic setting. We first adjust Mask3D (Schult et al., 2022) to remove semantic-related parts by discarding classification head and training it on only mask annotations. This Mask3D variant, denoted as Mask3D-CA, is compared to SOLE in class agnostic 3DIS on Scan Net200 (Rozenberszki et al., 2022). During the evaluation, we use all of the mask proposals without filtering the low quality masks. As shown in Tab. 12, SOLE significantly outperforms Mask3D-CA on the mask quality, demonstrating the effectiveness of semantic information in mask generation. We also show the performance of fullsupervised counterpart in Tab. 12, taking the class information into account but only evaluating the mask quality. The fully-supervised Mask3D (Schult et al., 2022) leverages classification logits to filter low-quality segments, which is their way to use semantic information to support mask generation. SOLE shows close performance to original Mask3D despite not filtering the low-quality Published as a conference paper at ICLR 2025 Table 13: Analysis on caption influence for multimodal associations. Scene Verse generates captions based on ground truth class labels, which can further improves the performance of SOLE. Method Caption GT Labels AP AP50 AP25 SOLE De Cap 20.1 28.1 33.6 SOLE Scene Verse 26.1 37.3 44.2 masks. When filtering the low-quality segments, the performance of SOLE can be further improved. Comparing the second and fourth row in Tab. 12, our SOLE shows robust performance regardless of whether filtering step is present or not. Since SOLE is aware of rich semantic information, it can already generate high quality of masks without the requirements of post processing. E.2 QUALITY OF SEMANTIC GUIDANCE SOLE leverages De Cap to generate mask captions without any class labels. However, although the generated captions contain rich semantic information, they are still not perfect due to the lack of class label. In Tab. 13, we replace the generated caption with captions in Scene Verse (Jia et al., 2024). Scene Verse (Jia et al., 2024) leverages LLM to generate mask captions given the mask and class labels. With the better captions, SOLE can be further improved by 6% in AP on Scan Net200 dataset. These results demonstrate that better semantic information indeed help to learn a better mask proposal modules and the label-free caption generator can still be improved, deserving investigation in the future work. F QUALITATIVE RESULTS Visualization for Segmentation Results. In Fig. 1 and Fig. 7, we present qualitative results, demonstrating that SOLE is capable of processing free-form language queries, including but not limited to visual questions, attributes description, and functional description. More qualitative results are presented by the video in supplementary material. Visualization for Mask Captions. We provide the visualization for different masks and corresponding generated captions in Fig. 8. The generated caption contains the semantic information of the 3D object as well as the location for better 3D segmentation. As shown in the examples, more than one nouns exist in the caption and thus we aggregate all the noun phrases with attention mechanism in mask-entity association. Published as a conference paper at ICLR 2025 leather texture. sofa with flower pattern. Green tree. What time is it? Turn on the light. I want to watch movie. Chairs near by the window. vending machine I m hungry. Throwing away the garbage. I wanna see outside. Place I can pee. I m sleepy. shower curtain I want to take a shower. sofa chairs Where can I sit? refrigerator stove stove hood Cooking the food. Figure 7: Qualitative results from SOLE. Our SOLE demonstrates open-vocabulary capability by effectively responding to free-form language queries, including visual questions, attributes description and functional description. Published as a conference paper at ICLR 2025 a microwave oven sitting inside of a small kitchen a trash can with a microwave on top of it a black refrigerator freezer sitting inside of a kitchen a bed with white pillows and a large white comforter on top a black suitcase with a piece of luggage sitting on a floor a brown chair sitting in front of a desk with a black table a box of various items sitting on a cluttered place a painting hanging on the wall of a living room a kitchen area with a sink and counter top near cabinets a computer monitor sitting on a desk in front of two computers a desk with a laptop, computer, papers and other items a plant with a white pot sitting on a table Figure 8: Qualitative examples for mask captions. Generated captions contain the semantic, appearance, and geometric relationship of the corresponding object.