# unionoverintersections_object_detection_beyond_winnertakesall__dcd8da7c.pdf Published as a conference paper at ICLR 2025 UNION-OVER-INTERSECTIONS: OBJECT DETECTION BEYOND WINNER-TAKES-ALL Aritra Bhowmik Pascal Mettes Martin R. Oswald Cess G. M. Snoek Atlas Lab, University of Amsterdam {a.bhowmik, p.s.m.mettes, m.r.oswald, c.g.m.snoek}@uva.nl This paper revisits the problem of predicting box locations in object detection architectures. Typically, each box proposal or box query aims to directly maximize the intersection-over-union score with the ground truth, followed by a winner-takesall non-maximum suppression where only the highest scoring box in each region is retained. We observe that both steps are sub-optimal: the first involves regressing proposals to the entire ground truth, which is a difficult task even with large receptive fields, and the second neglects valuable information from boxes other than the top candidate. Instead of regressing proposals to the whole ground truth, we propose a simpler approach: regress only to the area of intersection between the proposal and the ground truth. This avoids the need for proposals to extrapolate beyond their visual scope, improving localization accuracy. Rather than adopting a winner-takes-all strategy, we take the union over the regressed intersections of all boxes in a region to generate the final box outputs. Our plug-and-play method integrates seamlessly into proposal-based, grid-based, and query-based detection architectures with minimal modifications, consistently improving object localization and instance segmentation. We demonstrate its broad applicability and versatility across various detection and segmentation tasks. 1 INTRODUCTION Object detection is a long-standing challenge in computer vision, with the goal of spatially localizing and classifying objects in images by their bounding box (Burl et al., 1998; Viola & Jones, 2004; Felzenszwalb et al., 2010). Over the past decade, significant progress has been made, driven by advances in various stages of the detection pipeline. From the foundational R-CNN (Girshick et al., 2014) to Faster R-CNN (Ren et al., 2015), and paradigm-shifting YOLO architectures (Redmon et al., 2016; Ge et al., 2021; Bochkovskiy et al., 2020; Redmon & Farhadi, 2018; Huang et al., 2018), to the recent innovations by transformer networks (Zhu et al., 2021; Carion et al., 2020; Dai et al., 2021a; Nguyen et al., 2022), these developments have enhanced feature extraction and detection. Advances in loss functions, such as focal loss (Lin et al., 2017) and IOU-aware loss (Wu et al., 2020a; Zhang et al., 2021), have further addressed class imbalance and precise localization. Regardless of backbone or optimization choices, most modern methods generate a set of proposal boxes, grid-boxes, or queries. These are classified and spatially regressed, followed by ranking or non-maximum suppression to retain only the highest scoring box per location. In essence, each proposal solves the intersection-over-union alignment with the ground truth, followed by a winner-takes-all approach. We revisit this fundamental problem of object localization and candidate selection. We identify two issues with the common setup of object detection and propose a simple solution. First, the goal of proposal-based (Ren et al., 2015; He et al., 2017), grid-based (Redmon & Farhadi, 2018; Bochkovskiy et al., 2020) and query-based (Carion et al., 2020; Zhu et al., 2021) detectors, is to learn proposal boxes that independently represent ground truth objects. As shown in Figure 1, this often becomes an ill-posed problem. During the forward pass, proposal boxes typically capture only a portion of the ground truth, leading to an extrapolation challenge when regressing for perfect alignment. Second, there are multiple proposals or queries for each ground truth object, offering complementary views. While existing methods use box voting to select the top candidate, we focus on merging information from multiple proposals, leveraging their complementary strengths for improved detection. Published as a conference paper at ICLR 2025 Feature Encoder Regress Outside Scope Regress Inside Scope Ground Truth Regression Stage Winner-takes-all Combine Boxes Post-Processing Stage Competition Collaboration Traditional Visual Scope Input Image Union-over-Intersections Final Prediction Figure 1: Union-over-Intersections vs. Winner-takes-all. We introduce two simple modifications to the traditional object detection pipeline. First, in the regression stage, rather than requiring proposals to align with the entire ground truth, we adjust the targets to focus solely on their intersection with the ground truth. Second, in the post-processing stage, we perform Union-over-Intersections over the traditional practice of discarding less optimal bounding boxes. Our approach underscores the advantage of cooperative interaction among proposals, demonstrating that collaboration yields superior results over competitive exclusion. Our solution is straightforward: we decompose the problems of regressing to the entire ground truth from a single proposal and winner-takes-all candidate selection into easier to solve intersection and union problems. We do so by altering two steps of the object detection pipeline: (i) We set a new target for the regression of proposals: rather than predicting the entire ground truth, we only predict the region of intersection. Thus, we only regress towards the ground truth within the visual scope of the proposal. (ii) Given a group of proposals with predicted ground truth intersections, we form the final prediction by taking the union over the intersection regions. i.e., instead of selecting only the most confident proposal in a region, we use the wisdom of the crowd to form our final predictions. The two stages barely affect the existing object detection pipelines. In the current regression heads, we only need to change the target coordinates. In the current non-maximum suppression, we use the exact same way of grouping proposals, but instead of a winner-takes-all, we take the union-overintersections (Uo I) of all proposals in a group as the final prediction. Our approach can therefore be plugged into proposal-based, grid-based, and query-based object detectors, be it a convolutional or transformer architecture. Despite its technical simplicity, Uo I directly improves the detection performance. We show how our revisited approach improves canonical detection and instance segmentation methods across multiple datasets. We also demonstrate through a controlled experiment that our approach is ready to benefit from future improvements in classification and regression performance, and hope to establish it as the preferred formulation for the next generation of object detectors. 2 RELATED WORKS Object detection architectures. Object detection architectures can be broadly categorized into: single-stage, two-stage, and transformer-based detectors, each with their own characteristics and merits. Single-stage detectors such as SSD (Liu et al., 2016), the YOLO (Redmon et al., 2016) series (including YOLOv3 (Redmon & Farhadi, 2018), YOLOv4 (Bochkovskiy et al., 2020), YOLOX (Ge et al., 2021), YOLO-Lite (Huang et al., 2018)), Retina Net (Lin et al., 2017), Efficient Det (Tan et al., 2020), Corner Net (Law & Deng, 2018), Center Net (Duan et al., 2019), and FCOS (Tian et al., 2019) prioritize speed, directly predicting object classes and bounding boxes in one pass. Retina Net introduces Focal Loss for class imbalance and small object detection, while Efficient Det leverages Efficient Net (Tan & Le, 2019) backbones for scalable efficiency. Innovations in Corner Net, Center Net, and FCOS focus on anchor-free methods, further enhancing real-time efficiency. In the realm of two-stage object detectors, the R-CNN family has played a pivotal role. This family includes R-CNN (Girshick et al., 2014), Fast R-CNN (Girshick, 2015), and Faster R-CNN (Ren et al., 2015), which introduced end-to-end training with Region Proposal Networks. Additionally, Mask R-CNN (He et al., 2017) added segmentation capabilities, Cascade R-CNN (Cai & Vasconcelos, 2018) improved accuracy via a multi-stage framework, and Libra R-CNN (Pang et al., 2019) Published as a conference paper at ICLR 2025 addressed training imbalances. Enhancements include Trident Net (Paz et al., 2021), which handles scale variation with parallel branches, and methods like Grid R-CNN (Lu et al., 2019) for precise localization and Double-Head R-CNN (Wu et al., 2020b), which differentiates between the classification and bounding box regression, further enriching the field. Transformer-based models like DETR (Carion et al., 2020), Deformable DETR (Zhu et al., 2021), TSP-FCOS (Sun et al., 2021), Swin Transformer (Liu et al., 2021), and Vi T-FRCNN (Beal et al., 2020) have revolutionized object detection. DETR views detection as set prediction, Deformable DETR adds deformable attention for focused image analysis, TSP-FCOS integrates transformers with FCOS (Tian et al., 2019) for scale variance, Swin Transformer employs a hierarchical approach, and Vi T-FRCNN combines Vision Transformer (Dosovitskiy et al., 2021) with Faster R-CNN (Ren et al., 2015) for feature extraction. Follow-up work include Boxer (Nguyen et al., 2022), DINO (Zhang et al., 2022), DAB-DETR (Liu et al., 2022), UP-DETR (Dai et al., 2021b), Dynamic DETR (Dai et al., 2021a), and Conditional DETR (Meng et al., 2021), enhancing various aspects like training efficiency and dynamic predictions. Despite the diversity in approaches, a common aspect across these three paradigms is the presence of a classification and bounding box regression stage, in which each proposal needs to individually align with the entire ground truth object. We show that proposal-based, grid-based, and query-based methods benefit from our approach of first regressing to intersections, followed by a union-overintersection grouping. Object detection post-processing. After the regression stage, selection of the best regressed proposal from a set of candidate detections ensures accuracy, with Non-Maximum Suppression (Girshick, 2015; Ren et al., 2015; Redmon et al., 2016; Liu et al., 2016) (NMS) being a key technique for selecting the best proposals by eliminating redundant detections. Traditional NMS, while effective, has been surpassed by variants such as Soft-NMS (Bodla et al., 2017), which adjusts confidence scores instead of discarding detections, improving performance in crowded scenes. Learning NMS (Hosang et al., 2017) introduces adaptability by incorporating suppression criteria into the network training. Further refinements like Io U-aware NMS (Wu et al., 2020a; Zhang et al., 2021) and Distance-Io U (DIo U) NMS (Zheng et al., 2020) consider both overlap and spatial relationships between boxes for more precise detection. Additional post-processing techniques include Matrix NMS (Wang et al., 2020), which employs a matrix operation for efficient suppression, and Adaptive NMS (Liu et al., 2019), which dynamically adjusts thresholds based on object densities. Furthermore, ASAPNMS (Tripathi et al., 2020) optimizes processing speed, while Fitness NMS (Tychsen-Smith & Petersson, 2018) considers both detection scores and spatial fitness for more accurate suppression to enrich post-processing in object detection. Bounding box regression losses, crucial for refining predicted bounding boxes, have evolved to better align predictions with ground truth boxes. Io U loss (Yu et al., 2016; Jiang et al., 2018) directly optimizes the overlap between predicted and ground truth boxes, while GIo U loss (Rezatofighi et al., 2019) addresses scenarios where predicted boxes fall outside the ground truth, penalizing such deviations. DIo U (Zheng et al., 2020) and α-Io U (He et al., 2021) losses further enhance regression quality by incorporating factors like the distance between box centers and aspect ratio consistency. These advanced losses have become benchmarks for evaluating the quality of bounding box regressions and are integral to improving detection accuracy in modern object detection pipelines. The improved post-processing techniques address the limitation of selecting a single winner from each cluster of proposals at a location. However, all techniques still assume that each individual box is already fully aligned with the corresponding ground truth box and serve primarily to remove duplicate detections. Instead, we seek to obtain wisdom from the crowd by using all proposals for the same object instance to form the final object detection output. We do so by making minimal adjustments to the non-maximum suppression pipeline. 3 UNION-OVER-INTERSECTIONS IN OBJECT DETECTION Our Union-over-Intersections (Uo I) approach improves the object detection pipeline with minimal adjustments, without adding new methods or architectures. We first outline the standard pipeline, focusing on bounding box regression and top candidate selection, followed by two simple modifications, as shown in Figure 2. Problem statement. In traditional object detection, bounding box regression learns a mapping function f that transforms a proposal box P=(Px1, Py1, Px2, Py2) to approximate a ground truth box G=(Gx1, Gy1, Gx2, Gy2). This is typically achieved by minimizing the L1 or L2 loss between the Published as a conference paper at ICLR 2025 Set of Proposals P = p1, , p N Corresponding ground truths G = g1, , g N Corresponning detection scores S = s1, , s N Iou threshold k Loss and Targets for Regression : Loss = 𝑇 𝐵, where B = b1, , b N are the set of box predictions. T = g1, , g N Traditional targets T = p1 g1, , p N g N Our targets Main Difference: We regress only to the intersections, not the whole ground truth. begin D while P empty do m argmax S M bm ; H Pm ; C 𝑯 B B \ M ; P P \ H for i in range (len(P)) do if iou(M, bi) k then B B \ bi; S S \ si end if iou(H, pi) k then P P \ pi ; S S \ si ; C C bi M = top 5 (C) combined end end D D M end return D, S end Regression Stage Post-Processing Stage Main Difference: Union over discarding. Figure 2: Pseudo code demonstrating our minimal changes in the object detection pipeline. During regression, we adjust the target of the proposals from the entire ground truth to only the intersection with ground truth. In post-processing, we group boxes by proposal rather than regressed outcomes and merge regressed intersections, avoiding the discard of non-maximum boxes. predicted box and the ground truth. However, this approach presents a challenge: proposals P often capture only part of the ground truth G, requiring f to extend P beyond its visible scope. As f must infer parts of G not visible within P, this becomes a difficult task. While large receptive fields help, refining proposals through a union-based approach yields better results. Additionally, the set of proposals for a ground truth object, P={P1, P2, . . . , Pn}, provides complementary views, but non-maximum suppression or ranking selects only one candidate, discarding others. This assumes the intersection-over-union problem is already solved, merely excluding duplicates in the same region. To address these issues, we redefine box regression as an intersection learning task, aligning proposals with the intersecting parts of the ground truth boxes. Instead of discarding non-top proposals, we replace non-maximum suppression with a union operator, defining the final object as union-overintersection of all proposals in a region. Figure 1 illustrates how these tasks lead to final object detection. Intersection-based regression. We redefine the regression task as an intersection learning problem. Instead of regressing to the entire ground truth, each proposal regresses only to the visible part of the ground truth, i.e., the intersection between the proposal and ground truth. Let I=(Ix1, Iy1, Ix2, Iy2) be the intersection of boxes P and G. The task is to learn a mapping function f such that f (P) I. The intersection is computed as: Ix1 = max(Px1, Gx1) , Iy1 = max(Py1, Gy1) , (1) Ix2 = min(Px2, Gx2) , Iy2 = min(Py2, Gy2) . (2) The loss for this task is: Lintersection = X i {x1,y1,x2,y2} f (Pi) Ii t , (3) where t can be 1 or 2 depending on whether the L1 or L2 loss is used. Intersection-based grouping. Let P={P1, P2, . . . , Pn} be the set of all proposals for a ground truth object. The corresponding intersections with the ground truth are I={I1, I2, . . . , In}. Our goal is to combine these intersections into a bounding box using a function c : I R4, minimizing the Published as a conference paper at ICLR 2025 difference between the combined box and the ground truth: i {x1,y1,x2,y2} |ci(I) Gi| . (4) During inference, we refine the union of intersections to ensure a tight fit. Unlike traditional NMS, which groups regressed boxes, our method groups proposals, using the highest scoring proposal as a seed. This approach leverages the original proposals, as intersections often lack sufficient Io U for effective grouping, increasing the likelihood of fully encompassing the ground truth. Instead of discarding non-maxima, we take the minimum [xi 1, yi 1] and maximum [xi 2, yi 2] of all intersection regressions in the group. This forms a set of combined boxes B={B1, B2, . . . , Bm}. Each combined box Bj is a candidate bounding box for the object represented by group gj. By design, our method yields the same number of predictions as current detectors and is compatible with any NMS variant. Finally, we perform a regression step to refine the union-over-intersection box predictions. The regression function r : B R maps each combined box to a final regressed box, denoted as R, which represents the refined predictions: Lrefinement = i {x,y,w,h} |r(Bji) Tji| , (5) where Bji is the i-th coordinate of the combined box Bj, and Tji is the corresponding ground truth coordinate. The overall loss function is a combination of the intersection loss and refinement loss: Ltotal = Lintersection + λ Lrefinement , where λ, set to 0.5 in our experiments, controls the weight of the refinement loss. 4 EXPERIMENTS 4.1 DATASETS, EVALUATION, AND BASELINE DETECTOR IMPLEMENTATIONS Datasets. We conduct evaluations on COCO (Lin et al., 2014), covering two tasks: object detection and instance segmentation. The MS-COCO 2017 dataset is a key benchmark for object detection and instance segmentation, comprising 80 categories with 118k training and 5k evaluation images. It features a maximum of 93 object instances per image, with an average of 7 objects. Further results on PASCAL VOC (Everingham et al., 2010) are provided in the supplemental. For Pascal VOC, we leverage the 2007 and 2012 editions. It spans 20 object categories, with 5,011 training and 4,952 testing images in VOC2007, and an additional 11,540 training images in VOC2012. Evaluation criteria. We adhere to the standard evaluation protocols. For MS-COCO (Lin et al., 2014), we report the mean Average Precision (m AP) across Io U thresholds from 0.5 to 0.95 m AP, at specific Io U thresholds (0.5 and 0.75), multiple scales (small, medium and large), and Average Recall (AR) metrics. For PASCAL VOC (Everingham et al., 2010), we report the m AP metric at Io U thresholds of 0.5, 0.6, 0.7, 0.8 and 0.9. For instance segmentation, we report our results on MS-COCO with the m AP across Io U thresholds from 0.5 to 0.95 and across multiple scales (small, medium, large). Baseline detector implementations. We evaluate our Union-over-Intersections (Uo I) approach across five diverse object detection architectures: Faster R-CNN (Ren et al., 2015), Mask RCNN (He et al., 2017), Cascade R-CNN (Cai & Vasconcelos, 2018), YOLOv3 (Redmon & Farhadi, 2018), and Deformable DETR (Zhu et al., 2021), representing single-stage, two-stage, and transformer-based detectors. This comprehensive selection highlights the versatility of our method across different detection paradigms. We trained Faster R-CNN, Mask R-CNN, and Cascade R-CNN on COCO using standard configurations such as random horizontal flip, 512 proposals, and SGD optimization over 12 epochs, with a learning rate decay at epochs 8 and 11. YOLOv3 is trained with the Darknet-53 backbone for 273 epochs using common augmentation strategies and SGD. Deformable DETR is trained for 50 epochs with the Adam W optimizer. Published as a conference paper at ICLR 2025 Table 1: Detection comparison on MS-COCO. No matter the object detection method and backbone tested, when we replace the traditional box proposal selection process with our Union-over Intersections (Uo I), it results in improved accuracy for all metrics. Method Backbone m AP AP50 AP75 APS APM APL Faster R-CNN (Ren et al., 2015) R-50-fpn 37.4 58.1 40.4 21.2 41.0 48.1 Faster R-CNN w/ Uo I R-50-fpn 38.1 58.7 40.9 21.7 41.8 49.5 Faster R-CNN (Ren et al., 2015) R-101-fpn 39.4 60.1 43.1 22.4 43.7 51.1 Faster R-CNN w/ Uo I R-101-fpn 40.3 60.8 43.5 23.1 44.5 52.8 Mask R-CNN (He et al., 2017) R-50-fpn 38.2 58.8 41.4 21.9 40.9 49.5 Mask R-CNN w/ Uo I R-50-fpn 38.8 59.6 41.8 22.2 41.6 50.9 Mask R-CNN (He et al., 2017) R-101-fpn 39.8 60.3 43.4 23.1 43.8 52.5 Mask R-CNN w/ Uo I R-101-fpn 40.9 61.1 44.0 23.5 44.7 54.6 Cascade R-CNN (Cai & Vasconcelos, 2018) R-101-fpn 42.5 60.7 46.4 23.5 46.5 56.4 Cascade R-CNN w/ Uo I R-101-fpn 43.1 61.9 46.8 24.0 47.2 57.3 YOLOv3 (Redmon & Farhadi, 2018) Dark Net-53 33.7 56.6 35.3 19.4 36.8 44.3 YOLOv3 w/ Uo I Dark Net-53 34.5 57.5 35.9 19.9 37.3 45.2 Def-DETR (Zhu et al., 2021) R-50-fpn 44.3 63.2 48.6 26.8 47.7 58.8 Def-DETR w/ Uo I R-50-fpn 44.8 63.9 49.1 27.2 48.3 59.8 Inference cost. Our method adds minimal overhead. On Faster R-CNN, we achieve 14.1 fps (compared to 14.4 fps baseline) due to the extra regression stage. In Deformable DETR, switching to part-based regression leads to slightly longer training times: 23h for 50 epochs vs. 21h for baseline. 4.2 BENEFIT OF UNION-OVER-INTERSECTIONS To showcase the effectiveness of Union-over-Intersections in object detection pipelines, we perform a series of one-to-one comparisons. For Faster R-CNN, Mask R-CNN and Cascade R-CNN, we modify the regression targets to focus on object parts. For YOLO, we modify its object-to-grid assignment strategy by using an Io U-based criterion instead of the traditional center-based approach, assigning objects to multiple grid cells if they overlap sufficiently. Each anchor is tasked with regressing the part of the object it overlaps with best. In Deformable DETR, each query predicts a bounding box through the regression head after the decoder. Instead of matching these boxes to full ground truth boxes, we divide the ground truth into quadrants, assign queries to specific parts for part-based regression. Finally we apply our grouping and merging to all methods. While our method integrates with most architectures, adapting it to keypoint-based methods like FCOS requires more modifications, as they lack bounding box regions. Code is provided at https: //github.com/aritrabhowmik/Uo I. Detection comparison on MS-COCO. Table 1 presents the results of our Union-over-Intersections approach on MS-COCO (Lin et al., 2014), demonstrating consistent improvements across Faster R-CNN, Mask R-CNN, Cascade R-CNN, YOLOv3 and Deformable DETR. Our method enhances detection performance across Res Net-50 and Res Net-101 backbones, as well as different architectures, highlighting its broad applicability. In particular, we integrated our approach into Cascade R-CNN, which refines object proposals over multiple stages with increasing Io U thresholds. Unlike Cascade R-CNN s stage-by-stage refinement of individual proposals, our method refines the combined boxes after grouping within a single stage. We applied our intersection-based regression and union-over-intersection grouping in Cascade RCNN s final stage, improving performance without altering the earlier stages. This demonstrates how our method can complement even complex, multi-stage architectures. Moreover, our approach adapts seamlessly to different detection paradigms: working with proposals in two-stage detectors like Faster R-CNN, grid cells in one-stage detectors like YOLOv3, and queries in transformer-based detectors like Deformable DETR, with minimal modifications. This flexibility allows our method to enhance performance across various architectures with minimal changes. Published as a conference paper at ICLR 2025 Table 2: Integration of Io U-based losses with Uo I. Replacing the L1 loss with Io U-based losses such as GIo U, DIo U, and Alpha-Io U in our method for Faster R-CNN with a Res Net-50 backbone demonstrates consistent performance improvements, highlighting the flexibility and complementary nature of our approach. Io U Loss Type Base (m AP ) Base w/ Uo I (m AP ) Base (AP75 ) Base w/ Uo I (AP75 ) L1 37.4 38.1 40.4 40.9 GIo U 38.0 38.6 41.1 42.0 DIo U 38.1 38.8 41.1 41.9 Alpha-Io U 38.9 39.4 41.8 42.6 Table 3: Instance segmentation comparison on MS-COCO. Our Union-over-Intersections (Uo I) when applied to Mask R-CNN with Res Net50 and Res Net101 backbones improves the segmentation performance. Uo I is especially effective when segmenting large objects. Method Backbone m AP APS APM APL Mask R-CNN R-50 34.7 15.8 36.9 51.1 Mask R-CNN w/ Uo I R-50-fpn 35.3 16.2 37.5 51.7 Mask R-CNN R-101 36.0 16.7 39.1 53.0 Mask R-CNN w/ Uo I R-101-fpn 36.8 17.5 39.8 53.8 Table 4: Uo I is independent of the grouping method. Comparison across NMS variants with and without Uo I on Faster R-CNN using Res Net-50. Results are reported on the COCO validation set. NMS Type Base (m AP ) Base w/ Uo I (m AP ) Base (AP75 ) Base w/ Uo I (AP75 ) NMS 37.4 38.1 40.4 40.9 Cluster-NMS 37.6 38.4 40.4 41.0 Soft-NMS 38.2 38.8 40.9 41.7 Integration of Io U-based losses with Uo I. To evaluate the flexibility of our framework, we replaced the L1 regression loss in our method with Io U-based losses, including GIo U (Rezatofighi et al., 2019), DIo U (Zheng et al., 2020), and Alpha-Io U (He et al., 2021). The results, shown in Table 2, confirm that our Union-over-Intersections (Uo I) approach not only complements these advanced loss functions but also enhances their performance. These findings underscore the versatility of our method, which seamlessly integrates with various loss functions to achieve consistent improvements. Instance segmentation comparison on MS-COCO. We extended our Union-over-Intersections (Uo I) approach to the task of instance segmentation using Mask R-CNN with Res Net50 and Res Net101 backbones (Table 3). Our method consistently improves segmentation performance across various object sizes, with notable gains in mean Average Precision (m AP). Specifically, we observe an increase of 0.7 points for Res Net50 and 0.9 points for Res Net101 backbones. These improvements can be attributed to the effective merging of multiple proposals, which leads to more precise segmentation masks. The benefits are especially evident for larger objects, where tighter and more accurate localization is crucial. This demonstrates the versatility of our method, not only in object detection but also in enhancing instance segmentation tasks across different backbone architectures. 4.3 ABLATION STUDIES Next we perform a series of ablation studies to pinpoint the source of our improvements. Uo I is independent of the grouping method. While we use traditional NMS as the canonical grouping method in our experiments, our approach is not tied to any specific grouping strategy. To evaluate this, we conducted experiments using Cluster-NMS (Zheng et al., 2021), NMS, and Soft NMS (Bodla et al., 2017). From Table 4 we see that across all variants, our Union-over-Intersections (Uo I) approach delivers consistent improvements, demonstrating its robustness and flexibility. This highlights that our method can take advantage of advancements in grouping strategies, making it adaptable to different detection frameworks and improving performance across the board. Our improvements come from better localization only. We first investigate where our improved results come from. The final m AP depends both on accurate object localization and on correctly classifying objects. In Table 5, we dissect the classification and localization performance on Faster R-CNN with and without our approach. We find that the classification accuracy remains nearly Published as a conference paper at ICLR 2025 Table 5: Classification versus localization ablation on MS-COCO. Our approach improves the quality of the object localization without hampering classification accuracy, highlighting that unionover-intersection (Uo I) is more viable for object detection than winner-takes-all. Method Classification Accuracy [%] Localization m Io U [%] Faster R-CNN 76.4 53.7 Faster R-CNN w/ Uo I 76.5 64.4 Table 6: Localization Recall Precision (LRP) comparison for Faster R-CNN on various metrics. Replacing the traditional box proposal selection with Union-over-Intersections (Uo I) reduces LRP errors across all metrics. Method LRP (Error) LRPLoc LRPFP LRPFN Faster R-CNN 67.6 17.2 24.2 44.3 Faster R-CNN w/ Uo I 65.3 12.7 23.9 43.8 identical, while the localization performance obtains a clear leap in mean Io U. Hence, our changes positively affect the localization quality without hampering the classification performance. We also assess our method s performance using the Localization Recall Precision (LRP) metric, introduced by Oksuz et al. (2021) in Table 6. The LRP metric provides a comprehensive evaluation by combining localization error, false positives, and false negatives into a single error measure: LRP (Error): Overall error that incorporates all aspects. LRPLoc: Specifically measures localization error, reflecting how accurately bounding boxes match ground truth objects. LRPFP: Measures false positive errors. LRPFN: Measures false negative errors. Our Union-over-Intersections (Uo I) approach delivers a notable reduction in localization error (LRPLoc), reducing it from 17.2 to 12.7, demonstrating improved bounding box precision. Importantly, this improvement in localization is achieved without too many changes in false positive and false negative rates, where we observe only slight reductions in LRPFP (from 24.2 to 23.9) and LRPFN (from 44.3 to 43.8). The stability in false positives and false negatives suggests that the more accurate the model s class predictions become, the greater the overall improvement we can achieve with our method. Addressing NMS limitations without architectural change. While methods like one-to-few label assignment (Li et al., 2023) address NMS limitations by introducing entirely new architectures, our approach provides a simple, plug-and-play solution that can be applied to existing detection frameworks without architectural changes. For example, one-to-few achieves 40.9 m AP on the COCO validation set using Res Net101, through a modification of the Faster R-CNN baseline (39.4 m AP) with additional architectural changes. In contrast, our method can be seamlessly integrated into a stronger architecture like Cascade R-CNN (Cai & Vasconcelos, 2018) with the same Res Net101 backbone, achieving 43.1 m AP. This demonstrates that our approach not only avoids the complexity of redesigning architectures but also leverages existing frameworks to achieve superior results, which one-to-few cannot. Robustness to poor quality proposals. One might question whether directly regressing to the ground truth from box proposals is problematic, given the large receptive fields in modern networks. In Figure 3 (a), we ablate detection performance based on the quality of the top box proposals. When proposals have high overlap with the ground truth, both methods perform similarly. However, as proposal quality drops, standard pipelines struggle. Direct regression works well only with highquality proposals. Our approach, by looking beyond winner-takes-all, is more robust to lower-quality proposals, improving object localization. Regressing to intersections is simply an easier task. One of the key hypotheses in our work is that regressing to intersection areas simplify the task compared to regressing to the entire ground truth. As shown in Figure 3(b), this simplification results in lower regression loss during training. Notably, despite our dynamic target assignment to object parts, the training remains stable and Published as a conference paper at ICLR 2025 (d) Figure 3: Overview of four ablation studies on MS-COCO with Faster R-CNN. (a) traditional object detection struggles when the best proposal has a low initial overlap with the ground truth, whereas our Union-over-Intersection, which looks beyond the winner-takes-all, is more robust to variations in proposal quality. (b) regressing to intersections is a simpler task to optimize as is evident from the lower loss convergence. (c) the optimal number of top proposals per group for the union-over-intersections is five. (d) as object detector classification performance improves, there are more correctly classified proposals that each cover parts of the ground truth. Instead of selecting a single proposal and discarding the rest, it is more effective to use all proposals to create a more comprehensive representation of the object. Table 7: Test-time augmentation benefits us further. Both box voting and Uo I involve grouping predictions but differ in approach: box voting merges regressed boxes, while Uo I groups original proposals before merging intersections. Test-time augmentation enhances both methods, but Uo I remains effective without it, highlighting its robustness compared to box voting s reliance on test-time augmentation. Single-scaling uses an image scale of 666x400, while Multi-scaling incorporates both 666x400 and 2000x1200 as image scales. Test-time adaptation setting Method None Flip Flip + Single-scaling Flip + Multi-scaling Faster R-CNN w/ Box voting 37.5 37.7 38.0 38.8 Faster R-CNN w/ Uo I 38.1 38.2 38.6 39.3 achieves improved localization accuracy during inference, as evidenced by Table 5. We conclude that regressing to intersections not only simplifies object detection but also enhances final detection performance through the union-over-intersections strategy. Intermediate group size is best for union-over-intersection. Figure 3 (c) shows the object detection performance as a function of the number of proposals used per group when taking the union over intersections. Our approach consistently uses a maximum of five top proposals per region across all datasets and architectures, with all available proposals used if fewer than five are present. Performance improves from 2 to 5 proposals, then drops due to noisy, low-scoring proposals, achieving the best results with 5 proposals. This setting balances sufficient information for box merging while avoiding overly large boxes. Test-time augmentation benefits us further. We analyze the impact of test-time augmentation (TTA) on our Union-over-Intersections (Uo I) strategy and compare it with box voting (Roman et al., 2019), which involves confidence-weighted box merging. TTA involves applying multi-scale testing and horizontal flips during inference, generating additional predictions for aggregation. Box voting relies heavily on the increased diversity of predictions from TTA, while Uo I proves effective even without TTA and further benefits when combined with it. With better classification, Uo I becomes more powerful. We examine whether improvements in object classification will reduce or expand the gap between traditional methods and our approach. In an oracle experiment, we simulate varying proportions of known ground truth labels (20%, 40%, 50%, 60%, 80%, and 100%) while the remaining predictions come from the network. As shown in Figure 3 (d), as the accuracy of classification labels improves, our Union-over-Intersections method increasingly outperforms traditional techniques. This is because correctly classified proposals, each covering parts of the ground truth, combine to form a more comprehensive representation. In contrast, traditional methods, which rely on a single proposal, often miss portions of the object. These results highlight that as classification improves, the advantages of Uo I become more pronounced. Published as a conference paper at ICLR 2025 Table 8: Effectiveness of the second regression stage. Uo I inherently involves a second regression stage, where the combined intersections are further refined for improved performance. Adding just a second regression stage to the Faster R-CNN baseline alone does not yield similar benefits. Method m AP AP75 Faster R-CNN 37.4 40.4 Faster R-CNN w/ 2nd regression 37.4 40.5 Faster R-CNN w/ Uo I 38.1 40.9 Detection Results Limitations Faster R-CNN Faster R-CNN w/Uo I Figure 4: Qualitative analysis showcasing the effectiveness of our method applied to Faster R-CNN for object detection. Starting from the left, our method successfully removes the incorrect person prediction, precisely localizes the entire bird, and identifies the third cat as a distinct entity. However, in scenarios with clutter or a lot of overlap, as observed in the right image, our approach also consolidates multiple parrots into a single detection. Effectiveness of the second regression stage. We include an ablation study to analyze the necessity of the second regression stage in our method. Unlike baseline methods, where an additional regression head does not provide benefits due to redundant tasks, our approach assigns distinct roles to the regression stages. The first head focuses on parts of the object, while the second refines the union of these parts into a cohesive bounding box. As shown in Table 8, this distinction makes the second regression stage effective, further improving performance. Limitations. Our method benefits from multiple proposals, allowing the Union-over-Intersection strategy to unify their information effectively. However, for small objects, if only one proposal is available, our approach defaults to the baseline. Figure 4 highlights another limitation with crowded same-class objects, where closely positioned instances may merge into a single box ( 3 birds example). Advanced grouping strategies could mitigate this issue. 5 CONCLUSION In this paper, we introduced a plug-and-play approach to object detection that enhances localization by focusing on the intersection between proposals and ground truth and applying union-over-intersections instead of a winner-takes-all strategy. These simple changes, requiring minimal modifications to existing architectures, consistently improve performance across two-stage, single-stage, and transformer-based detectors for object detection as well as instance segmentation. Our method proves especially robust for stricter overlap thresholds and lower-quality proposals, making it highly flexible across various frameworks. By improving localization, our approach offers a practical solution that strengthens object detection and segmentation without architectural dependencies. Acknowledgement. This work has been financially supported by Tom Tom, the University of Amsterdam and the allowance of Top consortia for Knowledge and Innovation (TKIs) from the Netherlands Ministry of Economic Affairs and Climate Policy. We also extend our gratitude to the anonymous reviewers for their valuable feedback and insightful suggestions during the rebuttal stage, which considerably improved this work. Published as a conference paper at ICLR 2025 Josh Beal, Eric Kim, Eric Tzeng, Dong Huk Park, Andrew Zhai, and Dmitry Kislyuk. Toward transformer-based object detection. ar Xiv preprint ar Xiv:2012.09958, 2020. Alexey Bochkovskiy, Chien-Yao Wang, and Hong-Yuan Mark Liao. Yolov4: Optimal speed and accuracy of object detection. ar Xiv preprint ar Xiv:2004.10934, 2020. Navaneeth Bodla, Bharat Singh, Rama Chellappa, and Larry S Davis. Soft-nms improving object detection with one line of code. In Proceedings of the IEEE international conference on computer vision, pp. 5561 5569, 2017. Michael C. Burl, Markus Weber, and Pietro Perona. A probabilistic approach to object recognition using local photometry and global geometry. In Proceedings of the European conference on computer vision, 1998. Zhaowei Cai and Nuno Vasconcelos. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6154 6162, 2018. Nicolas Carion, Francisco Massa, Gabriel Synnaeve, Nicolas Usunier, Alexander Kirillov, and Sergey Zagoruyko. End-to-end object detection with transformers. In Proceedings of the European conference on computer vision, pp. 213 229. Springer, 2020. Xiyang Dai, Yinpeng Chen, Jianwei Yang, Pengchuan Zhang, Lu Yuan, and Lei Zhang. Dynamic detr: End-to-end object detection with dynamic attention. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 2988 2997, 2021a. Zhigang Dai, Bolun Cai, Yugeng Lin, and Junying Chen. Up-detr: Unsupervised pre-training for object detection with transformers. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 1601 1610, 2021b. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the International conference on learning representations, 2021. Kaiwen Duan, Song Bai, Lingxi Xie, Honggang Qi, Qingming Huang, and Qi Tian. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 6569 6578, 2019. Mark Everingham, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The pascal visual object classes (voc) challenge. International Journal of Computer Vision, 88: 303 338, 2010. Pedro F. Felzenszwalb, Ross B. Girshick, David Mc Allester, and Deva Ramanan. Object detection with discriminatively trained part-based models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010. Zheng Ge, Songtao Liu, Feng Wang, Zeming Li, and Jian Sun. Yolox: Exceeding yolo series in 2021. ar Xiv preprint ar Xiv:2107.08430, 2021. Ross Girshick. Fast r-cnn. In Proceedings of the IEEE international conference on computer vision, pp. 1440 1448, 2015. Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 580 587, 2014. Jiabo He, Sarah Erfani, Xingjun Ma, James Bailey, Ying Chi, and Xian-Sheng Hua. Alpha-Io U: A family of power intersection over union losses for bounding box regression. In Advances in neural information processing systems, volume 34, pp. 20230 20242, 2021. Published as a conference paper at ICLR 2025 Kaiming He, Georgia Gkioxari, Piotr Doll ar, and Ross Girshick. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision, pp. 2961 2969, 2017. Jan Hosang, Rodrigo Benenson, and Bernt Schiele. Learning non-maximum suppression. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4507 4515, 2017. Rachel Huang, Jonathan Pedoeem, and Cuixian Chen. Yolo-lite: a real-time object detection algorithm optimized for non-gpu computers. In Proceedings of the IEEE international conference on big data, pp. 2503 2510, 2018. Borui Jiang, Ruixuan Luo, Jiayuan Mao, Tete Xiao, and Yuning Jiang. Acquisition of localization confidence for accurate object detection. In Proceedings of the European conference on computer vision (ECCV), pp. 784 799, 2018. Hei Law and Jia Deng. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European conference on computer vision, pp. 734 750, 2018. Shuai Li, Minghan Li, Ruihuang Li, Chenhang He, and Lei Zhang. One-to-few label assignment for end-to-end dense detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 7350 7359, 2023. Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll ar, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In Proceedings of the European conference on computer vision, pp. 740 755. Springer, 2014. Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Doll ar. Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision, pp. 2980 2988, 2017. Shilong Liu, Feng Li, Hao Zhang, Xiao Yang, Xianbiao Qi, Hang Su, Jun Zhu, and Lei Zhang. Dab-detr: Dynamic anchor boxes are better queries for detr. In Proceedings of the international conference on learning representations, 2022. Songtao Liu, Di Huang, and Yunhong Wang. Adaptive nms: Refining pedestrian detection in a crowd. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 6459 6468, 2019. Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, and Alexander C Berg. Ssd: Single shot multibox detector. In Proceedings of the European conference on computer vision, pp. 21 37. Springer, 2016. Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 10012 10022, 2021. Xin Lu, Buyu Li, Yuxin Yue, Quanquan Li, and Junjie Yan. Grid r-cnn. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 7363 7372, 2019. Depu Meng, Xiaokang Chen, Zejia Fan, Gang Zeng, Houqiang Li, Yuhui Yuan, Lei Sun, and Jingdong Wang. Conditional detr for fast training convergence. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 3651 3660, 2021. Duy-Kien Nguyen, Jihong Ju, Olaf Booij, Martin R Oswald, and Cees GM Snoek. Boxer: Boxattention for 2d and 3d transformers. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 4773 4782, 2022. Kemal Oksuz, Baris Can Cam, Sinan Kalkan, and Emre Akbas. One metric to measure them all: Localisation recall precision (lrp) for evaluating visual detection tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(12):9446 9463, 2021. Jiangmiao Pang, Kai Chen, Jianping Shi, Huajun Feng, Wanli Ouyang, and Dahua Lin. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 821 830, 2019. Published as a conference paper at ICLR 2025 David Paz, Hengyuan Zhang, and Henrik I Christensen. Tridentnet: A conditional generative model for dynamic trajectory generation. In Proceedings of the international conference on intelligent autonomous systems, pp. 403 416. Springer, 2021. Joseph Redmon and Ali Farhadi. Yolov3: An incremental improvement. ar Xiv preprint ar Xiv:1804.02767, 2018. Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 779 788, 2016. Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in neural information processing systems, volume 28, 2015. Hamid Rezatofighi, Nathan Tsoi, Jun Young Gwak, Amir Sadeghian, Ian Reid, and Silvio Savarese. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 658 666, 2019. Solovyev Roman, Weimin Wang, and Tatiana Gabruseva. Weighted boxes fusion: ensembling boxes for object detection models. ar Xiv preprint ar Xiv:1910.13302, 2019. Zhiqing Sun, Shengcao Cao, Yiming Yang, and Kris M Kitani. Rethinking transformer-based set prediction for object detection. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 3611 3620, 2021. Mingxing Tan and Quoc Le. Efficient Net: Rethinking model scaling for convolutional neural networks. In International conference on machine learning, pp. 6105 6114. PMLR, 2019. Mingxing Tan, Ruoming Pang, and Quoc V Le. Efficient Det: Scalable and efficient object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10781 10790, 2020. Zhi Tian, Chunhua Shen, Hao Chen, and Tong He. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 9627 9636, 2019. Rohun Tripathi, Vasu Singla, Mahyar Najibi, Bharat Singh, Abhishek Sharma, and Larry Davis. Asap-nms: Accelerating non-maximum suppression using spatially aware priors. ar Xiv preprint ar Xiv:2007.09785, 2020. Lachlan Tychsen-Smith and Lars Petersson. Improving object localization with fitness nms and bounded iou loss. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 6877 6885, 2018. Paul Viola and Michael J. Jones. Robust real-time face detection. International Journal of Computer Vision, 2004. Xinlong Wang, Rufeng Zhang, Tao Kong, Lei Li, and Chunhua Shen. Solov2: Dynamic and fast instance segmentation. In Advances in neural information processing systems, volume 33, pp. 17721 17732, 2020. Shengkai Wu, Xiaoping Li, and Xinggang Wang. Iou-aware single-stage object detector for accurate localization. Image and Vision Computing, 97:103911, 2020a. Yue Wu, Yinpeng Chen, Lu Yuan, Zicheng Liu, Lijuan Wang, Hongzhi Li, and Yun Fu. Rethinking classification and localization for object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10186 10195, 2020b. Jiahui Yu, Yuning Jiang, Zhangyang Wang, Zhimin Cao, and Thomas Huang. Unitbox: An advanced object detection network. In Proceedings of the ACM international conference on multimedia, pp. 516 520, 2016. Published as a conference paper at ICLR 2025 Hao Zhang, Feng Li, Shilong Liu, Lei Zhang, Hang Su, Jun Zhu, Lionel M Ni, and Heung-Yeung Shum. Dino: Detr with improved denoising anchor boxes for end-to-end object detection. ar Xiv preprint ar Xiv:2203.03605, 2022. Haoyang Zhang, Ying Wang, Feras Dayoub, and Niko Sunderhauf. Varifocalnet: An iou-aware dense object detector. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8514 8523, 2021. Zhaohui Zheng, Ping Wang, Wei Liu, Jinze Li, Rongguang Ye, and Dongwei Ren. Distance-iou loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pp. 12993 13000, 2020. Zhaohui Zheng, Ping Wang, Dongwei Ren, Wei Liu, Rongguang Ye, Qinghua Hu, and Wangmeng Zuo. Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE Transactions on Cybernetics, 52(8):8574 8586, 2021. Xizhou Zhu, Weijie Su, Lewei Lu, Bin Li, Xiaogang Wang, and Jifeng Dai. Deformable detr: Deformable transformers for end-to-end object detection. In Proceedings of the international conference on learning representations, 2021. Published as a conference paper at ICLR 2025 A.1 ADDITIONAL RESULTS ON PASCAL VOC To provide a more comprehensive evaluation, we include results for Faster R-CNN, Mask R-CNN, Cascade R-CNN, YOLOv3 and Deformable DETR on the PASCAL VOC dataset. As shown in Table 9, our Union-over-Intersections (Uo I) strategy also consistently improves detection performance across all methods on PASCAL VOC, with again stronger gains observed at higher overlap thresholds. Table 9: Comparison on PASCAL VOC with additional methods. Adding Uo I improves performance across all methods, particularly for higher overlap thresholds. Method Backbone AP50 AP70 AP90 Faster R-CNN R-50-fpn 71.2 51.3 6.8 Faster R-CNN w/ Uo I R-50-fpn 72.9 55.3 11.3 Mask R-CNN R-50-fpn 73.1 52.7 7.9 Mask R-CNN w/ Uo I R-50-fpn 74.0 56.9 13.1 Cascade R-CNN R-50-fpn 77.5 54.1 10.5 Cascade R-CNN w/ Uo I R-50-fpn 78.8 57.9 14.8 YOLOv3 Dar Net-53 56.1 41.7 5.3 YOLOv3 w/ Uo I Dar Net-53 57.0 44.2 8.8 Deformable DETR R-50-fpn 83.2 68.5 22.1 Deformable DETR w/ Uo I R-50-fpn 84.1 70.3 26.5 A.2 EFFECT OF IOU THRESHOLD ON GROUPING STRATEGY To evaluate the impact of the Io U threshold (k) on our Uo I strategy, we conducted an ablation study using Faster R-CNN with a Res Net-101 backbone. The Io U threshold influences group sizes during the grouping stage: lowering the threshold (e.g., 0.3) increases group sizes, while raising it (e.g., 0.7) decreases group sizes. As shown in Table 10, the optimal number of proposals shifts to 4 for k=0.7 but remains at 5 for k=0.3. Importantly, the highest m AP is achieved with k=0.5, reaffirming the effectiveness of the default setting. These results further validate the robustness of our approach to the Io U threshold choice. Table 10: Effect of Io U threshold (k) on grouping strategy. Lowering or raising the Io U threshold shifts the optimal number of proposals, but the highest m AP is consistently achieved with k = 0.5. Number of Proposals Io U Threshold k 3 4 5 6 7 0.3 38.1 39.0 39.5 39.2 38.4 0.5 39.3 39.8 40.3 39.2 38.5 0.7 38.5 39.9 39.4 38.1 37.7 A.3 IMPACT OF ADDITIONAL CLASSIFICATION HEAD ON CONFIDENCE SCORES To assess the effect of adding a classification head for the combined boxes, we conducted an ablation using the Faster R-CNN baseline. As shown in Table 11, this modification resulted in marginal improvements in m AP and AP75. However, since the gains are minimal and the primary improvements in our approach stem from better localization (see Table 5 in the main paper), we did not include the additional classification stage in our main method. These results are provided here for completeness. A.4 EVALUATING FULL-OBJECT REGRESSION VERSUS PART-BASED REGRESSION To evaluate the implications of applying the Union of Proposals with the original regression objective, we conducted an experiment using full-object regression, where the model predicts the complete ground truth bounding boxes instead of focusing on part-based regions. As shown in the table Published as a conference paper at ICLR 2025 Table 11: Effect of adding a classification head for combined boxes. Adding a classification stage offers minimal improvements, reaffirming that the primary gains of our approach come from improved localization Method m AP AP75 Faster R-CNN w/ Uo I 38.1 40.9 Faster R-CNN w/ Uo I and 2nd classification stage 38.2 41.1 below, full-object regression results in looser bounding boxes after combining proposals, even with a second-stage regression. This leads to a considerable drop in performance compared to our partbased regression strategy (Uo I). The results confirm that focusing on parts of objects enables finer localization and better overall accuracy. Table 12: Comparison of full-object regression versus part-based regression (Uo I). Part-based regression achieves better performance by enabling finer localization. Method m AP AP75 Full-object regression 35.8 37.5 Part-based regression (Uo I) 38.1 40.9 A.5 ADAPTATION OF UOI FOR DEFORMABLE DETR AND YOLO Here we provide more details on how our approach integrates with Deformable DETR and YOLO. Code is provided at https://github.com/aritrabhowmik/Uo I. In the standard Deformable DETR, each query predicts a bounding box via the regression head after the decoder. These bounding boxes are then matched to full ground truth boxes using Hungarian matching, with the best query assigned to each ground truth box. Our method modifies this matching process. Specifically, we divide each ground truth box into four quadrants (top-left, top-right, bottomleft, bottom-right), treating these quadrants as part-level targets. During training, each query is assigned to a specific quadrant based on its Io U overlap, and the regression head is tasked with predicting the corresponding part of the ground truth. At inference, these part-level predictions are grouped using our Intersection-based Grouping strategy and merged into complete bounding boxes. Similarly, for YOLO, we adapt its object-to-grid assignment strategy. Instead of assigning an object to a single grid cell based on the object s center, we use an Io U-based criterion. An object is assigned to multiple grid cells if it overlaps sufficiently with their anchors. Each anchor predicts the part of the object it best overlaps with, and during inference, our grouping strategy merges these predictions into final bounding boxes. These modifications retain the original architecture of YOLO while enabling part-based regression and grouping. A.6 MORE QUALITATIVE RESULTS Published as a conference paper at ICLR 2025 Detection Results Limitations Faster R-CNN Faster R-CNN w/Uo I Figure 5: Qualitative analysis showcasing the effectiveness of our method applied to Faster R-CNN for object detection. Starting from the left, our Uo I approach diminishes false positives and achieves tighter localization for both the bird and the cow, outperforming the baseline method in detecting large objects. However, in scenarios with clutter or significant overlap, our method may also aggregate objects into a single group, as seen in the rightmost image with the two cars grouped as one in the right corner. Figure 6: Qualitative analysis of our method for instance segmentation on the COCO 2017 dataset. Our method merges information from multiple proposals, naturally leading to precise localization and segmentation masks.