# clusterfomer_clustering_as_a_universal_visual_learner__6e06a1d1.pdf CLUSTERFORMER: Clustering As A Universal Visual Learner James C. Liang Rochester Institute of Technology Yiming Cui University of Florida Qifan Wang Meta AI Tong Geng University of Rochester Wenguan Wang Zhejiang University Dongfang Liu Rochester Institute of Technology This paper presents CLUSTERFORMER, a universal vision model that is based on the CLUSTERing paradigm with Trans FORMER. It comprises two novel designs: ① recurrent cross-attention clustering, which reformulates the cross-attention mechanism in Transformer and enables recursive updates of cluster centers to facilitate strong representation learning; and ②feature dispatching, which uses the updated cluster centers to redistribute image features through similarity-based metrics, resulting in a transparent pipeline. This elegant design streamlines an explainable and transferable workflow, capable of tackling heterogeneous vision tasks (i.e., image classification, object detection, and image segmentation) with varying levels of clustering granularity (i.e., image-, box-, and pixel-level). Empirical results demonstrate that CLUSTERFORMER outperforms various well-known specialized architectures, achieving 83.41% top-1 acc. over Image Net-1K for image classification, 54.2% and 47.0% m AP over MS COCO for object detection and instance segmentation, 52.4% m Io U over ADE20K for semantic segmentation, and 55.8% PQ over COCO Panoptic for panoptic segmentation. For its efficacy, we hope our work can catalyze a paradigm shift in universal models in computer vision. 1 Introduction CLUSTERFORMER Figure 1: CLUSTERFORMER is a clustering-based universal model, offering superior performance over various specialized architectures. Computer vision has seen the emergence of specialized solutions for different vision tasks (e.g., Res Net [34] for image classification, Faster RCNN [70] for object detection, and Mask RCNN [33] for instance segmentation), aiming for superior performance. Nonetheless, neuroscience research [73, 65, 82, 5] has shown that the human perceptual system exhibits exceptional interpretive capabilities for complex visual stimuli, without task-specific constraints. This trait of human perceptual cognition diverges from current computer vision techniques [95, 44, 46], which often employ diverse architectural designs. Human vision possesses a unique attention mechanism that selectively focuses on relevant parts of the visual field while disregarding irrelevant information [81, 40]. This can be likened to a clustering approach [2, 3, 89], in which individual pixel points are decomposed and reorganized into relevant concepts to address various tasks. This essentially is a hierarchical process that involves combining Corresponding author. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). basic visual features, such as lines, shapes, and colors, to create higher-level abstractions of objects, scenes, and individuals [79, 59, 66, 27] . Inspired by the remarkable abilities of the human vision system, this work aims to develop a universal vision model that can replicate the unparalleled prowess. To this end, we employ a clustering-based strategy that operates at varying levels of granularity for visual comprehension. By solving different vision tasks (i.e., image classification, object detection, and image segmentation), we take into account the specificity at which visual information is grouped (i.e., image-, box-, and pixel-level). We name our approach, CLUSTERFORMER ( 3.2), as it utilizes a CLUSTERing mechanism integrated within the Trans FORMER architecture to create a universal network. The method begins by embedding images into discrete tokens, representing essential features that are grouped into distinct clusters. The cluster centers are then recursively updated through a recurrent clustering cross-attention mechanism that considers associated feature representations along the center dimension. Once center assignments and updates are complete, features are dispatched based on updated cluster centers, and then both are fed into the task head for the target tasks. CLUSTERFORMER enjoys a few attractive qualities. ❶Flexibility: CLUSTERFORMER is a clusteringanchored approach that accommodates a broad array of visual tasks with superior performance (see Fig. 1) under one umbrella. The core epistemology is to handle various tasks with different levels of granularity (e.g., image-level classification, box-level detection, pixel-level segmentation, etc.), moving towards a universal visual solution. ❷Transferability: The cluster centers generated by the CLUSTERFORMER encoder are directly employed by the task head as initial queries for clustering, allowing the entire architecture to transfer underlying representation for target-task predictions (see Table. 4). This elegant design facilitates the transferability of knowledge acquired from the upstream task (i.e., encoder trained on Image Net [72]) to downstream tasks (e.g., decoder trained on instance segmentation on COCO [49]). ❸Explainability: Regardless of the target tasks, CLUSTERFORMER s decision-making process is characterized by a transparent pipeline that continuously updates cluster centers through similarity-based metrics. Since the reasoning process is naturally derivable, the model inference behavior is ad-hoc explainable (see 4.2). This differs CLUSTERFORMER from most existing unified models [17, 44, 95] that fail to elucidate precisely how a model works. To effectively assess our method, we experimentally show: In 4.1.1, with the task of image classification, CLUSTERFORMER outperforms traditional counterparts, e.g., 0.13 0.39% top-1 accuracy compared with Swin Transformoer [53] on Image Net [72], by training from scratch. In 4.1.2, when using our Image Net-pretrained, our method can be expanded to the task of object detection and greatly improve the performance compared to Dino [96] over Swin Transformer on COCO [49] (0.8 1.1% m AP). In addition, our method can also adapt to more generic per-pixel tasks, a.k.a, semantic segmentation (see 4.1.3), instance segmentation (see 4.1.4), and panoptic segmentation (see 4.1.5). For instance, we achieve performance gains of 0.6 1.3% m Io U for semantic segmentation on ADE20K [101], 1.0 1.4% m AP for instance segmentation on MS COCO [49] and 1.5 1.7% PQ for panoptic segmentation on COCO Panoptic [42] compared with Mask2Former [17] over Swin Transformer. Our algorithm are extensively tested, and the efficacy for the core components is also demonstrated through a series of ablative studies outlined in 4.2, 2 Related Work Universal Vision Model. Transformers [81] have been instrumental in driving universal ambition, fostering models that are capable of tackling tasks of different specificity with the same architecture and embody the potential of these recent developments [23, 17, 16, 95, 96, 4, 80, 30, 57, 86] in the field. In the vision regime, mainstream research endeavors have been concentrating on the development of either encoders [53, 88] or decoders [44, 94]. The encoder is centered around the effort of developing foundation models [4, 53, 24, 22], trained on extensive data that can be adapted and fine-tuned to diverse downstream tasks. For instance, Swin Transformer [53] capably serves as a general-purpose backbone for computer vision by employing a hierarchical structure consisting of shifted windows; Vi T-22B [22], parameterizes the architecture to 22 billion and achieves superior performance on a variety of vision tasks through learning large-scale data. Conversely, research on decoders [23, 17, 16, 95, 94, 44, 96, 87, 50, 20, 52, 19, 21, 51, 93, 76, 37, 99, 25, 48] is designed to tackle homogeneous target tasks, by using queries to depict visual patterns. For instance, Mask2Former [17] incorporates mask information into the Transformer architecture and unifies various segmentation tasks (e.g., semantic, instance, and panoptic segmentation); Mask-DINO [44] extends the decoding process from detection to segmentation by directly utilizing query embeddings for target task predictions. Conceptually different, we streamline an elegant systemic workflow based on clustering and handle heterogeneous visual tasks (e.g., image classification, object detection, and image segmentation) at different clustering granularities. Clustering in Vision. Traditional clustering algorithms in vision [39, 28, 29, 55, 91, 1, 10, 61, 6, 58] can be categorized into the hierarchical and partitional modes. The hierarchical methods [62, 38] involve the modeling of pixel hierarchy and the iterative partitioning and merging of pixel pairs into clusters until reaching a state of saturation. This approach obviates the necessity of a priori determination of cluster quantity and circumvents the predicaments arising from local optima. [98, 12]. However, it exclusively considers the adjacent pixels at each stage and lacks the capacity to assimilate prior information regarding the global configuration or dimensions of the clusters. [69, 64]. In contrast, partitional clustering algorithms [78, 36] directly generate a flat structure with a predetermined number of clusters and exclusively assign pixels to a single cluster. This design exhibits a dynamic nature, allowing pixels to transition between clusters [11, 63]. By employing suitable measures, this approach can effectively integrate complex knowledge within cluster centers. As a powerful system, human vision incorporates the advantages of both clustering modes [89, 83, 67]. We possess the capability of grouping analogous entities at different scales. Meanwhile, we can also effectively categorize objects purely based on their shape, color, or texture, without having the hierarchical information. Drawing on the above insights, we reformulate the attention mechanism ( 3.2 ) in Transformer architectures [81] from the clustering s perspective to decipher the hierarchy of visual complexity. 3 Methodology 3.1 Preliminary Clustering. The objective of clustering is to partition a set of data points, denoted by X Rn d, into C distinct clusters based on their intrinsic similarities while ensuring that each data point belongs to only one cluster. Achieving this requires optimizing the stratification of the data points, taking into account both their feature and positional information, to form coherent and meaningful groupings. Clustering methodologies typically employ advanced similarity metrics, such as cosine similarity, to measure the proximity between data points and cluster centroids. Additionally, they consider the spatial locality of the points to make more precise group assignments. Cross-Attention for Generic Clustering. Drawing inspiration from the Transformer decoder architecture [81], contemporary end-to-end architecture [17, 9] utilize a query-based approach in which a set of K queries, C = [c1; ; c K] RK D, are learned and updated by a series of crossattention blocks. In this context, we rethink the term C to associate queries with cluster centers at each layer. Specifically, cross-attention is employed at each layer to adaptively aggregate image features and subsequently update the queries: C C + softmax HW (QC(KI) )V I, (1) where QC RK D, V I RHW D, KI RHW D represent linearly projected features for query, key, and value, respectively. The superscripts C and I denote the features projected from the center and image features, respectively. Motivated by [95], we follow a reinterpretation of the cross-attention mechanism as a clustering solver by considering queries as cluster centers and applying the softmax function along the query dimension (K) instead of the image resolution (HW): C C + softmax K(QC(KI) )V I. (2) 3.2 CLUSTERFORMER In this subsection, we present CLUSTERFORMER (see Fig. 2(a)). The model has a serial of hierarchical stages that enables multi-scale representation learning for universal adaptation. At each stage, image patches are tokenized into feature embedding [81, 53, 24], which are grouped into distinct clusters via a unified pipeline first recurrent cross-attention clustering and then feature dispatching. Recurrent Cross-Attention Clustering. Considering the feature embeddings I RHW D and initial centers C(0), we encapsulate the iterative Expectation-Maximization (EM) clustering process, C (0) C (0) C (0) C (1) C (1) C (T) C (T) E-step M-step E-step M-step M (0) M (0) M (0) M (T) ˆ M (T) ˆ M (t) M (t) M (t) ˆ ˆ M (t) Recurrent Cross-Attention Clustering Feature Dispatching Universal Adaptation Feature Dispatching C (T) C (T) C (T) Similarity (a) Recurrent Cross-Attention Clustering T iterations Stage 1 Stage L Figure 2: (a) Overall pipeline of CLUSTERFORMER. (b) Each Recurrent Cross-Attention Clustering layer carries out T iterations of cross-attention clustering (E-step) and center updating (M-step) (see Eq. 3). (c) The feature dispatching redistributes the feature embeddings on the top of updated cluster centers (see Eq. 6). consisting of T iterations, within a Recurrent EM Cross-Attention layer (see Fig. 2(b)): E-step: ˆ M (t) =softmax K(QC(t)(KI) ), M-step: C(t+1) = ˆ M (t)V I RK D, (3) where t {1, , T} and ˆ M [0, 1]K HW represents the soft cluster assignment matrix (i.e., probability maps of K clusters). As defined in Section 3.1, QC RK D denotes the query vector projected from the center C, and V I, KI RHW D correspond to the value and key vectors, respectively, projected from the image features I. The Recurrent Cross-Attention approach iteratively updates cluster membership ˆ M (i.e., E-step) and centers C (i.e., M-step). This dynamic updating strategy embodies the essence of partitional clustering. It enjoys a few appealing characteristics: Efficiency: While the vanilla self-attention mechanism has a time complexity of O(H2W 2D), the Recurrent Cross-Attention approach exhibits a lower bound of O(TKHWD). This is primarily due to the fact that TK HW (i.e., 4165 in Swin [53] vs. 1200 in ours). Specifically, considering the nature of the pyramid architecture [88, 53] during the encoding process, TK can indeed be much smaller than HW, especially in the earlier stages. It is important to note that during each iteration, merely the Q matrix requires an update, while the K and V matrices necessitate a single computation. Consequently, the whole model enjoys systemic efficiency (see Table 6c). Transparency: The transparency hinges on the unique role that cluster centers play in our Recurrent Cross-Attention mechanism. The cluster centers, derived through our clustering process, act as prototypes for the features they cluster. These prototypes serve as a form of a representative sample for each cluster, reflecting the most salient or characteristic features of the data points within that cluster. Moreover, the Recurrent Cross-Attention method adheres to the widely-established EM clustering algorithm, offering a lucid and transparent framework. This cluster center assignment behaves in a human-understandable manner (see Fig. 3) during representation learning and fosters ad-hoc explainability, allowing for a more intuitive understanding of the underlying relationships. Non-parametric fashion: The Recurrent Cross-Attention mechanism achieves a recursive nature by sharing the projection weights for query, key, and value across iterations. This approach effectively ensures recursiveness without the introduction of additional learnable parameters (see Table 6b). Since the overall architecture is hierarchical, Recurrent Cross-Attention is able to thoroughly explore the representational granularity, which mirrors the process of hierarchical clustering: Cl = RCAl(Il, Cl 0), (4) where RCA stands for the recurrent cross-attention layer. Il is the image feature map at different layers by standard pooling operation with H/2l W/2l resolution. Cl is the cluster center matrix for lth layer and Cl 0 is the initial centers at lth layer. The parameters for Recurrent Cross-Attention at different layers, i.e., {RCAl}L l=1, are not shared. In addition, we initialize the centers from image grids: [c(0) 1 ; ; c(0) K ] = FFN(Adptive_Pooling K(I)), (5) where FFN stands for Position-wise Feedforward Network which is an integral part of the Transformer architecture. It comprises two fully connected layers along with an activation function used in the hidden layer. Adptive_Pooling K(I) refers to select K feature centers from I using adaptive sampling, which calculates an appropriate window size to achieve a desired output size adaptively, offering more flexibility and precision compared to traditional pooling methods. Feature Dispatching. After the cluster assignment, the proposed method employs an adaptive process that dispatches each patch within a cluster based on similarity (see Fig. 2(c)), leading to a more coherent and representative understanding of the overall structure and context within the cluster. For every patch embedding pi I, the updated patch embedding p i is computed as: p i = pi + MLP( 1 k=0 sim(Ck, pi) Ck) (6) This equation represents the adaptive dispatching of feature embeddings by considering the similarity between the feature embedding and the cluster centers (C), weighted by their respective similarities. By incorporating the intrinsic information from the cluster centers, the method refines the feature embeddings, enhancing the overall understanding of the image s underlying structure and context. All feature representations are utilized for handling the target tasks in the decoding process. In 3.3, we discuss more details about the implementation of the ending tasks. 3.3 Implementation Details The implementation details and framework of CLUSTERFORMER are shown in (Fig. 2a). We followed the architecture and configuration of Swin Transformer [53]. The code will be available at here. Encoder. The encoding process is to generate presentation hierarchy, denoted as {Il} with l = {1, 2, 3, 4}, for a given image I. The pipeline begins with the feature embedding to convert the images into separate feature tokens. Subsequently, multi-head computing [81, 53] is employed to partition the embedded features among them. Center initialization (Eq. 5) is then adopted as a starting for initializing the cluster centers, and the recurrent cross-attention clustering (Eq. 3) is utilized to recursively update these centers. Once the centers have been updated, the features are dispatched based on their association with the updated centers (Eq. 6). The further decoding process leverage both the centers and the features, which guarantees well-rounded learning. Adaptation to Image Classification. The classification head is a single-layer Multilayer Perceptron (MLP) takes the cluster centers from the encoder for predictions. Adaptation to Detection and Segmentation. Downstream task head has six Transformer decoder layers with the core design of recurrent cross-attention clustering (Eq.4). Each layer has 3 iterations. 4 Experiment We evaluate our methods over five vision tasks viz. image classification, object detection, semantic segmentation, instance segmentation, and panoptic segmentation on four benchmarks. Image Net-1K for Image Classification. Image Net-1K[72] includes high-resolution images spanning distinct categories (e.g., animals, plants, and vehicles). Following conventional procedures, the dataset is split into 1.2M/50K/100K images for train/validation/test splits. MS COCO for Object Detection and Instance Segmentation. COCO [49] dataset features dense annotations for 80 common objects in daily contexts. Following standard practices [49], the dataset is split into 115K/5K/20K images for train2017/val2017/test-dev splits. ADE20K for Semantic Segmentation. ADE20K [101] dataset offers an extensive collection of images with pixel-level annotations, containing 150 diverse object categories in both indoor and outdoor scenes. The dataset comprises 20K/2K/3K images for train/val/test splits. COCO Panoptic for Panoptic Segmentation. The COCO Panoptic dataset [42] includes 80 thing categories and a carefully annotated set of 53 stuff categories. In line with standard practices [42], the COCO Panoptic dataset is split into 115K/5K/20K images for the train/val/test splits as well. The ensuing section commences by presenting the main results of each task ( 4.1), succeeded by a series of ablative studies ( 4.2), which aim to confirm the efficacy of each modulating design. 4.1 Main Results 4.1.1 Experiments on Image Classification Training. We use mmclassification2 as codebase and follow the default training settings. The default configuration for our model involves setting the number of centers to 100. To optimize the model s performance, we employ cross-entropy as the default loss function, which is widely used in classification tasks and helps in minimizing the difference between predicted probabilities and ground truth. For the training details, we run the model for 300 epochs, allowing sufficient time for the model to learn and converge. To manage the learning rate, we initialize it at 0.001 as default. The learning rate is then scheduled using a cosine annealing policy, which gradually decreases the learning rate over time. Due to limitations in our GPU capacity, we are constrained to set the total batch size at 1024. Models are trained from scratch on sixteen A100GPUs. Table 1: Classification top-1 and top-5 accuracy on Image Net [72] val (see 4.1.1 for details). Method #Params top-1 top-5 Context Cluster-Tiny[ICLR23][58] 5.3M 71.68% 90.49% Dei T-Tiny[ICML21][80] 5.72M 74.50% 92.25% PVi G-Tiny[Neur IPS22][31] 9.46M 78.38% 94.38% Res Net-50[CVPR2016][34] 25.56M 76.55% 93.06% Swin-Tiny[ICCV2021][53] 28.29M 81.18% 95.61% CLUSTERFORMER-Tiny 27.85M 81.31% 96.32% Context Cluster-Small[ICLR23][58] 14.0M 77.42% 93.69% Dei T-Small[ICML21][80] 22.05M 80.69% 95.06% PVi G-Small[Neur IPS22][31] 29.02M 82.00% 95.97% Res Net-101[CVPR2016][34] 44.55M 77.97% 94.06% Swin-Small[ICCV2021][53] 49.61M 83.02% 96.29% CLUSTERFORMER-Small 48.71M 83.41% 97.13% Results on Image Net. Table 1 illustrates our compelling results over different famous methods. CLUSTERFORMER exceeds the Swin Transformer [53] by 0.13% and 0.39% on Tinybased and Small-based models with fewer parameters (i.e., 27.85M vs. 28.29M and 48.71M vs. 49.61M), respectively. On top-5 accuracy, our approach also outperforms the Swin-Tiny and Swin-Small with gains of 0.71% and 0.84%, respectively. In addition, our margins over the Res Net family [34] are 3.44% 4.76% on top1 accuracy with on-par parameters (i.e., 27.85M vs. 25.56M and 48.71M vs. 44.55M). 4.1.2 Experiments on Object Detection Training. We use mmdetection3 as codebase and follow the default training settings. For a fair comparison, we follow the training protocol in [17]: 1) the number of instances centers is set to 100; 2) a linear combination of the L1 loss and the GIo U Loss is used as the optimization objective for bounding box regression. Their coefficients are set to 5 and 2, respectively. In addition, the final object centers are fed into a small FFN for object classification, trained with a binary cross-entropy loss. Moreover, we set the initial learning rate to 1 10 5, the training epoch to 50, and the batch size to 16. We use random scale jittering with a factor in [0.1, 2.0] and a crop size of 1024 1024. Test. We use one input image scale with shorter side as 800. Metric. We adopt AP, AP50, AP75, APS, APM, and APL. Performance Comparison. In Table 2, we present the numerical results for CLUSTERFORMER for object detection. We observe that it surpasses all counterparts [70, 7, 56, 77, 9, 75, 60, 102, 71, 96] with remarkable gains with respect to m AP. In particular, CLUSTERFORMER-Tiny exceeds the vanilla Deformable DETR [102], Sparse-DETR [71], and DINO [96] over Swin-T [53] by 6.5%, 3.4%, and 0.8% in terms of m AP, respectively. In addition, our approach also outperforms these methods over Swin-S [53], i.e., 54.2% vs 48.3% vs 49.9% vs 53.3% in terms of m AP, respectively. Notably, CLUSTERFORMER achieves impressive performance without relying on additional augmentation. 4.1.3 Experiments on Semantic Segmentation Training. We use mmsegmentation4 as codebase and follow the default training settings. The training process for semantic segmentation involves setting the number of cluster centers to match the number of semantic categories, which is 150 for ADE20K [101]. Following the approach employed in recent works [97, 17, 74], we adopt a combination of the standard crossentropy loss and an auxiliary dice loss for the loss function. By default, the coefficients for the cross-entropy and dice losses are set to 5 and 1, respectively. In addition, we configure the initial learning rate to 1 10 5, the number of training epochs to 50, and the batch size to 16. 2https://github.com/open-mmlab/mmclassification 3https://github.com/open-mmlab/mmdetection 4https://github.com/open-mmlab/mmsegmentation Table 2: Quantitative results on COCO [49] test-dev for object detection (see 4.1.2 for details). Algorithm Backbone Epoch m AP AP50 AP75 APS APM APL Faster R-CNN[Neur IPS15][70] Res Net-101 36 41.7 62.3 45.7 24.7 46.0 53.2 Cascade R-CNN[CVPR18] [7] Res Net-101 36 42.8 61.1 46.7 24.9 46.5 56.4 Grid R-CNN[CVPR19][56] Res Net-50 24 40.4 58.5 43.6 22.7 43.9 53.0 Efficient Det[CVPR20][77] Efficient-B3 300 45.4 63.9 49.3 27.1 49.5 61.3 DETR[ECCV20][9] Res Net-50 150 39.9 60.4 41.7 17.6 43.4 59.4 Sparse R-CNN[CVPR21][75] Res Net-101 36 46.2 65.1 50.4 29.5 49.2 61.7 Conditional DETR[ICCV21][60] Res Net-50 50 41.1 61.9 43.5 20.4 44.5 59.9 Deformable DETR [ICLR21][102] Swin-T 50 45.5 0.26 65.2 0.20 49.8 0.21 27.0 0.26 49.1 0.24 60.7 0.29 Swin-S 48.3 0.21 68.7 0.27 52.1 0.27 30.5 0.28 51.6 0.22 64.4 0.19 Sparse-DETR[ICLR22][71] Swin-T 50 48.6 0.24 69.6 0.20 53.5 0.23 30.1 0.27 51.8 0.21 64.9 0.29 Swin-S 49.9 0.21 70.3 0.27 54.0 0.26 32.5 0.22 53.6 0.28 66.2 0.25 DINO [ICLR23][96] Swin-T 50 51.2 0.26 68.4 0.25 55.3 0.26 31.3 0.24 55.1 0.38 65.8 0.26 Swin-S 53.3 0.27 70.9 0.38 57.6 0.23 33.8 0.23 56.4 0.32 66.9 0.26 CLUSTERFORMER Ours-Tiny 50 52.0 0.32 70.4 0.25 57.5 0.32 34.2 0.28 54.8 0.29 64.8 0.22 Ours-Small 54.2 0.33 71.8 0.16 59.1 0.17 35.6 0.28 57.2 0.20 67.4 0.18 Table 3: Quantitative results on ADE20K [101] val for semantic segmentation (see 4.1.3 for details). Algorithm Backbone Epoch m Io U FCN[CVPR2015][54] Res Net-50 50 36.0 Deeplab V3+[ECCV2018][15] Res Net-50 50 42.7 APCNet[CVPR2019][32] Res Net-50 100 43.4 SETR[CVPR2021][100] Vi T-L 100 49.3 Segmenter[ICCV2021][74] Vi T-B 100 52.1 Segformer[Neur IPS2021][90] MIT-B5 100 51.4 k Ma X-Deeplab[ECCV2022][95] Conv Ne Xt-T 100 48.3 0.15 Conv Ne Xt-S 51.6 0.23 Mask2Former[CVPR2022][17] Swin-T 100 48.5 0.24 Swin-S 51.1 0.21 CLUSTERFORMER Ours-Tiny 100 49.1 0.19 Ours-Small 52.4 0.23 Furthermore, we employ random scale jittering, applying a factor within the range of [0.5, 2.0], and utilize a crop size with a fixed resolution of 640 640 pixels. Test. During the testing phase, we re-scale the input image with a shorter side to 640 pixels without applying any additional data augmentation at test time. Metric. Mean intersection-over-union (m Io U) is used for assessing image semantic segmentation performance. Performance Comparison. Table 3 shows the results on semantic segmentation. Empriailly, our method compares favorably to recent transformer-based approaches [54, 15, 32, 100, 74, 90, 95, 17]. For instance, CLUSTERFORMER-Tiny surpasses both recent advancements, i.e., k Ma X-Deeplab [95] and Mask2Former [17] with Swin-T [53] (i.e., 49.1% vs. 48.3% vs. 48.5%), respectively. Moreover, CLUSTERFORMERSmall achieves 52.4% m Io U and outperforms all other methods in terms of m Io U, making it competitive with state-of-the-art methods as well. 4.1.4 Experiments on Instance Segmentation Training. We adopt the same training strategy for instance segmentation by following 4.1.2. For instance segmentation, we change the training objective by utilizing a combination of the binary cross-entropy loss and the dice Loss for instance mask optimization. Test. We use one input image scale with a shorter side of 800. Metric. We adopt AP, AP50, AP75, APS, APM, and APL. Performance Comparison.Table 4 presents the results of CLUSTERFORMER against famous instance segmentation methods [33, 8, 14, 43, 13, 26, 23, 18, 17, 44] on COCO test-dev. CLUSTERFORMER shows clear performance advantages over prior arts. For example, CLUSTERFORMER-Tiny outperforms the universal counterparts Mask2Former [17] by 1.4% over Swin-T [53] in terms of m AP and on par with the state-of-the-art method, Mask-Dino [44] with Swin-T backbone. Moreover, CLUSTERFORMER-Small surpasses all the competitors, e.g., yielding significant gains of 1.0% and 0.5% m AP compared to Mask2Former and Mask-Dino over Swin-S, respectively. Without bells and whistles, our method establishes a new state-of-the-art on COCO instance segmentation. 4.1.5 Experiments on Panoptic Segmentation Training. Following the convention [84, 17], we use the following objective for network learning: LPanoptic = λth Lth + λst Lst + λaux Laux, (7) Lth and Lst represent the loss functions for things and stuff, respectively. To ensure a fair comparison, we follow [95, 85] and incorporate an auxiliary loss calculated as a weighted sum of four different loss terms, specifically, a PQ-style loss, a mask-ID cross-entropy loss, an instance discrimination Table 4: Quantitative results on COCO [49] test-dev for instance segmentation (see 4.1.4 for details). Algorithm Backbone Epoch m AP AP50 AP75 APS APM APL Mask R-CNN[ICCV2017][33] Res Net-101 12 36.1 57.5 38.6 18.8 39.7 49.5 Cascade MR-CNN[PAMI2019][8] Res Net-101 12 37.3 58.2 40.1 19.7 40.6 51.5 HTC[CVPR2019][14] Res Net-101 20 39.6 61.0 42.8 21.3 42.9 55.0 Point Rend[CVPR2020][43] Res Net-50 12 36.3 56.9 38.7 19.8 39.4 48.5 Blend Mask[CVPR2020][13] Res Net-101 36 38.4 60.7 41.3 18.2 41.5 53.3 Query Inst[ICCV2021][26] Res Net-101 36 41.0 63.3 44.5 21.7 44.4 60.7 SOLQ[Neur IPS2021][23] Swin-L 50 46.7 72.7 50.6 29.2 50.1 60.9 Sparse Inst[CVPR2022][18] Res Net-50 36 37.9 59.2 40.2 15.7 39.4 56.9 Mask2Former[CVPR2022][17] Swin-T 50 44.5 0.16 67.3 0.15 47.7 0.24 23.9 0.20 48.1 0.16 66.4 0.15 Swin-S 46.0 0.21 68.4 0.22 49.8 0.24 25.4 0.19 49.7 0.22 67.4 0.24 Mask-Dino[CVPR2023][44] Swin-T 50 45.8 0.28 69.6 0.29 50.2 0.26 26.0 0.28 48.7 0.37 66.4 0.29 Swin-S 46.5 0.39 70.1 0.34 52.2 0.28 27.6 0.34 49.9 0.25 69.5 0.29 CLUSTERFORMER Ours-Tiny 50 45.9 0.26 69.1 0.21 49.5 0.18 25.2 0.22 50.1 0.24 68.8 0.24 Ours-Small 47.0 0.19 71.5 0.26 51.8 0.24 27.3 0.16 50.5 0.20 72.6 0.22 Table 5: Quantitative results on COCO Panoptic [42] val for panoptic segmentation (see 4.1.5 for details). Algorithm Backbone Epoch PQ PQTh PQSt m APTh pan m Io Upan Panoptic-FPN[CVPR2019][41] Res Net-101 20 44.0 52.0 31.9 34.0 51.5 UPSNet[CVPR2019][92] Res Net-101 12 46.2 52.8 36.5 36.3 56.9 Panoptic-Deeplab[CVPR2020][16] Xception-71 12 41.2 44.9 35.7 31.5 55.4 Panoptic-FCN[CVPR2021][45] Res Net-50 12 44.3 50.0 35.6 35.5 55.0 Max-Deeplab[CVPR2021][85] Max-L 55 51.1 57.0 42.2 CMT-Deeplab[CVPR2022][94] Axial-R104 55 54.1 58.8 47.1 Panoptic Segformer[CVPR2022][46] Res Net-50 24 49.6 0.25 54.4 0.26 42.4 0.25 39.5 0.20 60.8 0.21 Res Net-101 50.6 0.21 55.5 0.24 43.2 0.20 40.4 0.21 62.0 0.22 Mask2Former[CVPR2022][17] Swin-T 50 53.2 0.25 59.1 0.22 43.3 0.23 42.3 0.27 62.9 0.19 Swin-S 54.1 0.29 60.2 0.28 45.6 0.18 43.1 0.23 63.6 0.31 Mask Dino[CVPR2023][44] Swin-T 50 53.6 0.29 59.5 0.26 44.0 0.24 44.3 0.29 63.2 0.27 Swin-S 54.9 0.33 61.1 0.23 46.2 0.26 45.0 0.22 64.3 0.30 CLUSTERFORMER Ours-Tiny 50 54.7 0.22 60.8 0.31 46.1 0.20 43.4 0.25 64.0 0.20 Ours-Small 55.8 0.38 61.9 0.39 47.2 0.23 44.2 0.22 65.5 0.21 loss, and a semantic segmentation loss. More information about Laux can be found in [85, 95]. The coefficients λth, λst, and λaux are assigned the values of 5, 3, and 1, respectively. Furthermore, the final centers are input into a small feed-forward neural network (FFN) for semantic classification, which is trained using a binary cross-entropy loss. Moreover, we set the initial learning rate to 1 10 5, the number of training epochs to 50, and the batch size to 16. We also employ random scale jittering with a factor range of [0.1, 2.0] and a crop size of 1024 1024. Test. We use one input image scale with a shorter side of 800. Metric. We employ the PQ metric [42] and report PQTh and PQSt for the thing" and stuff" classes, respectively. To ensure comprehensiveness, we also include m APTh pan, which evaluates mean average precision on "thing" classes using instance segmentation annotations, and m Io Upan, which calculates m Io U for semantic segmentation by merging instance masks belonging to the same category, using the same model trained for the panoptic segmentation task. Performance Comparison. We perform a comprehensive comparison against two divergent groups of state-of-the-art methods: universal approaches [46, 17, 44] and specialized panoptic methods [41, 92, 16, 45, 85, 97, 94]. As shown in Table 5, CLUSTERFORMER outperforms both types of rivals. For instance, the performance of CLUSTERFORMER-Tiny clear ahead compared to Mask2Former [17] (i.e., 54.7% PQ vs. 53.2% PQ) and Mask-Dino [44] (i.e., 54.7% PQ vs. 53.6% PQ) on the top of Swin-T [53], and CLUSTERFORMER-Small achieves promising gains of 1.7% and 0.9% PQ against Mask2Former and Mask-Dino over Swin-S, respectively. Moreover, in terms of m APTh pan and m Io Upan, the CLUSTERFORMER also achieves outstanding performance beyond counterpart approaches. 4.2 Ablative Study This section ablates CLUSTERFORMER s key components on Image Net [72] and MS COCO [49] validation split. All experiments use the tiny model. Key Component Analysis. We first investigate the two major elements of CLUSTERFORMER, specifically, Recurrent Cross-Attention Clustering for center updating and Feature Dispatching for feature updating. We construct a BASELINE model without any center updating and feature dispatching technique. As shown in Table 6a, BASELINE achieves 74.59% top-1 and 91.73% top5 accuracy. Upon applying Recurrent Cross-Attention Clustering to the BASELINE, we observe consistent and substantial improvements for both top-1 accuracy (74.59% 80.57%) and top-5 accuracy (91.73% 95.22%). This highlights the importance of the center updating strategy Table 6: A set of ablative studies on Image Net [72] validation and MS COCO [49] test-dev split (see 4.2). The adopted designs are marked in red. Algorithm Component #Params top-1 top-5 BASELINE 21.73M 74.59 91.73 + Recurrent Cross-Attention Clustering 26.27M 80.57 95.22 + Feature Dispatching 23.46M 78.58 94.68 CLUSTERFORMER (both) 27.85M 81.31 96.32 (a) Key Component Analysiss Numbers (T) #Params top-1 top-5 1 81.06 96.23 2 81.22 96.29 3 81.31 96.32 4 81.33 96.33 (b) Number of Recursion Variant Cluster Center Updating Strategy #Params top-1 top-5 Cosine Similarity 23.88M 78.79 94.36 Vanilla Cross-Attention [81] 35.48M 79.67 94.95 Criss Cross-Attention [35] 34.16M 79.91 95.24 K-Means [95] 27.71M 80.96 95.57 Recurrent Cross-Attention 27.85M 81.31 96.32 (c) Recurrent Cross-Attention Clustering Head Dimension #Params top-1 top-5 16 17.25M 71.69 90.16 24 22.88M 75.37 92.45 32 27.85M 81.31 96.32 40 32.81M 82.21 97.09 48 38.14M 82.40 97.22 (d) Head Dimension Feature Dispatching #Params top-1 top-5 None 26.27M 80.57 95.22 Vanilla FC Layer 27.14M 80.83 95.47 Confidence-Based [68] 26.81M 80.69 95.30 FC w/ Similarity [58] 27.46M 80.96 95.84 Ours (Eq. 6) 27.85M 81.31 96.32 (e) Feature Dispatching Decoder Query Initialization m AP AP50 AP75 Free Parameters 44.2 66.3 46.4 Direct Feature Embedding [17] 44.5 67.3 47.2 Mixed Query Selection [44] 44.9 67.9 47.8 Scene-Adoptive Embedding [47] 45.1 67.8 48.0 Centers from Encoder (Ours) 45.9 69.1 49.5 (f) Decoder Query Initialization for instance segmentation and validates the effectiveness of our approach, even without explicitly performing clustering. Furthermore, after incorporating Feature Dispatching into the BASELINE, we achieve significant gains of 3.99% in top-1 accuracy and 2.95% in top-5 accuracy. Finally, by integrating both core techniques, CLUSTERFORMER delivers the best performance across both metrics. This indicates that the proposed Recurrent Cross-Attention Clustering and Feature Dispatching can work synergistically and validates the effectiveness of our comprehensive algorithmic design. Recurrent Cross-attention Clustering. We next study the impact of our Recurrent Cross-attention Clustering (Eq.4) by contrasting it with the cosine similarity updating, basic cross-attention [81], Criss-attention [35] and K-Means cross-attention [95]. As illustrated in Table 6c, our Recurrent Cross-Attention proves to be effective it outperforms the cosine similarity, vanilla, Criss and KMeans by 2.52%, 1.64%, 1.40% and 0.15% top-1 accuracy respectively, and efficient its #Params are significantly less than the other vanilla and Criss-attention and on par with K-Means, in line with our analysis in 3.2. To gain further insights into recursive clustering, we examine the effect of the recursion number T in Table 6b. We discover that performance progressively improves from 81.06% to 81.31% in top-1 accuracy when increasing T from 1 to 3, but remains constant after running additional iterations. We also observe that #Params increase as T increases. Consequently, we set T = 3 as the default to strike an optimal balance between accuracy and computation cost. Multi-head Dimension. We then ablate the head embedding dimension for the attention head in Table 6d. We find that performance significantly improves from 71.69% to 82.40% in top-1 accuracy when increasing the dimension from 16 to 48, but #Params steadily increase as the dimension grows. For a fair comparison with Swin [53], we set the head dimension to 32 as our default. Feature Dispatching. We further analyze the influence of our Feature Dispatching. As outlined in Table 6e, in a standard manner without any dispatching method, the model attains 80.57% top-1 accuracy and 95.22% top-5 accuracy. By applying a vanilla fully connected layer to update the feature, we witness a marginal increase of 0.26% in top-1 accuracy. Moreover, using the confidence-based updating method [68] and fully connected layer with similarity, the model demonstrates a noticeable enhancement in 0.12% and 0.39% top-1 accuracy, respectively. Last, our method yields significant performance advancements across both metrics, i.e., 81.31% top-1 and 96.32% top-5 accuracy. Decoder Query Initialization. Last, we examine the impact of query initialization in the decoder on a downstream task (i.e., instance segmentation) in Table 6f. For free parameter initialization, the base model can achieve 44.2% in terms of m AP. By applying direct feature embedding, the method has a slight improvement of 0.3% m AP. In addition, the model exhibits improvements in Figure 3: Visualization of center-feature assignment at the last stage of recurrent cross-attention clustering with the resolution of 7 by 7. The map displays distinct clusters, each containing features with similar representations. m AP, achieving 44.9% and 45.1%, respectively, by employing the mixed query selection [44] and scene-adoptive embedding [47]. Outstandingly, CLUSTERFORMER achieves the highest performance in all three metrices, i.e., 45.9% m AP, 69.1% AP50 and 49.5% AP75, respectively. The empirical evidence proves our design using the cluster centers from the encoder to derive the initial query for the decoder that facilitates the transferability for representation learning. Ad-hoc Explainability. We visualize the cluster assignment map for image classification in Fig. 3. This figure provides an insightful illustration of how CLUSTERFORMER groups similar features together. Each color represents a cluster of features that share common characteristics. 5 Conclusion This study adopts an epistemological perspective centered on the clustering-based paradigm, which advocates a universal vision framework named CLUSTERFORMER. This framework aims to address diverse visual tasks with varying degrees of clustering granularity. By leveraging insights from clustering, we customize the cross-attention mechanism for recursive clustering and introduce a novel method for feature dispatching. Empirical findings provide substantial evidence to support the effectiveness of this systematic approach. Based on its efficacy, we argue deductively that the proposed universal solution will have a substantial impact on the wider range of visual tasks when viewed through the lens of clustering. This question remains open for our future endeavors. Acknowledgement. This research was supported by the National Science Foundation under Grant No. 2242243. [1] Sameer Agarwal, Jongwoo Lim, Lihi Zelnik-Manor, Pietro Perona, David Kriegman, and Serge Belongie. Beyond pairwise clustering. In CVPR, 2005. [2] Merav Ahissar and Shaul Hochstein. The reverse hierarchy theory of visual perceptual learning. Trends in cognitive sciences, 8(10):457 464, 2004. [3] Valerie Ahl and Timothy FH Allen. Hierarchy theory: a vision, vocabulary, and epistemology. Columbia University Press, 1996. [4] Rishi Bommasani, Drew A Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, et al. On the opportunities and risks of foundation models. ar Xiv preprint ar Xiv:2108.07258, 2021. [5] Gijs Joost Brouwer and David J Heeger. Categorical clustering of the neural representation of color. Journal of Neuroscience, 33(39):15454 15465, 2013. [6] Xiao Cai, Feiping Nie, Heng Huang, and Farhad Kamangar. Heterogeneous image feature integration via multi-modal spectral clustering. In CVPR, 2011. [7] Zhaowei Cai and Nuno Vasconcelos. Cascade r-cnn: Delving into high quality object detection. In CVPR, 2018. [8] Zhaowei Cai and Nuno Vasconcelos. Cascade r-cnn: high quality object detection and instance segmentation. IEEE TPAMI, 43(5):1483 1498, 2019. [9] Nicolas Carion, Francisco Massa, Gabriel Synnaeve, Nicolas Usunier, Alexander Kirillov, and Sergey Zagoruyko. End-to-end object detection with transformers. In ECCV, 2020. [10] Mathilde Caron, Piotr Bojanowski, Armand Joulin, and Matthijs Douze. Deep clustering for unsupervised learning of visual features. In ECCV, 2018. [11] M Emre Celebi. Partitional clustering algorithms. Springer, 2014. [12] Antoni B Chan and Nuno Vasconcelos. Modeling, clustering, and segmenting video with mixtures of dynamic textures. IEEE TPAMI, 30(5):909 926, 2008. [13] Hao Chen, Kunyang Sun, Zhi Tian, Chunhua Shen, Yongming Huang, and Youliang Yan. Blendmask: Top-down meets bottom-up for instance segmentation. In CVPR, 2020. [14] Kai Chen, Jiangmiao Pang, Jiaqi Wang, Yu Xiong, Xiaoxiao Li, Shuyang Sun, Wansen Feng, Ziwei Liu, Jianping Shi, Wanli Ouyang, et al. Hybrid task cascade for instance segmentation. In CVPR, 2019. [15] Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig Adam. Encoder-decoder with atrous separable convolution for semantic image segmentation. In ECCV, 2018. [16] Bowen Cheng, Maxwell D Collins, Yukun Zhu, Ting Liu, Thomas S Huang, Hartwig Adam, and Liang-Chieh Chen. Panoptic-deeplab: A simple, strong, and fast baseline for bottom-up panoptic segmentation. In CVPR, 2020. [17] Bowen Cheng, Ishan Misra, Alexander G Schwing, Alexander Kirillov, and Rohit Girdhar. Masked-attention mask transformer for universal image segmentation. In CVPR, 2022. [18] Tianheng Cheng, Xinggang Wang, Shaoyu Chen, Wenqiang Zhang, Qian Zhang, Chang Huang, Zhaoxiang Zhang, and Wenyu Liu. Sparse instance activation for real-time instance segmentation. In CVPR, 2022. [19] Yiming Cui. Feature aggregated queries for transformer-based video object detectors. In CVPR, 2023. [20] Yiming Cui, Liqi Yan, Zhiwen Cao, and Dongfang Liu. Tf-blender: Temporal feature blender for video object detection. In ICCV, 2021. [21] Yiming Cui, Linjie Yang, and Haichao Yu. Learning dynamic query combinations for transformer-based object detection and segmentation. ICML, 2023. [22] Mostafa Dehghani, Josip Djolonga, Basil Mustafa, Piotr Padlewski, Jonathan Heek, Justin Gilmer, Andreas Steiner, Mathilde Caron, Robert Geirhos, Ibrahim Alabdulmohsin, et al. Scaling vision transformers to 22 billion parameters. ar Xiv preprint ar Xiv:2302.05442, 2023. [23] Bin Dong, Fangao Zeng, Tiancai Wang, Xiangyu Zhang, and Yichen Wei. Solq: Segmenting objects by learning queries. In Neur IPS, 2021. [24] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ICLR, 2021. [25] Jiemin Fang, Lingxi Xie, Xinggang Wang, Xiaopeng Zhang, Wenyu Liu, and Qi Tian. Msgtransformer: Exchanging local spatial information by manipulating messenger tokens. In CVPR, 2022. [26] Yuxin Fang, Shusheng Yang, Xinggang Wang, Yu Li, Chen Fang, Ying Shan, Bin Feng, and Wenyu Liu. Instances as queries. In ICCV, 2021. [27] David J Field, Anthony Hayes, and Robert F Hess. Contour integration by the human visual system: evidence for a local association field . Vision research, 33(2):173 193, 1993. [28] Hichem Frigui and Raghu Krishnapuram. A robust competitive clustering algorithm with applications in computer vision. IEEE TPAMI, 21(5):450 465, 1999. [29] Yoram Gdalyahu, Daphna Weinshall, and Michael Werman. Self-organization in vision: stochastic clustering for image segmentation, perceptual grouping, and image database organization. IEEE TPAMI, 23(10):1053 1074, 2001. [30] Anmol Gulati, James Qin, Chung-Cheng Chiu, Niki Parmar, Yu Zhang, Jiahui Yu, Wei Han, Shibo Wang, Zhengdong Zhang, Yonghui Wu, et al. Conformer: Convolution-augmented transformer for speech recognition. ar Xiv preprint ar Xiv:2005.08100, 2020. [31] Kai Han, Yunhe Wang, Jianyuan Guo, Yehui Tang, and Enhua Wu. Vision gnn: An image is worth graph of nodes. In Neur IPS, 2022. [32] Junjun He, Zhongying Deng, Lei Zhou, Yali Wang, and Yu Qiao. Adaptive pyramid context network for semantic segmentation. In CVPR, 2019. [33] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In ICCV, 2017. [34] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016. [35] Zilong Huang, Xinggang Wang, Lichao Huang, Chang Huang, Yunchao Wei, and Wenyu Liu. Ccnet: Criss-cross attention for semantic segmentation. In ICCV, 2019. [36] Anil K Jain and Richard C Dubes. Algorithms for clustering data. Prentice-Hall, Inc., 1988. [37] Varun Jampani, Deqing Sun, Ming-Yu Liu, Ming-Hsuan Yang, and Jan Kautz. Superpixel sampling networks. In ECCV, 2018. [38] Stephen C Johnson. Hierarchical clustering schemes. Psychometrika, 32(3):241 254, 1967. [39] Jean-Michel Jolion, Peter Meer, and Samira Bataouche. Robust clustering with applications in computer vision. IEEE TPAMI, 13(8):791 802, 1991. [40] Bela Julesz. A brief outline of the texton theory of human vision. Trends in Neurosciences, 7(2):41 45, 1984. [41] Alexander Kirillov, Ross Girshick, Kaiming He, and Piotr Dollár. Panoptic feature pyramid networks. In CVPR, 2019. [42] Alexander Kirillov, Kaiming He, Ross Girshick, Carsten Rother, and Piotr Dollár. Panoptic segmentation. In CVPR, 2019. [43] Alexander Kirillov, Yuxin Wu, Kaiming He, and Ross Girshick. Pointrend: Image segmentation as rendering. In CVPR, 2020. [44] Feng Li, Hao Zhang, Shilong Liu, Lei Zhang, Lionel M Ni, Heung-Yeung Shum, et al. Mask dino: Towards a unified transformer-based framework for object detection and segmentation. CVPR, 2023. [45] Yanwei Li, Hengshuang Zhao, Xiaojuan Qi, Liwei Wang, Zeming Li, Jian Sun, and Jiaya Jia. Fully convolutional networks for panoptic segmentation. In CVPR, 2021. [46] Zhiqi Li, Wenhai Wang, Enze Xie, Zhiding Yu, Anima Anandkumar, Jose M Alvarez, Ping Luo, and Tong Lu. Panoptic segformer: Delving deeper into panoptic segmentation with transformers. In CVPR, 2022. [47] James Liang, Tianfei Zhou, and Dongfang Liu. Clustseg: Clustering for universal segmentation. In ICML, 2023. [48] Weicong Liang, Yuhui Yuan, Henghui Ding, Xiao Luo, Weihong Lin, Ding Jia, Zheng Zhang, Chao Zhang, and Han Hu. Expediting large-scale vision transformer for dense prediction without fine-tuning. Neur IPS, 2022. [49] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, 2014. [50] Dongfang Liu, Yiming Cui, Wenbo Tan, and Yingjie Chen. Sg-net: Spatial granularity network for one-stage video instance segmentation. In CVPR, 2021. [51] Dongfang Liu, Yiming Cui, Liqi Yan, Christos Mousas, Baijian Yang, and Yingjie Chen. Densernet: Weakly supervised visual localization using multi-scale feature aggregation. In AAAI, 2021. [52] Dongfang Liu, James Liang, Tony Geng, Alexander Loui, and Tianfei Zhou. Tripartite feature enhanced pyramid network for dense prediction. IEEE TIP, 2023. [53] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In ICCV, 2021. [54] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In CVPR, 2015. [55] Le Lu and René Vidal. Combined central and subspace clustering for computer vision applications. In ICML, 2006. [56] Xin Lu, Buyu Li, Yuxin Yue, Quanquan Li, and Junjie Yan. Grid r-cnn. In CVPR, 2019. [57] Yawen Lu, Qifan Wang, Siqi Ma, Tong Geng, Yingjie Victor Chen, Huaijin Chen, and Dongfang Liu. Transflow: Transformer as flow learner. CVPR, 2023. [58] Xu Ma, Yuqian Zhou, Huan Wang, Can Qin, Bin Sun, Chang Liu, and Yun Fu. Image as set of points. In ICLR, 2023. [59] Celeste Mc Collough. Color adaptation of edge-detectors in the human visual system. Science, 149(3688):1115 1116, 1965. [60] Depu Meng, Xiaokang Chen, Zejia Fan, Gang Zeng, Houqiang Li, Yuhui Yuan, Lei Sun, and Jingdong Wang. Conditional detr for fast training convergence. In ICCV, 2021. [61] Marius Muja and David G Lowe. Scalable nearest neighbor algorithms for high dimensional data. IEEE TPAMI, 36(11):2227 2240, 2014. [62] Fionn Murtagh and Pedro Contreras. Algorithms for hierarchical clustering: an overview. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 2(1):86 97, 2012. [63] Satyasai Jagannath Nanda and Ganapati Panda. A survey on nature inspired metaheuristic algorithms for partitional clustering. Swarm and Evolutionary computation, 16:1 18, 2014. [64] Frank Nielsen and Frank Nielsen. Hierarchical clustering. Introduction to HPC with MPI for Data Science, pages 195 211, 2016. [65] Haluk Ö gmen, Thomas U Otto, and Michael H Herzog. Perceptual grouping induces nonretinotopic feature attribution in human vision. Vision Research, 46(19):3234 3242, 2006. [66] C Alejandro Parraga, Tom Troscianko, and David J Tolhurst. The human visual system is optimised for processing the spatial information in natural visual images. Current Biology, 10(1):35 38, 2000. [67] Yury Petrov, Matteo Carandini, and Suzanne Mc Kee. Two distinct mechanisms of suppression in human vision. Journal of Neuroscience, 25(38):8704 8707, 2005. [68] Yulei Qin, Juan Wen, Hao Zheng, Xiaolin Huang, Jie Yang, Ning Song, Yue-Min Zhu, Lingqian Wu, and Guang-Zhong Yang. Varifocal-net: A chromosome classification approach using deep convolutional networks. IEEE transactions on medical imaging, 38(11):2569 2581, 2019. [69] Chandan K Reddy and Bhanukiran Vinzamuri. A survey of partitional and hierarchical clustering algorithms. In Data clustering, pages 87 110. Chapman and Hall/CRC, 2018. [70] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In Neur IPS, 2015. [71] Byungseok Roh, Jae Woong Shin, Wuhyun Shin, and Saehoon Kim. Sparse detr: Efficient end-to-end object detection with learnable sparsity. ICLR, 2022. [72] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael S. Bernstein, Alexander C. Berg, and Fei-Fei Li. Imagenet large scale visual recognition challenge. IJCV, 115(3):211 252, 2015. [73] Dov Sagi. Perceptual learning in vision research. Vision research, 51(13):1552 1566, 2011. [74] Robin Strudel, Ricardo Garcia, Ivan Laptev, and Cordelia Schmid. Segmenter: Transformer for semantic segmentation. In ICCV, 2021. [75] Peize Sun, Rufeng Zhang, Yi Jiang, Tao Kong, Chenfeng Xu, Wei Zhan, Masayoshi Tomizuka, Lei Li, Zehuan Yuan, Changhu Wang, et al. Sparse r-cnn: End-to-end object detection with learnable proposals. In CVPR, 2021. [76] Teppei Suzuki. Clustering as attention: Unified image segmentation with hierarchical clustering. ar Xiv preprint ar Xiv:2205.09949, 2022. [77] Mingxing Tan, Ruoming Pang, and Quoc V Le. Efficientdet: Scalable and efficient object detection. In CVPR, 2020. [78] Yuliya Tarabalka, Jón Atli Benediktsson, and Jocelyn Chanussot. Spectral spatial classification of hyperspectral imagery based on partitional clustering techniques. IEEE transactions on geoscience and remote sensing, 47(8):2973 2987, 2009. [79] Simon Thorpe, Denis Fize, and Catherine Marlot. Speed of processing in the human visual system. nature, 381(6582):520 522, 1996. [80] Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herve Jegou. Training data-efficient image transformers and distillation through attention. In ICML, 2021. [81] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Neur IPS, 2017. [82] Mai-Anh T Vu, Tülay Adalı, Demba Ba, György Buzsáki, David Carlson, Katherine Heller, Conor Liston, Cynthia Rudin, Vikaas S Sohal, Alik S Widge, et al. A shared vision for machine learning in neuroscience. Journal of Neuroscience, 38(7):1601 1607, 2018. [83] George Wald. Human vision and the spectrum. Science, 101(2635):653 658, 1945. [84] Huiyu Wang, Yukun Zhu, Hartwig Adam, Alan Yuille, and Liang-Chieh Chen. Ma X-Deep Lab: End-to-end panoptic segmentation with mask transformers. In CVPR, 2021. [85] Huiyu Wang, Yukun Zhu, Hartwig Adam, Alan Yuille, and Liang-Chieh Chen. Max-deeplab: End-to-end panoptic segmentation with mask transformers. In CVPR, 2021. [86] Wenguan Wang, Cheng Han, Tianfei Zhou, and Dongfang Liu. Visual recognition with deep nearest centroids. ICLR, 2023. [87] Wenguan Wang, James Liang, and Dongfang Liu. Learning equivariant segmentation with instance-unique querying. Neur IPS, 2022. [88] Wenhai Wang, Enze Xie, Xiang Li, Deng-Ping Fan, Kaitao Song, Ding Liang, Tong Lu, Ping Luo, and Ling Shao. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In ICCV, 2021. [89] Hugh R Wilson. Computational evidence for a rivalry hierarchy in vision. Proceedings of the National Academy of Sciences, 100(24):14499 14503, 2003. [90] Enze Xie, Wenhai Wang, Zhiding Yu, Anima Anandkumar, Jose M Alvarez, and Ping Luo. Segformer: Simple and efficient design for semantic segmentation with transformers. Neur IPS, 2021. [91] Xuanli Lisa Xie and Gerardo Beni. A validity measure for fuzzy clustering. IEEE TPAMI, 13(08):841 847, 1991. [92] Yuwen Xiong, Renjie Liao, Hengshuang Zhao, Rui Hu, Min Bai, Ersin Yumer, and Raquel Urtasun. Upsnet: A unified panoptic segmentation network. In CVPR, 2019. [93] Jiarui Xu, Shalini De Mello, Sifei Liu, Wonmin Byeon, Thomas Breuel, Jan Kautz, and Xiaolong Wang. Groupvit: Semantic segmentation emerges from text supervision. In CVPR, 2022. [94] Qihang Yu, Huiyu Wang, Dahun Kim, Siyuan Qiao, Maxwell Collins, Yukun Zhu, Hartwig Adam, Alan Yuille, and Liang-Chieh Chen. Cmt-deeplab: Clustering mask transformers for panoptic segmentation. In CVPR, 2022. [95] Qihang Yu, Huiyu Wang, Siyuan Qiao, Maxwell Collins, Yukun Zhu, Hatwig Adam, Alan Yuille, and Liang-Chieh Chen. k-means mask transformer. ECCV, 2022. [96] Hao Zhang, Feng Li, Shilong Liu, Lei Zhang, Hang Su, Jun Zhu, Lionel M Ni, and Heung Yeung Shum. Dino: Detr with improved denoising anchor boxes for end-to-end object detection. ar Xiv preprint ar Xiv:2203.03605, 2022. [97] Wenwei Zhang, Jiangmiao Pang, Kai Chen, and Chen Change Loy. K-net: Towards unified image segmentation. Neur IPS, 2021. [98] Ying Zhao, George Karypis, and Usama Fayyad. Hierarchical clustering algorithms for document datasets. Data mining and knowledge discovery, 10:141 168, 2005. [99] Minghang Zheng, Peng Gao, Renrui Zhang, Kunchang Li, Xiaogang Wang, Hongsheng Li, and Hao Dong. End-to-end object detection with adaptive clustering transformer. BMVC, 2021. [100] Sixiao Zheng, Jiachen Lu, Hengshuang Zhao, Xiatian Zhu, Zekun Luo, Yabiao Wang, Yanwei Fu, Jianfeng Feng, Tao Xiang, Philip HS Torr, et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In CVPR, 2021. [101] Bolei Zhou, Hang Zhao, Xavier Puig, Sanja Fidler, Adela Barriuso, and Antonio Torralba. Scene parsing through ade20k dataset. In CVPR, 2017. [102] Xizhou Zhu, Weijie Su, Lewei Lu, Bin Li, Xiaogang Wang, and Jifeng Dai. Deformable detr: Deformable transformers for end-to-end object detection. In ICLR, 2021.