# dynamic_clustering_convolutional_neural_network__2dc06025.pdf Dynamic Clustering Convolutional Neural Network Tanzhe Li1,2, Baochang Zhang3,4, Jiayi Lyu5, Xiawu Zheng1,2, Guodong Guo6, Taisong Jin1,2,7 1Key Laboratory of Multimedia Trusted Perception and Effcient Computing, Ministry of Education of China, Xiamen University, China. 2School of Informatics, Xiamen University, China. 3Hangzhou Research Institute, School of Artificial Intelligence, Beihang University, China. 4 Nanchang Institute of Technology, China. 5 School of Engineering Science, University of Chinese Academy of Sciences, China. 6 Ningbo Institute of Digital Twin, Eastern Institute of Technology, Ningbo, China. 7 Key Laboratory of Oracle Bone Inscriptions Information Processing, Ministry of Education of China, Anyang Normal University, China. litanzhe@stu.xmu.edu.cn,{jintaisong,zhengxiawu}@xmu.edu.cn, bczhang@buaa.edu.cn Convolutional neural networks (CNNs) have been playing a dominant role in computer vision. However, the existing approaches of using local window modeling in popular CNNs lack flexibility and hinder their ability to capture long-range dependencies of objects in an image. To overcome these limitations, we propose a novel CNN architecture, termed Dynamic Clustering Convolutional Neural Network (DCCNe Xt). The proposed DCCNe Xt takes a unique approach by employing global clustering to group image patches with similar semantics into clusters that are then convolved using the shared convolution kernels. To address the high computational complexity of global clustering, the feature vectors from each patch s subspace are extracted for efficient clustering, which makes the proposed model widely compatible with the downstream vision tasks. The extensive experiments of image classification, object detection, instance segmentation, and semantic segmentation on the benchmark datasets demonstrate that the proposed DCCNe Xt outperforms the mainstream Convolutional Neural Networks (CNNs), Vision Transformers (Vi Ts), Vision Multi-layer Perceptrons (MLPs), Vision Graph Neural Networks (GNNs), and Vision Mambas (Vi Ms). We anticipate that this study will provide a new perspective and a promising avenue for the design of convolutional neural networks. Code https://github.com/ltzovo/DCCNe Xt Introduction Convolutional neural networks (CNNs) (Le Cun et al. 1998) have been proposed for handwritten digit recognition in 1998. However, due to limitations in computational power and available data, it was not until Alex Net (Krizhevsky, Sutskever, and Hinton 2012) won the 2012 Image Net challenge that CNNs gained widespread attention and sparked the deep learning revolution. For a considerable period of time, CNNs (Simonyan and Zisserman 2014; He et al. 2016; Huang et al. 2017; Tan and Le 2019; Howard et al. 2017) architectures commonly utilize 3 3 convolution kernels as The corresponding author. Copyright 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. main building blocks due to the superior performance over larger convolution kernels with equivalent parameters and computational cost. Recently, Vision Transformers (Dosovitskiy et al. 2021; Touvron et al. 2021a; Wang et al. 2021; Liu et al. 2021; Touvron et al. 2021b) have achieved remarkable success in computer vision by leveraging global or large local receptive field of attention (Vaswani et al. 2017), which has resulted in a period where CNNs have been overshadowed by the success of Vision Transformers. It is time for CNNs to fight back. Inspired by the success of Vision Transformers, some recent studies (Trockman and Kolter 2022; Liu et al. 2022; Woo et al. 2023; Ding et al. 2022; Liu et al. 2023) have revisited the use of large convolution kernels, such as the famous Conv Ne Xt (Liu et al. 2022) , which utilizes 7 7 deep separable convolution (Howard et al. 2017; Chollet 2017) to increase receptive field and incorporates a series of improvements that enable CNNs to achieve competitive performance against Vi Ts. More recently, some studies (Smith et al. 2023; Woo et al. 2023) have demonstrated that CNNs achieve performance comparable to or even better than Vision Transformers in largescale supervised pre-training and unsupervised pre-training vision tasks. This challenges researchers previous perception that CNNs are not as effective as Vision Transformers in handling large-scale data. Although the modern CNNs have greatly improved by leveraging larger receptive fields of 7 7 convolution kernels , Conv Ne Xt (Liu et al. 2022) shows that larger 9 9 and 11 11 convolution kernels can result in performance saturation and decline, which indicates that convolution kernels that are too dense are not conducive to feature extraction. Despite the subsequent Rep LKNet (Ding et al. 2022) and SLa K (Liu et al. 2023) expanding the kernel to an astonishing 31 31 and 51 51 through reparameterization (Ding et al. 2021) and sparse decomposition, this super-sized kernel only brings a minor performance enhancement while the computational cost is difficult to accept. Thus, the 7 7 convolution kernel is still the norm for CNN-based backbones. However, the basic modeling scheme of 7 7 convolution still operates within a fixed-size local window, which limits the global receptive field of CNNs and is not flexible for cap- The Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI-25) turing the long-range dependencies of objects in an image. To efficiently address the aforementioned limitations of the standard convolution, we propose a novel global modeling tool, termed Dynamic Clustering Convolution (DCConv). As shown in Figure 1, for the H W image patches, we treat each patch as a clustering center, after calculating the ℓ2-norm distance between each clustering center and other patches, the K 1 patches with the shortest distance is chosen to join the cluster, resulting in H W clusters of size K. Furthermore, the convolution operation is performed on each cluster using a shared set of convolution kernels, resulting in the output of new feature maps. Considering that the clustering operation is based on the entire image and computation of the ℓ2-norm distance has the quadratic computational complexity, the proposed DCConv extracts a subset of feature vectors for patch clustering, which significantly reduces the computational cost for high-resolution downstream vision tasks. Our proposed dynamic clustering convolution differs greatly from standard convolution: (1) Standard convolution operates on locally adjacent image regions, which makes it difficult to capture long-range dependencies of the objects. In contrast, the proposed dynamic clustering convolution performs clustering globally, providing a larger global receptive field that can capture dependencies among objects that are farther apart. (2) The proposed convolution operation is sparsely distributed over globally important image patches, which has good sparsity property. (3) Cluster is a widespread data structure, which makes that our proposed DCConv not constrained by the structure or modality of the data and can flexibly perform global modeling like Transformer (Vaswani et al. 2017). This offers many possibilities for the adaptation of convolutional neural networks to the complex learning tasks in real-world applications. Based on the proposed DCConv, we develop a series of vision backbones, namely DCCNe Xt. To validate the efficacy of our method, we conduct extensive experiments on benchmark datasets to evaluate the efficacy of the proposed DCCNe Xt in various vision tasks, including image classification, object detection, instance segmentation, and semantic segmentation. The experimental results demonstrate the effectiveness of the proposed method. For instance, DCCNe Xt B2 achieves a top-1 accuracy of 82.8% on the Image Net classification task, outperforming the representative CNNs (Conv Ne Xt V2-T (Woo et al. 2023) 82.5%) with fewer parameters and FLOPs, in downstream tasks such as object detection, instance segmentation, and semantic segmentation, which rely on global modeling, our DCCNe Xt-B2 demonstrated improvements of 2.1 APb, 1.8 APm, and 0.8 m IOU, respectively, compared to Conv Ne Xt V2-T. Meanwhile, our model consistently outperforms various types of backbones (Vision Transformers, Vision MLPs, Vision GNNs, and Vision Mambas) in a series of vision tasks. Related Work Convolutional Neural Networks After Alex Net (Krizhevsky, Sutskever, and Hinton 2012) won the Image Net image classification challenge in 2012, the potential of CNNs was the first widely acknowledged by researchers. Since then, different extensions and variants of CNNs have been proposed to employ various vision tasks. For instance, VGG (Simonyan and Zisserman 2014) proposed repeatedly using simple building blocks, which provides the basic rules to guide the subsequent architecture design. Google Net (Szegedy et al. 2015) and Res Ne Xt (Xie et al. 2017) demonstrate the efficacy of multi-branching in the building blocks. Res Net (He et al. 2016) is a significant advance in CNN architectures by utilizing residual connections to address the challenge of deep neural networks. Furthermore, CNNs are designed for lighter and faster architectures, including but not limited to Mobile Net (Howard et al. 2017), Shuffle Net (Zhang et al. 2018), and Efficient Net (Tan and Le 2019). Recently, Conv Ne Xt (Liu et al. 2022; Woo et al. 2023) has shown impressive performance and competitive results with Vision Transformers, Rep LKNet (Ding et al. 2022) and SLa K (Liu et al. 2023) have enlarged the convolutional neural network s kernel to an astonishing 31 31 and 51 51. The aforementioned studies demonstrate the importance of CNNs as a vision architecture for the vision community. In contrast to these works, our objective is not to use larger convolution kernels, but rather to sparsely distribute convolution kernels among globally important patches. Thus, we employ global clustering to dynamically represent the image as clusters, followed by performing convolution on these clusters to capture long-range dependencies. Vision Transformers With the impressive achievements of Transformer (Vaswani et al. 2017) in the domain of natural language processing (NLP), the computer vision community seeks to harness the capabilities of this potential model. However, the dissimilarities between the structure of images and sentences result in the incorporation of only the attention mechanism module from the Transformer into the convolutional neural network (Hu, Shen, and Sun 2018; Woo et al. 2018; Wang et al. 2020; Hou, Zhou, and Feng 2021). The subsequent emergence of Vi T (Dosovitskiy et al. 2021) has demonstrated that Transformer (Vaswani et al. 2017) architectures can be highly effective in computer vision tasks. Despite its successes, Vi T (Dosovitskiy et al. 2021) suffers from the following limitations including the absence of inductive bias, dependence on large quantities of data, and difficulties in applying it to high-resolution images due to the computational complexity of self-attention (Vaswani et al. 2017). To address the aforementioned limitations of Vi T (Dosovitskiy et al. 2021), Dei T (Touvron et al. 2021a) leverages knowledge distillation (Hinton, Vinyals, and Dean 2015) and multiple training strategies to enable effective training on Image Net-1K (Russakovsky et al. 2015), Swin Transformer (Liu et al. 2021) and PVT (Wang et al. 2021) utilize the pyramid architectures and reduce the computational complexity of self-attention to tackle the challenges with high-resolution images. Furthermore, some research works like Poolformer (Yu et al. 2022c,d) and MLP-Mixer (Tolstikhin et al. 2021) adopt a different approach by replacing the self-attention in Transformer with a pooling layer Index Select Convolution Convolutional Kernel Image Patches Cluster 1 Cluster N Cluster 2 Cluster N Cluster 3 Cluster 1 Cluster 2 Figure 1: Illustration of dynamic clustering convolution. H: height of the feature map, W: width of the feature map, C: feature dimension, d: Downsampling interval, K: convolution kernel size of dynamic clustering convolution, Flatten: flattening the feature vectors along the spatial dimensions. Downsamp: extracting a set of sub-vectors for each patch. Topk Index: obtaining the index of clusters using the Top-K algorithm. Index Select: selecting feature vectors using the index of clusters. or MLP layer can still achieve remarkable results. These research efforts indicate that the success of Transformer in computer vision is attributed to its architecture, rather than the self-attention. Clustering in Computer Vision Clustering algorithms (Jampani et al. 2018; Huang and Li 2020) have been widely applied in computer vision, especially in image segmentation (Ren and Malik 2003; Li and Chen 2015; Yu et al. 2022b; Ma et al. 2022; Yu et al. 2022a), but their utilization in general vision backbones is quite constrained. Recently, Context-Cluster (Ma et al. 2023) introduced clustering algorithm into vision backbones, achieving remarkable results. However, due to computational complexity, Context-Cluster restricts clustering regions to local windows, similar to the Swin Transformer (Liu et al. 2021), hindering the model s ability to capture global dependencies. Furthermore, each image patch can only be assigned to one cluster center, which results in multiple image patches within the same cluster receiving the same semantic information, thereby limiting the model s expressive capacity. Compared to Context-Cluster, our clustering approach is more aggressive, where each patch is treated as a clustering center, and each patch can be assigned to multiple cluster centers, enabling each image patch to obtain distinct semantic information. At the same time, our proposed subvector downsampling allows clustering to operate globally, greatly enhancing the model s ability to capture long-range dependencies. The aforementioned innovations enable our DCCNe Xt to achieve significantly better performance in image classification compared to Context-Cluster. Methodology Network Architecture The network architecture of the proposed method is illustrated in Figure 2. Taking the DCCNe Xt-B1 version as an example, it first divides the input RGB image into patches using a 7 7 convolution with a stride of 4. Each patch is considered as a cluster center and has a feature dimension of 48. Our model consists of four stages in total. In the first two stages, the 7 7 depthwise separable convolution (Howard et al. 2017; Chollet 2017) is used to extract local features, while the latter two stages utilize dynamic clustering convolution to extract crucial global features. As the stages progress, adjacent patches are merged into a new patch using the downsamp layer, reducing the feature map to one-fourth of its original size and increasing the feature dimension of the patches. Following the design principles of previous pyramid structures, we mainly stack the blocks on the third stage, with each block consisting of a 7 7 depthwise separable convolution or dynamic clustering convolution and an FFN, respectively used for feature extraction and transformation. More details of our DCCNe Xt series are shown in Table 1. Models Channels Blocks Params (M) FLOPs (G) DCCNe Xt-B0 [32, 64, 160, 256] [3, 3, 5, 2] 5.5 0.8 DCCNe Xt-B1 [48, 96, 240, 384] [2, 2, 6, 2] 11.4 1.6 DCCNe Xt-B2 [64, 128, 320, 512] [3, 3, 9, 3] 26.7 4.0 DCCNe Xt-B3 [96, 192, 384, 576] [3, 12, 18, 3] 56.8 13.0 DCCNe Xt-B4 [96, 192, 384, 768] [5, 12, 20, 5] 86.7 15.7 Table 1: Detailed settings of DCCNe Xt series. Dynamic Clustering Convolution Converting Image to Patches: For the popular architectures, whether CNNs, Vision Transformers, Vision MLPs, or Vision GNNs, the input images are generally converted into patches before subsequent processing. Following the aforementioned strategy, the proposed method converts the original image into different patches first. For an image with input size H W 3, the image is divided into different patches. Each patch is considered as a cluster center in the clustering operation, resulting in a set of cluster centers C = {c1, c2, ..., cn}. By applying learnable parameters W, the original patch features are transformed into vectors, resulting in a set of cluster center features X = {x1, x2, ..., xn}, where xi RD, D is the feature dimension and i = 1, 2, , n. DCConv Block Patch Embedding Conv Blocks Blocks Conv Blocks Stage 1 Stage 4 Stage 3 Stage 2 Figure 2: The framework of the DCCNe Xt (DCCNe Xt-B1). Patch Embedding: network layer that makes patch embedding from the original image. H: height of the feature map. W: width of the feature map. D: feature dimension, FFN: feed-forward neural network (Vaswani et al. 2017). LN: layer normalization (Ba, Kiros, and Hinton 2016). DCConv: dynamic clustering convolution. Conv: 7 7 depthwise separable convolution (Howard et al. 2017; Chollet 2017). Dynamic Clustering: For each cluster center, we calculate the ℓ2-norm distance between each cluster center and other patches. The ℓ2-norm distance is defined as follows: distance(x, y) = (x y)2 , (1) where x is the cluster center, y is a patch other than the cluster center, x RD, y RD. After calculating the ℓ2-norm distances between each cluster center and the other patches, we obtain a distance matrix M Rn n for each cluster center and other patches. Next, based on this matrix M, we use a Top-K algorithm to select the K 1 patches that are closest in distance to each cluster center. Consequently, we obtain the index of patches within each cluster. This process can be represented as follows: idx = Topk Index(M), (2) where idx is the index of patches within clusters and idx Rn k . Based on the obtained index idx , we select the feature vectors of clusters from patch features, resulting in a feature vector matrix X Rn k d that is convenient for subsequent convolution. Within this feature vector matrix X, patches belonging to the same cluster are arranged in ascending order based on their distances from the cluster center. The above process can be represented as follows: X = Index Select(idx, X), (3) where X is the patch feature, idx is index of patches within clusters. Efficient Dynamic Clustering: Our clustering operation involves computing the ℓ2-norm distance between each cluster center and other patches, resulting in a quadratic complexity: Ω(DC) = h2w2C + 2hw C, (4) where h and w respectively are the height and width of the feature map, C is the feature dimension, and DC is the Dynamic Clustering. The above computation complexity is intractable for highresolution images. Considering that the feature vectors in the latter two stages are typically long and contain redundancy when calculating ℓ2-norm distance, we use subvectors Vsub = {a1, a1+d, a1+2d, ..., ac} of the feature vector V = {a1, a2, a3, ..., ac} to compute ℓ2-norm distance, where Vsub RC//d and V RC. Thus, it can result in a significant reduction in computational cost when processing high-resolution images: Ω(EDC) = h2w2 where h and w respectively are the height and width of the feature map, C is the feature dimension, and EDC is the efficient dynamic clustering, d represents the sampling interval for sub-vectors and is set to 8 in our model. Performing Convolution on Clusters: After obtaining n clusters, we perform convolution using a set of shared convolution kernels. To reduce the number of parameters, depthwise separable convolution (Howard et al. 2017; Chollet 2017) is commonly used in CNNs. Following this approach, the proposed dynamic clustering convolution model performs grouped convolution along the channels for improving computational efficiency. The proposed dynamic clustering convolution is defined as follows: k=0 X i w+j,k,c Wk,c + bc, (6) where Y is the output feature, X is the feature of patches within the cluster, W is the weight of the convolution kernels, b is the bias of the convolution kernels, K is the size of the cluster, i is the row coordinate of the patch in the image, j is the column coordinate of the patch in the image, w is the width of the feature map, k is the coordinate of the patch within the cluster, and c is the coordinate of the channel. Convolution FFN: To enhance the expressive capability of local information, inspired by the work (Chu et al. 2023; Islam, Jia, and Bruce 2020; Li et al. 2021; Wang et al. 2022), we insert 3 3 depthwise separable convolution (DWConv) (Howard et al. 2017; Chollet 2017) into the model s FFN to improve the model s ability for local modeling: Y = DWConv(Linear C 4C(Y )), (7) Y = Linear4C C{GELU(Y )}, (8) where Linear4C C means linear layer with input channels of 4 C and output channels of C, GELU denotes GELU (Hendrycks and Gimpel 2016) activation function. Experiments Image Classification Experimental settings: We use Image Net-1K (Russakovsky et al. 2015) to conduct the image classification experiments. Image Net-1K (Russakovsky et al. 2015) is also called the ISLVRC 2012 dataset, with 1K classes containing 1.28M training images and 50K validation images. We implement the proposed method by using Py Torch (Paszke et al. 2019) and Timm (Wightman et al. 2019). For a fair comparison, we follow the experimental parameter settings that are widely used in Dei T (Touvron et al. 2021a) and train the proposed model on a 2242-resolution image with 300 epochs. We employ the Adam W (Loshchilov and Hutter 2017) optimizer using a cosine decay learning rate scheduler with 20 epochs of linear warm-up. Data augmentation and regularization techniques include Rand Augmentation (Cubuk et al. 2020), Mixup (Zhang et al. 2017), Cut Mix (Yun et al. 2019), Random Erasing (Zhong et al. 2020), Weight Decay, Label Smoothing (Szegedy et al. 2016), and Stochastic Depth (Huang et al. 2016). We do not use Exponential Moving Average(EMA) (Polyak and Juditsky 1992), which does not improve the final performance of the model. Results: Table 2 lists the comparison results of the proposed DCCNe Xt with the representative backbones of different types on Image Net-1K (Russakovsky et al. 2015). DCCNe Xt outperforms common CNNs (Conv Ne Xt (Liu et al. 2022), Conv Ne Xt V2 (Woo et al. 2023) and SLa K (Liu et al. 2023)), Vision Transformers (FLatten-Swin (Han et al. 2023a) and Slide-PVT (Pan et al. 2023)), Vision MLPs (Cycle MLP (Chen et al. 2022)), Vision GNN (Vi HGNN (Han et al. 2023b)), and Vision SSM (VMamba (Liu et al. 2024)) with similar parameters and FLOPs. Furthermore, we observe that our approach far surpasses the state-of-the-art clustering-based vision architecture FEC-Small (Chen et al. 2024), even surpassing it by up to 3.1 points at a model size of around 5M (75.8% vs. 72.7%). When the model scales up to around 50M and 90M, our DCCNe Xt-B3 and DCCNe Xt B4 achieves top-1 accuracies of 84.3% and 84.5% respectively, surpassing various models of different types, indicating the strong scalability of our model. The above results highlight the excellence of our DCCNe Xt architecture. Object Detection and Instance Segmentation Settings: We conduct the object detection and instance segmentation experiments on COCO 2017 (Lin et al. 2014), Model Type Params (M)FLOPs (G)Top-1 (%) PVTv2-B0 (Wang et al. 2022) Vi Ts 3.4 0.6 70.5 TNT-Ti (Han et al. 2021) Vi Ts 6.1 1.4 73.9 Context-Cluster-Ti (Ma et al. 2023) Clustering 5.3 1.0 71.8 FEC-Small (Chen et al. 2024) Clustering 5.5 1.4 72.7 DCCNe Xt-B0 (ours) CNNs 5.5 0.8 75.8 PVTv2-B1 (Wang et al. 2022) Vi Ts 13.1 2.1 78.7 Cycle MLP-B1 (Chen et al. 2022) MLPs 15.0 2.1 78.9 Vi HGNN-Ti (Han et al. 2023b) GNNs 12.3 2.3 78.9 Context-Cluster-Small (Ma et al. 2023) Clustering 14.0 2.6 77.5 FEC-Base (Chen et al. 2024) Clustering 14.4 3.4 78.1 Res Net-18 (He et al. 2016) CNNs 12.0 1.8 70.6 DCCNe Xt-B1 (ours) CNNs 11.4 1.6 79.7 Swin-T (Liu et al. 2021) Vi Ts 29 4.5 81.3 FLatten-Swin-T (Han et al. 2023a) Vi Ts 29 4.5 82.1 Slide-PVT-S (Pan et al. 2023) Vi Ts 22.7 4.0 81.7 PVTv2-B2 (Wang et al. 2022) Vi Ts 25.4 4.0 82.0 QFormerh-T (Zhang et al. 2024) Vi Ts 29 4.6 82.5 Cycle MLP-B2 (Chen et al. 2022) MLPs 27 3.9 81.6 Vi HGNN-S (Han et al. 2023b) GNNs 28.5 6.3 82.5 Context-Cluster-Medium (Ma et al. 2023)Clustering 27.9 5.5 81.0 FEC-Large (Chen et al. 2024) Clustering 28.3 6.5 81.2 VMamba-T (Liu et al. 2024) SSMs 22 5.6 82.2 Res Net-50 (He et al. 2016) CNNs 25.6 4.1 79.8 SLa K-T (Liu et al. 2023) CNNs 30 5.0 82.5 Conv Ne Xt-T (Liu et al. 2022) CNNs 28.6 4.5 82.1 Inception Ne Xt-T (Yu et al. 2024) CNNs 28 4.2 82.3 Conv Ne Xt V2-T (Woo et al. 2023) CNNs 28.6 4.5 82.5 DCCNe Xt-B2 (ours) CNNs 26.7 4.0 82.8 Swin-S (Liu et al. 2021) Vi Ts 50 8.7 83.0 FLatten-Swin-S (Han et al. 2023a) Vi Ts 51 8.7 83.5 PVTv2-B3 (Wang et al. 2022) Vi Ts 45.2 6.9 83.2 QFormerh-S (Zhang et al. 2024) Vi Ts 51 8.9 84.0 Cycle MLP-B4 (Chen et al. 2022) MLPs 52 10.1 83.0 Vi HGNN-M (Han et al. 2023b) GNNs 52.4 10.7 83.4 VMamba-S (Liu et al. 2024) SSMs 44 11.2 83.5 Res Net-152 (He et al. 2016) CNNs 60.2 11.5 81.8 SLa K-S (Liu et al. 2023) CNNs 55 9.8 83.8 Conv Ne Xt-S (Liu et al. 2022) CNNs 50 8.7 83.1 Inception Ne Xt-S (Yu et al. 2024) CNNs 49 8.4 83.5 DCCNet-B3 (ours) CNNs 56.8 13.0 84.3 Swin-B (Liu et al. 2021) Vi Ts 88 15.4 83.5 FLatten-Swin-B (Han et al. 2023a) Vi Ts 89 15.4 83.8 PVTv2-B5 (Wang et al. 2022) Vi Ts 82 11.8 83.8 QFormerh-B (Zhang et al. 2024) Vi Ts 90 15.7 84.1 Cycle MLP-B5 (Chen et al. 2022) MLPs 76 12.3 83.2 Vi HGNN-B (Han et al. 2023b) GNNs 94.4 18.1 83.9 VMamba-B (Liu et al. 2024) SSMs 75 18.0 83.7 SLa K-B (Liu et al. 2023) CNNs 95 17.1 84.0 Conv Ne Xt-B (Liu et al. 2022) CNNs 89 15.4 83.8 Inception Ne Xt-B (Yu et al. 2024) CNNs 87 14.9 84.0 Conv Ne Xt V2-B (Woo et al. 2023) CNNs 89 15.4 84.3 DCCNet-B4 (ours) CNNs 86.7 15.7 84.5 Table 2: Results of DCCNe Xt and other backbones on Image Net-1K (Russakovsky et al. 2015). which contains 118K training images, 5K validation images, and 41K test images. We use MMDetection (Chen et al. 2019) to implement object detection and instance segmentation. For a fair comparison, we use the model pre-trained on Image Net-1K (Russakovsky et al. 2015) as the backbone, and the object detection and instance segmentation frameworks are Retina Net (Lin et al. 2017) and Mask R-CNN (He et al. 2017), respectively. Results: We list the results in Table 3. For Retina Net (Lin et al. 2017) based object detection, we report the Average Precision (AP) at different IOU thresholds (50%, 75%) and three different object sizes (small, medium, and large). From the results, we observe that the proposed DCCNe Xt consistently outperforms other models in AP at different thresholds, especially excelling in the detection of medium and large objects. Even in small object detection, the model Backbone Retina Net 1 Param (M)FLOPs (G) AP AP50 AP75 APS APM APL Res Net-50 37.7 239.3 36.3 55.3 38.6 19.3 40.0 48.8 PVT-Small 34.2 226.5 40.4 61.3 44.2 25.0 42.9 55.7 Cycle MLP-B2 36.5 230.9 40.6 61.4 43.2 22.9 44.4 54.5 Swin-T 38.5 244.8 41.5 62.1 44.2 25.1 44.9 55.5 Conv Ne Xt V2-T 38.8 243.2 41.7 62.5 44.6 25.2 45.2 55.5 DCCNe Xt-B2 (ours) 33.8 247.1 43.0 63.8 46.1 25.6 47.4 56.1 Backbone Mask R-CNN 1 Param (M)FLOPs (G) APb APb 50 APb 75 APm APm 50 APm 75 Res Net-50 44.2 260.1 38.0 58.6 41.4 34.4 55.1 36.7 FEC-Large 47.1 - 39.9 62.5 43.2 37.3 59.5 39.5 PVT-Small 44.1 245.1 40.4 62.9 43.8 37.8 60.1 40.3 Slide-PVT-S 42.0 269.0 42.8 65.9 46.7 40.1 63.1 43.1 Cycle MLP-B2 46.5 249.5 42.1 64.0 45.7 38.9 61.2 41.8 Swin-T 47.8 264.0 42.2 64.6 46.2 39.1 61.6 42.0 Conv Ne Xt V2-T 48.1 262.1 42.5 63.5 46.5 38.8 60.7 41.7 DCCNe Xt-B2 (ours) 43.7 281.1 44.6 66.1 49.0 40.6 63.4 43.5 Table 3: Object detection and instance segmentation results on COCO val2017 (Lin et al. 2014). The proposed DCCNe Xt is compared with the other backbones using Retina Net (Lin et al. 2017) and Mask R-CNN (He et al. 2017) frameworks. We calculate FLOPs with 1280 800 input size. demonstrates strong performance. This indicates that our model possesses good local modeling capabilities alongside its powerful long-range modeling capabilities. For Mask RCNN (He et al. 2017) based object detection and instance segmentation, we report bounding box and mask Average Precision (APb and APm) at different IOU thresholds (50%, 75%). Our approach demonstrates a clear advantage across all metrics, highlighting its outstanding performance in both object detection and instance segmentation tasks. Semantic Segmentation Settings: We choose ADE20K (Zhou et al. 2017) to conduct the semantic segmentation. The ADE20K includes 20K training images, 2K validation images, and 3K test images covering 150 semantic categories. We use MMSEG (Contributors 2020) as the framework and Semantic FPN (Kirillov et al. 2019) as the segmentation head. In the training phase, the backbone is initialized with the weights pretrained on Image Net-1K (Russakovsky et al. 2015), and the newly added layers are initialized with Xavier (Glorot and Bengio 2010). We use Adam W (Loshchilov and Hutter 2017) to optimize our model with an initial learning rate of 1e-4. Following the common practices (Kirillov et al. 2019; Chen et al. 2017), we train our models for 40k iterations with a batch size of 32. The learning rate is decayed following the polynomial decay schedule with a power of 0.9. In the training phase, we randomly resize and crop the image to 512 512. In the testing phase, we rescale the image with a shorter side of 512 pixels. Results: As shown in Table 4, we observe that the proposed DCCNe Xt outperforms the representative backbones of different types in semantic segmentation, including Conv Ne Xt V2 (Woo et al. 2023) (CNN), Swin Transformer (Liu et al. 2021) and Slide-PVT (Pan et al. 2023) (Vision Transformer), Cycle MLP (Chen et al. 2022) (Vision MLP). Due to the ab- Backbone Param (M)FLOPs (G)m IOU (%) Cycle MLP-B2 (Chen et al. 2022) 30.6 - 42.4 PVT-Small (Wang et al. 2021) 28.2 44.5 39.8 Slide-PVT-S (Pan et al. 2023) 26.0 - 42.5 Swin-T (Liu et al. 2021) 31.9 - 41.5 Pool Former-S36 (Yu et al. 2022c) 35.0 48.0 42.0 FEC-Large (Chen et al. 2024) 31.9 - 40.5 Res Net-50 (He et al. 2016) 28.5 45.6 36.7 Inception Ne Xt-T (Yu et al. 2024) 28.0 44.0 43.1 Conv Ne Xt V2-T (Woo et al. 2023) 32.2 45.2 43.7 DCCNe Xt-B2 (ours) 27.8 43.3 44.5 Table 4: Results of semantic segmentation on ADE20K validation set, - represents no relevant data. We calculate FLOPs with input size 512 512 for Semantic FPN (Kirillov et al. 2019). sence of relevant data, we do not compare the semantic segmentation results with Vision GNN. In general, DCCNe Xt performs competently in semantic segmentation and outperforms various backbones. Ablation Studies The convolution kernel size. The size of the convolution kernel is a crucial factor affecting the performance of dynamic clustering convolution. A convolution kernel that is too small may result in an insufficient receptive field, whereas a convolution kernel that is too large with inadequate sparsity may result in performance saturation and degradation. We tune kernel size K from 4 to 18 and show the experimental results in Table 5. From Table 5, it is evident that the optimal performance is obtained when the convolution kernel size is set to 12 and 18. K 4 8 12 16 12 and 18 Top-1(%) 78.1 78.5 78.6 78.5 78.7 Table 5: Image Net-1K results of different sizes of the convolution kernel. The basic architecture is DCCNe Xt-B1 and the number of training epochs is set to 150. 12 and 18 denote that the convolutional kernel sizes used in the third and fourth stages are 12 and 18, respectively. The effects of modules in DCCNe Xt. As shown in Table 6, we conduct the ablation experiments on three key modules, namely dynamic clustering convolution, Convolution FFN, and subvector sampling. It can be seen that without dynamic clustering convolution for global feature extraction, the performance is significantly worse (77.4% vs. 75.3%). For DCCNe Xt, the performance of the model with Conv FFN is improved by 1.3%, which indicates that local information is crucial for DCCNe Xt. When the model is processing high-resolution images (object detection is generally tested on a resolution of 1280 800 to calculate FLOPs), subvector sampling can greatly reduce 50% computational cost (39.2 G vs. 79.5 G) on the resolution of 1280 800 while ensuring the performance. Method Params FLOPs (224 224) FLOPs (1280 800) Image Net top-1 acc. DCConv None 11.2 M 1.52 G 31.0 G 75.3% +DCConv 11.3 M 1.65 G 78.9 G 77.4% ++Conv FFN 11.4 M 1.67 G 79.5 G 78.7% +++Subvector Sampling 11.4 M 1.58 G 39.2 G 78.7% Table 6: The effects of modules in DCCNe Xt on Image Net1K. The basic architecture is DCCNe Xt-B1 and the number of training epochs is set to 150. (a) Input image. (b) An cluster in the 7th block. (c) An cluster in the 16th block. Figure 3: An example of the cluster distribution after dynamic clustering for different blocks, where the red patches are the cluster centers and the orange patches are the other patches in the cluster. (a) Convolution kernel weights (b) Convolution kernel weights in the 7th block. in the 16th block. Figure 4: Heat map visualization of dynamic clustering convolution kernel weights for different blocks. The horizontal coordinate represents the subscript of the convolution kernel weights, while the vertical coordinate represents the weight value. To better reflect the general trend of the weights of the convolution kernels at different positions, where the weights are first taken absolutely and then averaged over the channels. Visualization To better comprehend the working principles of the proposed dynamic clustering convolution, we visualize the clusters in DCCNe Xt-B2. As shown in Figure 3, we present clusters at different depths (the 7th and 16th blocks). Our observation is that the proposed model clusters image patches with similar semantics globally into the same cluster. Furthermore, we visualize the convolution kernels of dynamic clustering convolution in DCCNe Xt-B2. Figure 4 shows the weights of convolution kernels in different layers (the 7th and 16th blocks). From Figure 4, we observe that the convolution weights of patches closer to the cluster center in the same cluster are larger. These results indicate that the proposed dynamic clustering convolution can assign different weights to the patches in the cluster based on the patches contributions to the cluster. In this paper, we have proposed a novel convolutional neural network, termed Dynamic Clustering Convolutional Neural Network (DCCNe Xt). Different from the prevalent scheme of modeling local windows using standard convolutions, the proposed method dynamically clusters image patches globally into multiple clusters and then convolves each cluster with a shared set of convolution kernels. Thus, the proposed method has a more expressive capability to capture long-range dependencies of objects, which are lacking in standard convolutions. Furthermore, the subspaces of feature vectors are extracted to perform clustering, which significantly reduces the computational cost of the model on high-resolution images. Thus, the proposed method can be applied to various downstream tasks involving images with high resolutions. The extensive experiments on image classification, object detection, instance segmentation, and semantic segmentation have demonstrated the superiority of the proposed method. We hope this proposed novel flexible vision architecture can offer a promising avenue for convolutional neural networks to be widely adapted to various real-world vision tasks. Acknowledgements The work was supported by the National Key Research and Development Program of China (No. 2023YFC3306401) and the National Natural Science Foundation of China (Nos. 62072386, 62076016). This research was also supported by Zhejiang Provincial Natural Science Foundation of China (No. LD24F020007), Beijing Natural Science Foundation of China (Nos. L223024 and L244043), One Thousand Plan projects in Jiangxi Province (No. Jxsq2023102268), supported by Yunnan Provincial Major Science and Technology Special Plan Project (No. 202402AD080001), supported by Henan Province key research and development project (No. 231111212000) , supported by the Open Foundation of Henan Key Laboratory of General Aviation Technology of China (No. ZHKF-230212) and the Open Foundation the Key Laboratory of Oracle Information Processing of Ministry of Education of China (No. OIP2024E002). References Ba, J.; Kiros, J.; and Hinton, G. 2016. Layer Normalization. Chen, G.; Li, X.; Yang, Y.; and Wang, W. 2024. Neural clustering based visual representation learning. In CVPR, 5714 5725. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. 2019. MMDetection: Open mmlab detection toolbox and benchmark. ar Xiv preprint ar Xiv:1906.07155. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; and Yuille, A. L. 2017. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE PAMI, 40(4): 834 848. Chen, S.; Xie, E.; Ge, C.; Liang, D.; and Luo, P. 2022. Cyclemlp: A mlp-like architecture for dense prediction. In ICLR. Chollet, F. 2017. Xception: Deep learning with depthwise separable convolutions. In CVPR, 1251 1258. Chu, X.; Tian, Z.; Zhang, B.; Wang, X.; and Shen, C. 2023. Conditional Positional Encodings for Vision Transformers. In ICLR. Contributors, M. 2020. MMSegmentation: Openmmlab semantic segmentation toolbox and benchmark. Cubuk, E. D.; Zoph, B.; Shlens, J.; and Le, Q. V. 2020. Randaugment: Practical automated data augmentation with a reduced search space. In CVPRW, 702 703. Ding, X.; Zhang, X.; Han, J.; and Ding, G. 2022. Scaling up your kernels to 31x31: Revisiting large kernel design in cnns. In CVPR, 11963 11975. Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; and Sun, J. 2021. Repvgg: Making vgg-style convnets great again. In CVPR, 13733 13742. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. 2021. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR. Glorot, X.; and Bengio, Y. 2010. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, 249 256. JMLR. Han, D.; Pan, X.; Han, Y.; Song, S.; and Huang, G. 2023a. Flatten transformer: Vision transformer using focused linear attention. In ICCV, 5961 5971. Han, K.; Xiao, A.; Wu, E.; Guo, J.; Xu, C.; and Wang, Y. 2021. Transformer in transformer. Neur IPS, 34: 15908 15919. Han, Y.; Wang, P.; Kundu, S.; Ding, Y.; and Wang, Z. 2023b. Vision HGNN: An Image is More than a Graph of Nodes. In ICCV, 19878 19888. He, K.; Gkioxari, G.; Doll ar, P.; and Girshick, R. 2017. Mask r-cnn. In ICCV, 2961 2969. He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In CVPR, 770 778. Hendrycks, D.; and Gimpel, K. 2016. Gaussian error linear units (gelus). ar Xiv preprint ar Xiv:1606.08415. Hinton, G.; Vinyals, O.; and Dean, J. 2015. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531. Hou, Q.; Zhou, D.; and Feng, J. 2021. Coordinate attention for efficient mobile network design. In CVPR, 13713 13722. Howard, A. G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; and Adam, H. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. ar Xiv preprint ar Xiv:1704.04861. Hu, J.; Shen, L.; and Sun, G. 2018. Squeeze-and-excitation networks. In CVPR, 7132 7141. Huang, G.; Liu, Z.; Van Der Maaten, L.; and Weinberger, K. Q. 2017. Densely connected convolutional networks. In CVPR, 4700 4708. Huang, G.; Sun, Y.; Liu, Z.; Sedra, D.; and Weinberger, K. Q. 2016. Deep networks with stochastic depth. In ECCV, 646 661. Springer. Huang, Z.; and Li, Y. 2020. Interpretable and accurate finegrained recognition via region grouping. In CVPR, 8662 8672. Islam, A.; Jia, S.; and Bruce, N. D. B. 2020. How Much Position Information Do Convolutional Neural Networks Encode. ar Xiv preprint ar Xiv:2001.08248. Jampani, V.; Sun, D.; Liu, M.-Y.; Yang, M.-H.; and Kautz, J. 2018. Superpixel sampling networks. In ECCV, 352 368. Kirillov, A.; Girshick, R.; He, K.; and Doll ar, P. 2019. Panoptic feature pyramid networks. In CVPR, 6399 6408. Krizhevsky, A.; Sutskever, I.; and Hinton, G. E. 2012. Imagenet classification with deep convolutional neural networks. In Neur IPS, 1097 1105. Le Cun, Y.; Bottou, L.; Bengio, Y.; and Haffner, P. 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11): 2278 2324. Li, Y.; Zhang, K.; Cao, J.; Timofte, R.; and Gool, L. V. 2021. Local Vi T: Bringing Locality to Vision Transformers. ar Xiv preprint ar Xiv:2104.05707. Li, Z.; and Chen, J. 2015. Superpixel segmentation using linear spectral clustering. In CVPR, 1356 1363. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; and Doll ar, P. 2017. Focal loss for dense object detection. In ICCV, 2980 2988. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Doll ar, P.; and Zitnick, C. L. 2014. Microsoft coco: Common objects in context. In ECCV, 740 755. Springer. Liu, S.; Chen, T.; Chen, X.; Chen, X.; Xiao, Q.; Wu, B.; K arkkainen, T.; Pechenizkiy, M.; Mocanu, D. C.; and Wang, Z. 2023. More Conv Nets in the 2020s: Scaling up Kernels Beyond 51x51 using Sparsity. In ICLR. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; and Liu, Y. 2024. Vmamba: Visual state space model. ar Xiv preprint ar Xiv:2401.10166. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; and Guo, B. 2021. Swin transformer: Hierarchical vision transformer using shifted windows. In ICCV, 10012 10022. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; and Xie, S. 2022. A convnet for the 2020s. In CVPR, 11976 11986. Loshchilov, I.; and Hutter, F. 2017. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101. Ma, X.; Qin, C.; You, H.; Ran, H.; and Fu, Y. 2022. Rethinking network design and local geometry in point cloud: A simple residual MLP framework. ar Xiv preprint ar Xiv:2202.07123. Ma, X.; Zhou, Y.; Wang, H.; Qin, C.; Sun, B.; Liu, C.; and Fu, Y. 2023. Image as Set of Points. ICLR. Pan, X.; Ye, T.; Xia, Z.; Song, S.; and Huang, G. 2023. Slide Transformer: Hierarchical Vision Transformer with Local Self-Attention. In CVPR, 2082 2091. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. 2019. Pytorch: An imperative style, high-performance deep learning library. Neur IPS, 32. Polyak, B. T.; and Juditsky, A. B. 1992. Acceleration of stochastic approximation by averaging. SIAM Journal on Control and Optimization, 30(4): 838 855. Ren; and Malik. 2003. Learning a classification model for segmentation. In ICCV, 10 17. IEEE. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. 2015. Imagenet large scale visual recognition challenge. IJCV, 115: 211 252. Simonyan, K.; and Zisserman, A. 2014. Very deep convolutional networks for large-scale image recognition. ar Xiv preprint ar Xiv:1409.1556. Smith, S. L.; Brock, A.; Berrada, L.; and De, S. 2023. Conv Nets Match Vision Transformers at Scale. ar Xiv preprint ar Xiv:2310.16764. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; and Rabinovich, A. 2015. Going deeper with convolutions. In CVPR, 1 9. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; and Wojna, Z. 2016. Rethinking the inception architecture for computer vision. In CVPR, 2818 2826. Tan, M.; and Le, Q. 2019. Efficientnet: Rethinking model scaling for convolutional neural networks. In ICML, 6105 6114. PMLR. Tolstikhin, I. O.; Houlsby, N.; Kolesnikov, A.; Beyer, L.; Zhai, X.; Unterthiner, T.; Yung, J.; Steiner, A.; Keysers, D.; Uszkoreit, J.; et al. 2021. Mlp-mixer: An all-mlp architecture for vision. Neur IPS, 34: 24261 24272. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; and J egou, H. 2021a. Training data-efficient image transformers & distillation through attention. In ICML, 10347 10357. PMLR. Touvron, H.; Cord, M.; Sablayrolles, A.; Synnaeve, G.; and J egou, H. 2021b. Going deeper with image transformers. In ICCV, 32 42. Trockman, A.; and Kolter, J. Z. 2022. Patches are all you need? ar Xiv preprint ar Xiv:2201.09792. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; and Polosukhin, I. 2017. Attention is all you need. Neur IPS, 30. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; and Hu, Q. 2020. ECA-Net: Efficient channel attention for deep convolutional neural networks. In CVPR, 11534 11542. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; and Shao, L. 2021. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In ICCV, 568 578. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; and Shao, L. 2022. Pvt v2: Improved baselines with pyramid vision transformer. CVM, 8(3): 415 424. Wightman, R.; et al. 2019. Pytorch image models. Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I. S.; and Xie, S. 2023. Convnext v2: Co-designing and scaling convnets with masked autoencoders. In CVPR, 16133 16142. Woo, S.; Park, J.; Lee, J.-Y.; and Kweon, I. S. 2018. Cbam: Convolutional block attention module. In ECCV, 3 19. Xie, S.; Girshick, R.; Doll ar, P.; Tu, Z.; and He, K. 2017. Aggregated residual transformations for deep neural networks. In CVPR, 1492 1500. Yu, Q.; Wang, H.; Kim, D.; Qiao, S.; Collins, M.; Zhu, Y.; Adam, H.; Yuille, A.; and Chen, L.-C. 2022a. Cmt-deeplab: Clustering mask transformers for panoptic segmentation. In CVPR, 2560 2570. Yu, Q.; Wang, H.; Qiao, S.; Collins, M.; Zhu, Y.; Adam, H.; Yuille, A.; and Chen, L.-C. 2022b. k-means Mask Transformer. In ECCV, 288 307. Springer. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; and Yan, S. 2022c. Metaformer is actually what you need for vision. In CVPR, 10819 10829. Yu, W.; Si, C.; Zhou, P.; Luo, M.; Zhou, Y.; Feng, J.; Yan, S.; and Wang, X. 2022d. Metaformer baselines for vision. ar Xiv preprint ar Xiv:2210.13452. Yu, W.; Zhou, P.; Yan, S.; and Wang, X. 2024. Inceptionnext: When inception meets convnext. In CVPR, 5672 5683. Yun, S.; Han, D.; Oh, S. J.; Chun, S.; Choe, J.; and Yoo, Y. 2019. Cutmix: Regularization strategy to train strong classifiers with localizable features. In ICCV, 6023 6032. Zhang, H.; Cisse, M.; Dauphin, Y. N.; and Lopez-Paz, D. 2017. mixup: Beyond empirical risk minimization. ar Xiv preprint ar Xiv:1710.09412. Zhang, Q.; Zhang, J.; Xu, Y.; and Tao, D. 2024. Vision transformer with quadrangle attention. TPAMI. Zhang, X.; Zhou, X.; Lin, M.; and Sun, J. 2018. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In CVPR, 6848 6856. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; and Yang, Y. 2020. Random erasing data augmentation. In AAAI, 13001 13008. Zhou, B.; Zhao, H.; Puig, X.; Fidler, S.; Barriuso, A.; and Torralba, A. 2017. Scene parsing through ade20k dataset. In CVPR, 633 641.