# taskdisruptive_background_suppression_for_fewshot_segmentation__bbc0538f.pdf Task-Disruptive Background Suppression for Few-Shot Segmentation Suho Park, Su Been Lee, Sangeek Hyun, Hyun Seok Seong, Jae-Pil Heo* Sungkyunkwan University {shms0706, leesb7426, hsi1032, gustjrdl95, jaepilheo}@skku.edu Few-shot segmentation aims to accurately segment novel target objects within query images using only a limited number of annotated support images. The recent works exploit support background as well as its foreground to precisely compute the dense correlations between query and support. However, they overlook the characteristics of the background that generally contains various types of objects. In this paper, we highlight this characteristic of background which can bring problematic cases as follows: (1) when the query and support backgrounds are dissimilar and (2) when objects in the support background are similar to the target object in the query. Without any consideration of the above cases, adopting the entire support background leads to a misprediction of the query foreground as background. To address this issue, we propose Task-disruptive Background Suppression(TBS), a module to suppress those disruptive support background features based on two spatial-wise scores: query-relevant and target-relevant scores. The former aims to mitigate the impact of unshared features solely existing in the support background, while the latter aims to reduce the influence of target-similar support background features. Based on these two scores, we define a query background relevant score that captures the similarity between the backgrounds of the query and the support, and utilize it to scale support background features to adaptively restrict the impact of disruptive support backgrounds. Our proposed method achieves state-of-the-art performance on standard few-shot segmentation benchmarks. Our official code is available at github.com/Suho Park0706/TBSNet. 1 Introduction With the advance of deep learning, semantic segmentation (Long, Shelhamer, and Darrell 2015; Chen et al. 2017; Zhao et al. 2017; Seong et al. 2023) has achieved remarkable performance. However, its performance relies on abundant data for target classes and degrades noticeably with insufficient data. To resolve this, Few-Shot Segmentation (FSS) has been proposed (Shaban et al. 2017a) to build a model adaptable for novel classes with only a few number of labeled data. Briefly, FSS aims to learn a novel class with a small number of labeled images, called a support set, to segment an unlabeled image, called a query set. *Corresponding author Copyright 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. (a) Problematic cases caused by disruptive support features Ideal cases Problematic cases SF is similar to QF SB is dissimilar to QF & QB SB is similar to QF Query & Fg mask Support & Fg mask Query Support Baseline Ours (b) Visualization of regions for top-5 support features similar to query foreground Query Support Baseline Ours SB is similar to QB QF QB SF SB : Query Foreground : Query Background : Support Foreground : Support Background Figure 1: The abbreviation Q, S, F, B are query, support, foreground and background, respectively. (a) Two types of disruptive SB can occur: (1) SB dissimilar to all Q can introduce ambiguity in the relationship between QB and SB. (2) SB more similar to QF than QB can cause misclassification of QF as background. (b) We visualize the top-5 support regions similar to QF at the feature level, the closer to the top-1 the darker. In the first column, we illustrate cases with query-irrelevant SB, while the second column demonstrates scenarios where SB highly resembles QF. In both cases, our approach effectively suppresses disruptive SB that exhibits high similarity to QF, as indicated by the yellow boxes. A more detailed analysis is shown in Section 5.3. In Few-Shot Segmentation, affinity learning is a mainstream technique (Min, Kang, and Cho 2021; Shi et al. 2022; Zhang et al. 2021; Wang et al. 2020), learning pixel-wise correlations. Early methods (Wang et al. 2020; Min, Kang, and Cho 2021) utilize only the support foreground (SF) features to compute the correlations with the query features (Q). However, they overlook that the support background (SB) features also contain contextual information helpful for distinguishing between the query foreground (QF) and background (QB). For instance, the sky can be particularly use- The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Query-relevant score Support Query Query-relevant feat. Query background- relevant score Target-relevant score Support-target Support Target-relevant Scoring low high Score magnitude Target-relevant feat. Query-relevant Scoring Figure 2: Illustration of two types of scoring processes. Query-relevant scoring aims to reduce the influence of unshared features solely existing in the support background (yellow region) and target-relevant scoring is to restrict the influence of target-similar support background features (red bus). Query background-relevant score is computed based on these two scores to focus on shared background features (blue and skyblue regions in support image). ful to segment airplanes. Therefore, the most recent techniques (Zhang et al. 2021; Shi et al. 2022) leverage the entire set of support features (S), encompassing both foreground (SF) and background (SB). However, it is important to note that not every SB feature is beneficial for distinguishing between the QF and QB. Since a background can contain diverse objects, the dissimilarity between the QB and SB can disrupt the segmentation. Moreover, the SB may include objects similar to the target object in the query image. In such cases, the high similarity between the SB and QF may mislead the model to predict QF pixels as QB. Therefore, we need to filter out these harmful SB pixels. In this context, we propose Task-disruptive Background Suppression (TBS), a module designed to suppress harmful background features within a support set to enhance the segmentation of the query. Specifically, to determine the utility of background features, TBS defines two spatial-wise scores: query-relevant and target-relevant scores. The queryrelevant score determines the similarity between each SB feature and the entire Q features, calculated using the crossattention module (Doersch, Gupta, and Zisserman 2020). This score allows us to identify SB pixels relevant to the Q image while filtering out irrelevant ones. However, SB features similar to the QF may still persist, potentially leading to the misprediction of the QF mask. To address it, we introduce the target-relevant score for each SB feature, indicating the degree of similarity with the target object features in the support. Similar to the query-relevant score, the targetrelevant score is derived through the cross-attention between each SB feature and SF feature. By combining these two scores, we define a query background-relevant score; the relevance score of SB features to QB. As a result, we suppress task-disruptive SB features that are irrelevant to a given QF or similar to the target object class. This is achieved by multiplying these scores with SB features. To sum up, our contributions are summarized as follows: For the first time, we define the advantageousness of support background features based on the relation between background regions of query and support images. We propose a novel module, Task-disruptive Background Suppression (TBS) that restricts certain background features within the support set that do not contribute to precise segmentation of the query image. Our method achieves state-of-the-art performance over baselines, and its effectiveness is validated by various ablation studies. 2 Related Work 2.1 Few-Shot Segmentation The methods of Few-Shot Segmentation (FSS) can be categorized into two main streams: prototype-based and affinity learning methods. Prototype-based methods (Li et al. 2021; Zhang et al. 2020; Wang et al. 2019; Liu et al. 2020) represent the foreground objects in the support set as single or multiple prototypes. They classify the pixels of the query image into foreground and background based on their similarity to the prototypes. However, these methods may lead to deteriorated segmentation results, since they lose information about support objects while summarizing the images with only a few representative features. On the other hand, affinity learning methods (Min, Kang, and Cho 2021; Hong et al. 2022; Wang, Sun, and Zhang 2023) leverage pixel-level dense correlations between the object features of support set and query features. Moreover, recent techniques (Zhang et al. 2021; Shi et al. 2022) have found that the background features of the support set are also useful in distinguishing between the foreground and background of the query. Therefore, they compute correlations with query features using not only the object features but also the background features. Although they show impressive performance by exploiting whole support features, they still overlook that some background features are not useful for classifying the foreground and background of the query image. On the other hand, our method assigns low weights to those task-disruptive background features to prevent the pixel of the query image from being misclassified. 2.2 Feature Suppression in FSS The cross-attention module in few-shot segmentation is generally employed to constrain the disruptive support features in distinguishing between the query foreground and background (Wang, Sun, and Zhang 2023; Peng et al. 2023; Zhang et al. 2021). For example, Cy CTR (Zhang et al. 2021) utilizes bidirectional cross-attention to identify the most similar support feature for each query feature and vice versa. Consequently, when the category of a support feature (the starting feature) differs from the class of the most similar support feature (the ending feature) for the corresponding query feature, the starting feature is identified as potentially disruptive and is mitigated. similary, ABCNet (Wang, Sun, and Zhang 2023) suppresses unuseful support features by indirectly comparing the support and query features through computing cross-attention with a reference pool. However, these methods may struggle to demonstrate effectiveness especially when the support background is similar to the query The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Convolution Block Query-relevant scoring Shared weight Cross Attention Target-relevant scoring Score magnitude Shared weight Cross Attention Score magnitude f : Subtraction f : Multiplication 𝐹𝑆: Support feature 𝐹𝑄: Query feature 𝐹𝑂: Masked Support feature 𝑀𝑆: Support Mask Figure 3: Illustration of Task-disruptive Background Suppression (TBS) module. TBS incorporates two scoring processes to compute query-relevant and target-relevant scores. By representing support features F S based on query features F Q by the cross-attention, we can get the query-relevant scores RQ which signify how well each support feature describes the query feature. On the other hand, the target-relevant score RT assesses whether each support background feature is similar to the target object feature, obtained by representing the support feature with the support object feature. By employing the score refinement by RQ RT, query background relevant score is computed to activate regions that are query-relevant but targetirrelevant. Consequently, by applying RB to the original support features F S through multiplication, we can obtain adaptive support features AS. foreground. This limitation arises because they do not consider the categorically derived relationship between features. Although our method also restrains disruptive support features like existing methods based on cross-attention, we utilize the relationship between the support foreground and background for the first time. By incorporating this novel relationship in addition to the relation between the support and query, we can discover task-disruptive support features with consideration of only the similarity between the support background and the query background. 3 Our Method The overall architecture of our method is illustrated in Fig. 3. Affinity learning methods in FSS require precise pairing between the query and support features, especially based on their binary class (foreground and background). However, as noted in the Introduction, backgrounds of different images typically consist of various objects and may not share common characteristics, in contrast to the foregrounds. Thus, in this paper, we aim to refine support background features by considering the following two conditions. First, the support background features similar to the query background features are preferred. Another is that the support background features should be distinguished from the query object features. In other words, these two conditions intend to reduce the gap between backgrounds of support and query, and enhance the disparity between background and foreground simultaneously, encouraging the query and support features to be well-clustered according to their binary class. To meet these requirements, we introduce two representativeness scores: query-relevant score RQ and target-relevant score RT, which are pixel-level importance scores for support background features. Specifically, RQ signifies whether the support background features can describe the query features. Thus, the support background features well represented by the query features would have high RQ. However, since the query features used in computing RQ contain both the query background and foreground, it cannot be guaranteed that RQ is derived only from the background. To resolve this, we define RT which is highly activated when the support background features are similar to the target object features in the support set. By subtracting RT from RQ, we filter out the scores activated by the query target object at RQ. We define the filtered scores as background-relevant score RB which restrains task-disruptive support background features containing the information of the target object. We utilize RB to selectively suppress the influence of the support background features. In the following sections, we provide the problem definition and a detailed explanation of the aforementioned scores The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) for refining the support background features. 3.1 Problem Definition As a standard formulation of the few-shot segmentation problem, we have two disjoint datasets: Dtrain for training a model and Dtest for evaluating a learned model. Each dataset consists of distinct object classes Ctrain and Ctest without any overlap (Ctrain Ctest = ). Generally, the training and testing of few-shot segmentation are composed of several episodes. Each episode consists of K labeled images and an unlabeled image, i.e., K-shot episode, where all images contain objects of the same category c randomly sampled from the dataset. Specifically, the labeled images are called a support set S = IS j , M S j K j=1, and the unlabeled image is named a query set Q = IQ, M Q . Here, I is an image and M denotes a corresponding ground-truth binary mask, containing a value of 1 for foreground object regions, belonging to category c, and 0 for the others. The goal of the few-shot segmentation is to predict M Q based on S. 3.2 Spatial-wise Representativeness Scores For features of the j-th support image F S j computed by a feature extractor, we spatially divide F S j into support object features F O j and support background features F B j by ground-truth mask M S j , as follows: F O j = f S j,h,w|m S j,h,w = 1 F B j = f S j,h,w|m S j,h,w = 1 , (1) where f S j,h,w and m S j,h,w denote the value spatially located at (h, w) of F S j and M S j , respectively. Among these two sorts of features, F O j can be treated as valuable features for the task, since they only contains the categorical information of target objects that we need to detect in the query set Q. In contrast, F B j may include both useful and unuseful features, so thus it is crucial to prevent the unuseful features within F B j from being utilized in the segmentation process. Therefore, we introduce a query-relevant score RQ of how much each pixel well represents query-relevant information. To measure the relevance of each pixel, we utilize the cross-attention module (Doersch, Gupta, and Zisserman 2020). However, unlike the existing cross-attention, we reconstruct the support features F S j based on the query features F Q. Specifically, reconstructed support features F S j are computed as follows: F S j = V F Q Softmax Q F S j K F Q T where d is the channel dimension of projection, and Q, K, and V are linear heads for query, key, and value, respectively. Since the reconstruction quality of F S j would be high when the support feature is similar to the query features, we define the similarity between F S j and V F S j with additional linear projection as the RQ j which represents queryrelevance score of j-th support feature F S j , as follows: RQ j = CS L( F S j ), L(V F S j ) , (3) where CS ( , ) is the cosine-similarity function, and L is a linear projection layer. Although we can estimate how much each pixel of support features is relevant to query features based on RQ, it is not solely derived from its similarity with the query background. In this case, the query-relevance score alone can be problematic, since the high similarity between the support background and query object induces the pixels of the query objects to be predicted as the background. To resolve this issue, we define a complementary score, a target-relevant score RT, which indicates whether each pixel is similar to the target object. We first compute reconstructed support features ˆF S j with the support object features F O j , as follows: ˆF S j = V F O j Softmax Q F S j K F O j T Then, RT j is determined the similarity between ˆF S j and V F O j , as follows: RT j = CS L( ˆF S j ), L(V F S j ) . (5) Note that, the parameters of projection heads such as L, Q, K, and V are shared with Eq. 2 3.3 Task-disruptive Background Suppression We first define a query background-relevant score RB through the subtraction of RT from RQ and multiply it with a support background mask 1 M S element-wisely. Since the subtraction operation signifies the removal of target object relevance from query relevance, the score RB reflects the similarity of each pixel in the support background features to the query background features. We convert this score into spatial-wise weights to suppress disruptive support background features as follows: RB j = b RB j , (6) where b is a shallow convolution block for refinement and the architecture of it is explained in Section 4.2. Recall that our goal is to refine the support background features to resolve an undesirable background matching problem. However, directly multiplying the score map RB j with the support feature F S j would result in the incomplete preservation of foreground features. To prevent this, we replace the value of RB j corresponding to M O j with 1, denoting it as RB j . Consequently, we obtain adaptive support features AS j by multiplying RB j with F S j . We then utilize AS j in the subsequent segmentation process instead of F S j . 4 Experiments 4.1 Datasets and Evaluation Metrics We utilize PASCAL-5i (Shaban et al. 2017b) and COCO20i(Nguyen and Todorovic 2019) following the prior works (Zhang et al. 2021; Shi et al. 2022; Peng et al. 2023; Wang, Sun, and Zhang 2023). PASCAL-5i combines data The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Method Backbone 1-shot/5-shot F-0 F-1 F-2 F-3 m Io U FB-Io U CWT (Lu et al. 2021) 56.9/62.6 65.2/70.2 61.2/68.8 48.8/57.2 58.0/64.7 - Do G-LSTM (Azad et al. 2021) 57.0/57.3 67.2/68.5 56.1/61.5 54.3/56.3 58.7/60.9 - DAN (Wang et al. 2020) 54.7/57.9 68.6/69.0 57.8/60.1 51.6/54.9 58.2/60.5 71.9/72.3 HSNet (2021) 67.3/71.8 72.3/74.4 62.0/67.0 63.1/68.3 66.2/70.4 77.6/80.6 Cy CTR (Zhang et al. 2021) 67.2/71.0 71.1/75.0 57.6/58.5 59.0/65.0 63.7/67.4 - Cy CTR + TBS (Ours) 67.8/72.3 70.9/74.7 57.8/59.5 59.6/65.5 64.0/68.0 - DCAMA (Shi et al. 2022) 65.4/70.7 71.4/73.7 63.2/66.8 58.3/61.9 64.6/68.3 77.6/80.8 DCAMA + TBS (Ours) 68.5/72.3 72.0/74.1 63.8/68.4 59.5/67.2 65.9/70.5 77.7/81.3 HSNet (2021) 67.9/72.2 74.0/77.5 60.3/64.0 67.0/72.6 67.3/71.6 77.9/81.2 DCAMA (Shi et al. 2022) 72.2/75.7 73.8/77.1 64.3/72.0 67.1/74.8 69.3/74.9 78.5/82.9 DCAMA + TBS (Ours) 74.7/76.5 74.4/76.5 66.1/71.4 69.5/76.3 71.2/75.2 80.0/83.4 Table 1: Experimental results on the PASCAL-5i dataset with Res Net-101 and Swin-B backbones. Method Backbone 1-shot/5-shot F-0 F-1 F-2 F-3 m Io U FB-Io U HSNet (2021) 43.6/50.1 49.9/58.6 49.4/56.7 46.4/55.1 47.3/55.1 72.5/76.1 DCAMA (Shi et al. 2022) 49.5/55.4 52.7/60.3 52.8/59.9 48.7/57.5 50.9/58.3 73.2/76.9 DCAMA + TBS (Ours) 49.6/53.7 54.3/62.9 54.1/59.3 51.3/58.2 52.3/58.5 74.2/76.8 Table 2: Experimental results on the COCO-20i dataset with Swin-B backbone. from PASCAL VOC 2012 (Williams 2010) and SDS (Hariharan et al. 2014), comprising 20 categories. In contrast, COCO-20i is a subset of COCO (Lin et al. 2014) and is comprised of 80 categories. And, each dataset is divided into four folds where it has the same number of categories that do not overlap with others. Hence, each fold of PASCAL-5i and COCO-20i have 5 and 20 classes, respectively. To evaluate the model s adaptability to novel classes, we adopt a cross-validation scheme where each fold is selected as Dtest and others are used as Dtrain. Then, we evaluate the model with mean intersection over union (m Io U) and foreground-background intersection over union (FBIo U) for 1000 episodes randomly sampled from Dtest. 4.2 Implementation Details To verify the high adaptability of TBS, we apply it to two baseline models: DCAMA (Shi et al. 2022) and Cy CTR (Zhang et al. 2021). For fair comparisons with baselines, we adopt Res Net-101 pretrained on Image Net and Swin-Transformer pre-trained on Image Net 1K as a feature extractor. In the case of DCAMA with Swin-Transformer, we apply TBS at scales of 1 8, 1 16, and 1 32 to align with the scale used in DCAMA s cross-attention mechanism. However, for DCAMA with Res Net-101, we utilize TBS only in 1 16 and 1 32 scales due to memory limitation. On the other hand, since Cy CTR was verified only on Res Net, we conducted experiments on Res Net-101, not Swin-Transformer. Unlike DCAMA which adopts multi-level features, Cy CTR utilizes single-level features generated by combining features from 3and 4-th blocks. Therefore, we suppress only those combined features by using TBS. Many hyperparameters, i.e., optimizer, learning rate, batch size, etc., are the same as the baseline. We describe the flow of the convolution block for converting the query background-relevant score into the spatialwise weight. It first concatenates the input score map and the layer normalization of the score map to reference the spatial-wise distribution of scores. Subsequently, two consecutive layers of 1 1 convolution without any non-linear activation are employed. This projects the input score maps into 256 channels and squeezes it into 1 channel again. After that, a sigmoid function is applied to make the score map into the range between 0 and 1. 4.3 Experimental Results Quantitative Results. We evaluate our proposed method by comparing it with previous techniques designed for fewshot segmentation. As illustrated in Table 1, recent affinity learning models, specifically Cy CTR and DCAMA, already exhibit comparable performances. Upon incorporating TBS into these approaches, a consistent improvement over baseline model performances is observed, resulting in the state-of-the-art scores. This improvement remains consistent across various evaluation metrics and different quantities of labeled images on the PASCAL-5i dataset. Similar trends are observed in the 1-shot scenario of COCO-20i. As demonstrated in Table 2, TBS consistently enhances DCAMA s performance across all folds, providing the best performance. While its impact is less pronounced in the 5-shot scenario compared to the 1-shot scenario, where it shows substantial effectiveness, TBS still succeeds in improving the average m Io U of DCAMA. As a result, TBS surpasses the existing state-of-the-art performance in three out of four quantitative benchmark scenarios in the context of few-shot segmentation. This verifies the effectiveness of suppressing disruptive support, particu- The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) (b) 5-shot scenario Support GT DCAMA Ours (a) 1-shot scenario Support GT DCAMA Ours Figure 4: Qualitative comparison results of DCAMA and DCAMA + TBS (Ours) on PASCAL-5i dataset with Swin-B. larly in situations of extreme data scarcity. Qualitative Results. In addition to the quantitative results, we report qualitative results to intuitively show the effectiveness of TBS. Compared with DCAMA, our results include fewer mispredicted pixels regardless of the number of support images as shown in Fig. 4. Especially, when objects in the support background are not present in the query background, our model outperforms DCAMA. This validates that our method appropriately suppresses unnecessary support background. Additional in-depth analysis of it is provided in Section 5.4. QS TS m Io U - - 73.8 - 73.4 - 74.1 74.4 (a) Effect of two spatialwise scores. QS and TS denote the queryand targetrelevant scores. Method Metric AA SF&QF 0.213 SB&QB 0.886 Avg. 0.550 SF&QF 0.325 SB&QB 0.820 Avg. 0.573 (b) Averaged Attention score (AA) (See Section 5.3). Table 3: Results of ablation studies 5 Further Analysis In this section, we conduct ablation studies and provide an in-depth analysis of our method. For most ablation studies, we use the PASCAL-5i dataset in the 1-shot scenario with Swin-B as the backbone network, except for Section 5.2. Additionally, m Io U is adopted for metric which is one of the most standard metrics in few-shot segmentation. 5.1 Ablation Study on Main Components Table 3a presents the effects of queryand target-relevant scores in fold-1. In the second row, the query-relevant score reduces m Io U by 0.4%. We argue that the query-relevant score can inherently possess a negative effect by emphasizing some of the support backgrounds similar to the target object. Therefore, suppressing the support background with only the query-relevant score might be harmful to performance. Thus, the query-relevant score should be used in conjunction with the target-relevant score. On the contrary, as shown in the third row, the target-relevant score enhances the baseline by 0.3%. It means that some support background similar to the target object resembles the query foreground, making feature suppression advantageous. Importantly, adopting both scores can further boost performance, affirming their complementary nature. 5.2 Varying K for K-shot We validated the merits of our method using standard evaluation protocol with 1 and 5 labeled images (i.e., K = 1 and 5). We also conducted experiments with varying numbers of labeled images and the results are provided in Fig. 6. As reported, the benefits of our method are especially highlighted at low shots in terms of relative performance improvements. It confirms that our method is notably more effective in scenarios with severe data scarcity, aligning well with the requirements of few-shot learning. 5.3 Quality of Feature Matching Following our motivation, TBS suppresses the taskdisruptive features, enhancing the similarity between query and support foreground (QF and SF), and also between query and support background (QB and SB). To analyze these changes in similarity caused by TBS, we examine the cross-attention map in the segmentation model (DCAMA). Specifically, we average the attention scores corresponding to the QF & SF pairs to capture their similarity, and perform the same computation for the QB & SB pairs as well. As shown in Table 3b, we observe that the proposed method achieves a higher averaged attention score compared The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Query Support 𝑅B (a) None-object (b) Irrelevant-objects Query Support 𝑅B (c) Relevant-objects Query Support 𝑅B Figure 5: Visualization of the background-relevant scores before and after the score refinement module. The color from red to blue represents high to low scores. RB denotes score maps before refinement, while RB denotes score maps after refinement. 1 2 3 4 5 6 7 8 9 10 DCAMA DCAMA + Ours Number of Shot Figure 6: K-shot results with varying K. to the baseline. This observation implies that ours successfully improves the similarity between foreground objects. However, for the attention score of background pairs, ours achieve lower score compared to baseline. We suspect that unintended suppression of useful background occurs, resulting in lower attention scores of background pairs. Nevertheless, we verify the improvement when averaging the scores of both pairs and highlight the significant enhancement in attention scores of foreground pairs. 5.4 Visualization of Background-relevant Score We visualize background-relevant scores under diverse conditions to verify TBS in mitigating task-disruptive support background regions, as shown in Fig. 5. In the first scenario where objects do not exist in support background, as demonstrated in Fig. 5 (a), TBS only restrains the support object boundaries that are treated as background. The next scenario is when objects within the support background are not present in the query background. As shown in Fig. 5 (b), TBS assigns low scores to these regions, demonstrating the impact of query-relevant scoring. In the last scenario, shared objects are present in both the support and query backgrounds. These shared objects act as helpful features since they enhance the similarity between query and support backgrounds. As depicted in Fig. 5 (c), we verify that TBS grants high scores to leverage these features for segmentation. More importantly, TBS exhibits its effectiveness even in scenarios where relevant and irrelevant objects coexist within the support background (last row of Fig. 5 (c)). In such a case, we expect that the segmentation model may exploit only the relevant objects while suppressing the irrelevant ones to enhance background similarity. As shown in the last row of Fig. 5 (c), although the support backgrounds contain both people and a dog, only the dog is suppressed, as people are present in the query image while the dog is not. To sum up, TBS effectively suppresses task-disruptive support backgrounds in various conditions. Additionally, we compare background-relevant scores before and after convolution block to analyze the effectiveness of the refinement module. Notably, scores before refinement tend to have small variations between background-relevant and -irrelevant objects, potentially hindering effective suppression due to a subtle difference in suppression power. In contrast, the refined score maps present a wider spectrum of magnitudes, decisively influencing the determination of whether pixels in the support background warrant suppression. Furthermore, without the refinement module, the score distribution across the query-support image pairs exhibits significant diversity. This disparity is evident when comparing the background score distributions from Fig. 5 (a) and (c). Conversely, the distribution of scores after refinement attains consistency across the query-support pairs, enhancing their stability as input for the segmentation model. 6 Conclusion In this paper, we introduce a Task-Disruptive Background Suppression module designed to mitigate the problem that query-irrelevant and target-similar features in support background regions. To suppress such features, we present two types of scores. First, a query-relevant score is employed to filter out irrelevant support background pixels whose similarity between query images is low. Second, a target-relevant score is used for detecting support backgrounds that are similar to the support foreground. Based on these two score maps, we could suppress task-disruptive backgrounds in the support set. Finally, experiments conducted on standard benchmarks show the effectiveness of our model. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Acknowledgments This work was supported in part by MSIT/IITP (No. 20220-00680, 2019-0-00421, 2020-0-01821, 2021-0-02068), and MSIT&KNPA/KIPo T (Police Lab 2.0, No. 210121M06). References Azad, R.; Fayjie, A. R.; Kauffmann, C.; Ben Ayed, I.; Pedersoli, M.; and Dolz, J. 2021. On the texture bias for few-shot cnn segmentation. In Proceedings of the IEEE/CVF winter conference on applications of computer vision, 2674 2683. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; and Yuille, A. L. 2017. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE transactions on pattern analysis and machine intelligence, 40(4): 834 848. Doersch, C.; Gupta, A.; and Zisserman, A. 2020. Crosstransformers: spatially-aware few-shot transfer. Advances in Neural Information Processing Systems, 33: 21981 21993. Hariharan, B.; Arbel aez, P.; Girshick, R.; and Malik, J. 2014. Simultaneous detection and segmentation. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part VII 13, 297 312. Springer. Hong, S.; Cho, S.; Nam, J.; Lin, S.; and Kim, S. 2022. Cost Aggregation with 4D Convolutional Swin Transformer for Few-Shot Segmentation. In European Conference on Computer Vision, 108 126. Li, G.; Jampani, V.; Sevilla-Lara, L.; Sun, D.; Kim, J.; and Kim, J. 2021. Adaptive prototype learning and allocation for few-shot segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 8334 8343. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Doll ar, P.; and Zitnick, C. L. 2014. Microsoft coco: Common objects in context. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, 740 755. Springer. Liu, Y.; Zhang, X.; Zhang, S.; and He, X. 2020. Part-Aware Prototype Network for Few-Shot Semantic Segmentation. In European Conference on Computer Vision, 142 158. Long, J.; Shelhamer, E.; and Darrell, T. 2015. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, 3431 3440. Lu, Z.; He, S.; Zhu, X.; Zhang, L.; Song, Y.-Z.; and Xiang, T. 2021. Simpler is better: Few-shot semantic segmentation with classifier weight transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 8741 8750. Min, J.; Kang, D.; and Cho, M. 2021. Hypercorrelation squeeze for few-shot segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, 6941 6952. Nguyen, K.; and Todorovic, S. 2019. Feature weighting and boosting for few-shot segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 622 631. Peng, B.; Tian, Z.; Wu, X.; Wang, C.; Liu, S.; Su, J.; and Jia, J. 2023. Hierarchical Dense Correlation Distillation for Few-Shot Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 23641 23651. Seong, H. S.; Moon, W.; Lee, S.; and Heo, J.-P. 2023. Leveraging Hidden Positives for Unsupervised Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 19540 19549. Shaban, A.; Bansal, S.; Liu, Z.; Essa, I.; and Boots, B. 2017a. One-shot learning for semantic segmentation. ar Xiv preprint ar Xiv:1709.03410. Shaban, A.; Bansal, S.; Liu, Z.; Essa, I.; and Boots, B. 2017b. One-shot learning for semantic segmentation. ar Xiv preprint ar Xiv:1709.03410. Shi, X.; Wei, D.; Zhang, Y.; Lu, D.; Ning, M.; Chen, J.; Ma, K.; and Zheng, Y. 2022. Dense cross-query-and-support attention weighted mask aggregation for few-shot segmentation. In European Conference on Computer Vision, 151 168. Springer. Wang, H.; Zhang, X.; Hu, Y.; Yang, Y.; Cao, X.; and Zhen, X. 2020. Few-shot semantic segmentation with democratic attention networks. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XIII 16, 730 746. Springer. Wang, K.; Liew, J. H.; Zou, Y.; Zhou, D.; and Feng, J. 2019. Panet: Few-shot image semantic segmentation with prototype alignment. In proceedings of the IEEE/CVF international conference on computer vision, 9197 9206. Wang, Y.; Sun, R.; and Zhang, T. 2023. Rethinking the Correlation in Few-Shot Segmentation: A Buoys View. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 7183 7192. Williams, E. M. V. G. L. 2010. CK Winn J Zisserman A The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis, 88(2): 303. Zhang, G.; Kang, G.; Yang, Y.; and Wei, Y. 2021. Few-shot segmentation via cycle-consistent transformer. Advances in Neural Information Processing Systems, 34: 21984 21996. Zhang, X.; Wei, Y.; Yang, Y.; and Huang, T. S. 2020. Sg-one: Similarity guidance network for one-shot semantic segmentation. IEEE transactions on cybernetics, 50(9): 3855 3865. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; and Jia, J. 2017. Pyramid scene parsing network. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2881 2890. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)