# multilabel_testtime_adaptation_with_bound_entropy_minimization__626bccd1.pdf Published as a conference paper at ICLR 2025 MULTI-LABEL TEST-TIME ADAPTATION WITH BOUND ENTROPY MINIMIZATION Xiangyu Wu1,2, Feng Yu1, Qing-Guo Chen2, Yang Yang1 , Jianfeng Lu1 1Nanjing University of Science and Technology 2Alibaba International Digital Commerce Group Mainstream test-time adaptation (TTA) techniques endeavor to mitigate distribution shifts via entropy minimization for multi-class classification, inherently increasing the probability of the most confident class. However, when encountering multi-label instances, the primary challenge stems from the varying number of labels per image, and prioritizing only the highest probability class inevitably undermines the adaptation of other positive labels. To address this issue, we investigate TTA within multi-label scenario (ML TTA), developing Bound Entropy Minimization (BEM) objective to simultaneously increase the confidence of multiple top predicted labels. Specifically, to determine the number of labels for each augmented view, we retrieve a paired caption with yielded textual labels for that view. These labels are allocated to both the view and caption, called weak label set and strong label set with the same size k. Following this, the proposed BEM considers the highest top-k predicted labels from view and caption as a single entity, respectively, learning both view and caption prompts concurrently. By binding top-k predicted labels, BEM overcomes the limitation of vanilla entropy minimization, which exclusively optimizes the most confident class. Across the MSCOCO, VOC, and NUSWIDE multi-label datasets, our ML TTA framework equipped with BEM exhibits superior performance compared to the latest SOTA methods, across various model architectures, prompt initialization, and varying label scenarios. The code is available at https://github.com/Jinx630/ML-TTA. 1 INTRODUCTION The advent of vision-language models (VLMs) (Radford et al., 2021; Li et al., 2023; Zeng et al., 2024; Yang et al., 2024a) has facilitated remarkable generalization capabilities by pretraining on massive datasets. Nonetheless, VLMs such as CLIP (Radford et al., 2021), require sophisticated prompt learning when confronted with considerable discrepancies between training and testing domains, to prevent performance degradation due to distribution shifts occurring during testing time. Fortunately, recent advancements (Shu et al., 2022; Feng et al., 2023; Ma et al., 2023; Liu et al., 2024b; Zhang et al., 2024b; Zhao et al., 2024a; Karmanov et al., 2024; Yoon et al., 2024; Gao et al., 2024) allow for immediate adaptation to any distribution of test instance during testing time, which is known as Test-Time Adaptation (TTA). As pioneering works, TPT (Shu et al., 2022) and its enhancement, Diff TPT (Feng et al., 2023), select a set of confident augmented views, learning instance-level prompt for each test instance. DART (Liu et al., 2024b) and DMN (Zhang et al., 2024b), to fully utilize the encountered knowledge from past samples, design dual-modal knowledge retention prompts and dynamic dual-memory networks, respectively, to adaptively incorporate historical knowledge. The central premise of these methods is entropy minimization, which aims to minimize inconsistency and uncertainty over the model predictions, and further increase the prediction probability of the highest confidence class, a theory that is readily demonstrable. Although entropy loss is advantageous for TTA as an uncertainty metric, a natural question arises: Can it be reliably applied to instances with multiple positive labels? As illustrated in Figure 1 (a), for the positive label set {keyboard, phone, remote, mouse, book}, compared to CLIP, all methods consistently boost the probability of the most confident class, keyboard. Nonetheless, TPT (Shu Corresponding author Published as a conference paper at ICLR 2025 keyboard book 0 mouse remote phone people vase bottle clock refrigerator RLCF ML-TTA RLCF ML-TTA logit logit RLCF ML-TTA 1 3 5 10 Number of labels per image Positive labels of image (a). Changes of output logits compared to CLIP (b). Results on varying number of labels Figure 1: (a). Compared to CLIP (Radford et al., 2021), ML TTA increases all positive label logits simultaneously, while others focus only on top-1 class. (b). Comparison of various methods on images with varying numbers. Compared to CLIP, as the number of labels per image rises, the adaptability of TPT (Shu et al., 2022) and RLCF (Zhao et al., 2024a) in handling multi-label images shows a marked decrease. et al., 2022) and RLCF (Zhao et al., 2024a) adversely impair the remaining positive labels. This indicates that existing TTA methods primarily focus on increasing the confidence of top-1 label, leading to insufficient adaptation for other positive labels. Given this, we expect to treat the highest top-k positive labels as a single label, aiming to simultaneously increase the predicted confidence of multiple top-k labels. However, positive label sets are not known in advance in real applications. Based on the preceding discussion, we investigate the TTA within multi-label scenario (ML TTA) and propose a novel theoretical optimization objective named Bound Entropy Minimization (BEM), which posits that when the highest top-k predicted labels (k being the size of positive label set) share identical probabilities, the entropy loss will uniformly increase the probabilities of all top-k classes. Consider a multi-label test image with a set of augmented views, to determine the number of positive labels for each view, we retrieve a paired caption with derived textual labels for each view, which then serves as weak label set of size k for the corresponding view. Furthermore, owing to the aligned visual-language space of CLIP (Radford et al., 2021), texts can be treated as pseudo-images with known positive labels, a premise corroborated by recent academic research (Guo et al., 2023; Zhao et al., 2024b; Li et al., 2024a; Wu et al., 2024). Drawing inspiration from these findings, we conceptualize each paired caption as a pseudo-view possessing a known label set, termed strong label set, of the same size k, since the textual labels are directly derived from captions. Upon determining the weak label set for each view and the strong label set for each paired caption, the proposed BEM objective binds the highest top-k predicted labels as a single label for both view and caption. By optimizing the view prompt and caption prompt, the model is encouraged to concurrently increase the confidence of the top-k classes. Additionally, since some augmented views and paired captions may fail to capture the target label area, leading to misleading predictions, we adopt confidence selection utilized in TPT (Shu et al., 2022) to filter out noisy views and captions with high entropy (i.e., low confidence). Consequently, in this paper, starting from TPT, the developed ML TTA framework equipped with BEM endows the CLIP s adaptability of multi-label instances during testing. Our contributions are summarized as follows: We examine the Multi-Label Test-Time Adaptation (ML TTA) and propose Bound Entropy Minimization (BEM), which simultaneously increase the probabilities of all highest top labels. BEM binds weak label set of view and strong label set of the caption as a single label, respectively, learning instance-level view and caption prompts for adapting multi-label test instances. On the MSCOCO, VOC, and NUSWIDE datasets, ML TTA outperforms the original CLIP model as well as other state-of-the-art TTA methods designed for multi-class classification, across various model architectures, prompt initialization, and varying label scenarios. 2 RELATED WORK 2.1 TEST-TIME ADAPTION Test-time adaptation (TTA) (Zhang et al., 2022; Shu et al., 2022; Ma et al., 2023; Karmanov et al., 2024; Zhao et al., 2024a; Lee et al., 2024; Chi et al., 2024; Ma et al., 2024) enables models to adapt Published as a conference paper at ICLR 2025 changing distributions during testing time without accessing to the source domain data or extensive target domain data. Within the spectrum of TTA settings, e.g., fully TTA (Wang et al., 2021; Zhao et al., 2023), online TTA (Lee & Chang, 2024; Lee et al., 2024), continuous TTA (Liu et al., 2024a; Song et al., 2023), and prior TTA (Wei et al., 2023; 2024), online TTA (Shu et al., 2022; Karmanov et al., 2024; Zhao et al., 2024a) focuses on adapting to individual samples and is particularly valuable in many application domains, such as autonomous driving, where weather conditions are constantly changing, and road monitoring, where traffic patterns are continually evolving. MEMO (Zhang et al., 2022) is the pioneering work that proposes consistent predictions across diverse augmented views. Following this, TPT (Shu et al., 2022) notably enhances the generalization capabilities of the CLIP (Radford et al., 2021) model to unseen test data by entropy minimization. Swap Prompt (Ma et al., 2023) utilizes online and target prompts, enhancing the CLIP s adaptability by preserving historical information and alternating prediction. In contrast, TDA (Karmanov et al., 2024) adapts to streaming input data by constructing a dynamic key-value cache from historical data. RLCF (Zhao et al., 2024a) incorporates reinforcement learning to distill knowledge into more compact models. Among these works, MEMO (Zhang et al., 2022), TPT (Shu et al., 2022), and RLCF (Zhao et al., 2024a) are particularly challenging, as the model is reset after adapting a test instance, obviating the need to retain historical knowledge, and thereby accommodating continuously shifting test distributions. Nonetheless, these methods are primarily designed for multi-class classification and may not be as effective in the more common multi-label scenario. 2.2 PROMPT LEARNING IN VLMS Visual-language models (VLMs) (Li et al., 2021; Wu et al., 2022; Yang et al., 2024b; Li et al., 2023; Wan et al., 2024; Zeng et al., 2024; Huang et al., 2024), trained on massive image-text pairs (Sharma et al., 2018; Schuhmann et al., 2022), have demonstrated remarkable proficiency in cross-task learning. To further enhance the transfer abilities of CLIP (Radford et al., 2021), researchers have developed various prompt learning techniques (Zhou et al., 2022b;a; Fu et al., 2024; Li et al., 2024b; Wu et al., 2024). For instance, the groundbreaking work Co Op (Zhou et al., 2022b), and its advancement Co Co Op (Zhou et al., 2022a), are the first to propose optimizing context vectors to improve the generalization capabilities of CLIP. Maple (Khattak et al., 2023) introduces a multimodal prompt learning method, designed to recalibrate both visual and language modalities. Dept (Zhang et al., 2024a) and Prompt KD (Li et al., 2024b) take on the challenge from the perspectives of knowledge retention and distillation, respectively, to promote robust generalization on novel tasks. Exploiting the aligned visual-language space of CLIP (Radford et al., 2021), TAI-DPT (Guo et al., 2023), PVP (Wu et al., 2024) and RC-TPL (Zhao et al., 2024b) propose to regard texts as images for prompt tuning in zero-shot multi-label image classification. Investigations like Dual Co Op (Sun et al., 2022), Dual Co Op++ (Hu et al., 2023), and VLPL (Xing et al., 2024) consider more intricate tasks, enhancing multi-label classification capabilities in the partial-label scenario. In contrast, our study focuses on a training-free paradigm, termed multi-label test-time adaptation, which obviates the need for the source training data and is exclusively at the testing instance level. In Sec. 3.1, we review the entropy minimization widely used in TTA. In Sec. 3.2, we highlight the issue that vanilla entropy minimization predominantly increases the probability of top-1 predicted label and propose a new proposition Bound Entropy Minimization (BEM). In Sec. 3.3, we present a Multi-Label Test-Time Adaptation (ML TTA) framework, incorporating BEM, which binds the highest top predicted labels of both augmented views and paired captions as an individual single label. ML TTA consists of view-caption constructing (Sec. 3.3.1) and label binding (Sec. 3.3.2). 3.1 PRELIMINARIES The purpose of Test-Time Adaptation is to utilize each test instance once for immediate adaptation before inference, without any prior assumptions about the test data distribution. For the TTA of VLMs, let Mθ denote the CLIP model trained on the training dataset Dtrain = {(xtrain i , ytrain i ) | xtrain i X train, ytrain i Ytrain}M train i=1 . The TTA approach, TPT (Shu et al., 2022), incorporates the Marginal Entropy Minimization (MEM) objective to adapt Mθ using a solitary instance xtest from the testing dataset Dtest = {(xtest i , ytest i ) | xtest i X test, ytest i Ytest}M test i=1 . Published as a conference paper at ICLR 2025 Given a test instance xtest and a set A of N random augmentation functions, xtest is first augmented N times to generate a set of different views, represented as Xtest = {xtest j | xtest j = Aj(xtest)}N j=1. TTA aims to minimize the marginal entropy of these augmented views, encouraging the model to perform consistent and confident predictions. The entropy of an augmented view is defined as: H(p( |xtest j )) = l=1 p(y = l|xtest j ) log(p(y = l|xtest j ), (1) where l Ytest and L is the number of labels in Ytest. The core principle of TPT is to minimize the marginal entropy of the prediction probability distributions of selected confident augmented views by a ratio τ, thereby encouraging the model to make consistent predictions. After obtaining the average entropy of these confident views, denoted as H, TPT updates the prompt using a single gradient descent step based on H and performs immediate inference on this test instance. Once inference is done, the model s prompt and optimizer are reset promptly for adaptation to the next test instance. Owing to its simplicity and effectiveness, Marginal Entropy Minimization has emerged as a de facto standard in modern TTA. 3.2 BOUND ENTROPY MINIMIZATION It can be observed that the TPT method selects a subset of confident augmented views with lower entropy (i.e., high confidence) from Xtest, continually minimizing the average entropy of these confident views to maintain consistent model predictions across these views. With respect to vanilla entropy minimization within TTA, the following proposition holds. Proposition 1. Consider the output logits of a confident view x, denoted as s = (s1, s2, . . . , s L), where, without loss of generality, we assume s1 > s2 > > s L. It can be deduced that the entropy loss H = H(p( |x)) decreases as s1 increases, and H increases as the sum of the remaining logits, Srest = PL i=2 si, decreases. Formally, this relationship can be expressed as: s1 < 0 and srest H = H Srest > 0. (2) A detailed proof is provided in the Appendix. Following a single gradient descent update step, we can derive s(t+1) 1 = s(t) 1 α s1H and S(t+1) rest = S(t) rest α Srest H, where α denotes the learning rate. Therefore, Proposition 1 indicates that the nature of entropy loss is to increase the probability of the most confident class while diminishing the cumulative probability of the rest classes. Hence, when adapting to single-label test instances, the goal of vanilla entropy minimization is to solely maximize the probability of the top-1 predicted label, disregarding changes in the probabilities of the remaining labels. In contrast, in the context of multi-label test-time adaptation, where the test instance may include a set of positive labels Lp = {lp1, lp2, ..., lpk}. In this case, regardless of whether the top-1 predicted label is the element of the positive label set Lp, the entropy loss will inevitably decrease the prediction probabilities of the other positive labels within Lp while increasing the probability of the most confident class. This may lead to the model overemphasizing the top-1 predicted label and inadequately adapting to the other positive labels. Therefore, for test-time adaptation in multi-label data, we propose the following proposition, termed Bound Entropy Minimization. Proposition 2. Consider the output logits of a confident view x, denoted as s = (s1, s2, . . . , s L), where, without loss of generality, we assume s1 > s2 > > s L. We define the modified logits as s = (s 1, s 2, . . . , s L), where s i = ai +si for i k with ai = s1 si and s i = si for i > k. Here, ai is a constant value that does not participate in differentiation, resulting in s i = s1 for all i k. Let Srest = PL i=k+1 si. For the modified logits s , we define the modified probability p = Softmax(s ), and the modified entropy as H = PL i=1 p i log p i. It follows that: si < 0, i k and srest H = H Srest > 0. (3) A detailed proof is provided in Appendix. Likewise, after one step of gradient descent optimization, the prediction probabilities of all top-k predicted labels will further increase due to si H < 0 for all i <= k and srest H > 0. Therefore, from Proposition 2, to be robust against distribution shifts with multiple labels, it is crucial to determine the number of positive labels for adapting multi-label test instances. In the following subsection, we will introduce a novel Multi-Label Test-Time Adaptation framework by employing proposition 2 and incorporating text captions into the adaptation system. Published as a conference paper at ICLR 2025 Text Description Base A big dog sits next to a cat. ... A cat sits on a gray table. Vase and cup is in front of bed. Nouns Filter Labels Count Counts: 2 Counts: 2 Counts: 3 ... Weak Label Binding Strong Label Binding 2 1 3 1 2 1 2 View Prompt Caption Prompt Multi-label Image s1(x1) s2(x2) s N(x N) s1(t1) s2(t2) s N(t N) s1(x1) s2(x2) s N(y|x N) s1(t1) s2(t2) s N(y|t N) : Frozen Parameters : Frozen Parameters : Tuned Parameters : Tuned Parameters Figure 2: Overview of proposed multi-label test-time adaption. 3.3 MULTI-LABEL TEST-TIME ADAPTATION 3.3.1 VIEW-CAPTION CONSTRUCTING Benefiting from the aligned space of CLIP, any image can be assigned a most similar caption from an offline text description base based on similarity retrieval. As depicted in Figure 2, given a test image xtest and a collection of random augmentation functions A = {A1, A2, ..., AN}, xtest is first augmented N times to generate a set of different augmented views. For each augmented view, we retrieve the most similar caption from an offline text description database to serve as its paired caption. The views generating and caption allocating can be expressed as: Xtest = {xtest i | xtest i = Ai(xtest)}N i=1, T test = {ttest i | ttest i = Ri(xtest i )}N i=1, (4) where Ai and Ri represents augmentation and retrieval by computing similarity. To streamline the retrieval process, we directly utilize the method proposed in PVP (Wu et al., 2024), which employs LLama-2-7B (Touvron et al., 2023) to construct the text description base, each text is a description of a natural scene containing several categories. Then, CLIP is used to extract text embeddings and construct an offline database of size B d, where B denotes the number of test descriptions and d denotes the embedding dimension. More details of the text description base construction are provided in the appendix. The goal of TTA is to calibrate the model for a single unlabeled test instance. Clearly, a single instance is insufficient for tuning the entire CLIP model to learn domain-specific knowledge. Consequently, as shown in Figure 2, akin to prompt tuning paradigm, we design two identical prompts, referred to as view prompt and caption prompt, denoted by V and C, respectively. Treating prompt tuning at test-time as a way to furnish customized context for individual test instances. Benefiting from the aligned space of CLIP, the representations of images and texts share similar semantic information, therefore, the paired caption can be considered as a pseudo image with accurate textual labels, encouraging the model to learn visual-related knowledge and complementary information from views and captions jointly. For L categories, we initialize the view and caption prompts with template a photo of a [CLS]j , in which [CLS]j represents the j-th label name, e.g., dog or cat, results in vj and cj. Once the paired views and captions are obtained, we compute the logits for each view xtest i on L view prompts and for each caption ttest i on L caption prompts as below: sxtest ij = Enc I(xtest i ), Enc T(vj) , sttest ij = Enc T(ttest i ), Enc T(cj) , (5) where Enc I and Enc T represent the frozen image encoder and text encoder of CLIP, , signifies the dot product. As stated in proposition 2, the crux of adapting multi-label instance lies in identifying the size of positive label set for each view xtest i and caption ttest i . Published as a conference paper at ICLR 2025 Algorithm 1: Label Binding Algorithm Input: Logits si before label binding and the size of weak label set kxi. Output: Modified logits si after label binding. 1 mi = maxj sij ; 2 for j = 1 to L do 3 aij = detach (mi sij) Detach from gradient. ; 4 if Rank(sij,si) kxi then 5 sij = aij + sij Bind sij if j-th label is in highest top-kxi predicted labels. ; 8 sij = sij ; 11 si = ( si0, si1, , si L) 3.3.2 LABEL BINDING Obviously, the positive label set for xtest i is not feasible to obtain directly. Fortunately, the textual labels for ttest i , which we refer to as strong label set, can be readily derived through noun filtering, e.g., A truck drives past a black car with a suitcase on top. with extracted strong label set being truck, car, suitcase. Moreover, this set can also serve as a pseudo-positive label set, termed the weak label set, for xtest i . Consequently, we treat the size of strong label set as the top-k bound highest logits of captions, akin to views. The binding operation for sxtest ij and sttest ij can be expressed as: sxtest ij =((mxtest i sxtest ij )+sxtest ij ) I(Rank(sxtest ij ,sxtest i ) kxtest i )+sxtest ij I(Rank(sxtest ij ,sxtest i ) >kxtest i ), sttest ij =((mttest i sttest ij )+sttest ij ) I(Rank(sttest ij ,sttest i ) kttest i )+sttest ij I(Rank(sttest ij ,sttest i ) >kttest i ), (6) where ((mxtest i sxtest ij ) + sxtest ij ) employs stop-gradient operation follow VQ-VAE van den Oord et al. (2017), sxtest i = (sxtest i1 , sxtest i2 , ..., sxtest i L ) and sttest i = (sttest i1 , sttest i2 , ..., sttest i L ) denotes the logits before binding, mxtest i and mttest i denotes the maximum logit of sxtest i and sttest i , respectively, I( ) denotes the indicator function, and Rank(s, s) indicates the descending rank of s within s, kxtest i and kttest i denotes the size of weak label set of i-th augmented view and strong label set of i-th paired caption. The algorithm process of label binding is presented in algorithm 1. We provide a detailed label binding process using a 3-class classification task in the Appendix. To reduce the noise brought by random augmentation and the noise in the caption caused by noisy views, we employ confidence selection to filter out noisy views and captions with higher entropy (i.e., lower confidence). Such noisy views may, due to random cropping augmentation, exclude the target label area, leaving only irrelevant background information. Similarly, the retrieved paired captions for these noisy views will lack any pertinent textual labels. We selected views and captions with lower predicted entropy by a ratio τ, yielding {ˇxtest i }τN i=1 for views and {ˇttest i }τN i=1 for captions. Taking views ˇxtest i as an example, the probability of ˇxtest i on L labels denoted as p = Softmax( sˇxtest i i ), the average predicted entropy of the filtered low-entropy views can be expressed as: H ˇxtest avg = 1 τN l=1 p(y = l|ˇxtest i ) log(p(y = l|ˇxtest i )) Subsequently, the bound entropy optimization objective of view prompt V is to minimize the predicted entropy through H ˇxtest avg . For the objective of caption prompt C, we replace ˇxtest i in Eq.(7) with confident captions ˇttest i to obtain Hˇttest avg . 3.3.3 OVERALL OBJECTIVE OF ML TTA ML TTA calculates the predicted bound entropy of confident augmented views and paired captions, optimizing both view prompt and caption prompt with a single step of gradient descent, and simultaneously increasing the probability of highest top predicted labels. Then, the overall bound entropy Published as a conference paper at ICLR 2025 loss is given by: HBEM = H ˇxtest avg + Hˇttest avg . (8) After optimizing the prompts, ML TTA immediately infers the test instance xtest and resets the parameters of the prompts (V and C) and the state of optimizer to adapt to the next test instance. During the inference phase, we separately compute the similarity between the view prompt V and the test instance xtest, as well as the similarity between the caption prompt C and the test instance xtest, and directly add these two similarities to obtain the final prediction result. 4 EXPERIMENT 4.1 EXPERIMENTAL SETUP Benchmarks. We utilize the widely employed CLIP (Radford et al., 2021) model as source model and select the multi-label datasets VOC (Everingham et al., 2010), MSCOCO (Lin et al., 2014), and NUSWIDE (Chua et al., 2009) as target domains. The VOC dataset includes 20 categories, covering both VOC2007 and VOC2012 versions, which contain 4,952 and 5,823 test images, respectively. The MSCOCO dataset extends the category range to 80, and for testing purposes, we employ the validation sets of COCO2014 with 40,504 images and COCO2017 with 5,000 images, as the test set labels are not accessible. The NUSWIDE dataset includes 81 categories with a total of 83,898 test images of lower resolution, which presents a broader category spectrum than MSCOCO. Implementation details. All experiments are based on the CLIP model, encompassing RN50, RN101, Vi T-B/32, and Vi T-B/16 architectures, each consisting of an image encoder and a corresponding text encoder. For the initialization of the view and caption prompts, we employ the token embedding of the a photo of a hard prompt as initialization weights and another using learned prompts from Co Op (Zhou et al., 2022b) and Ma PLE (Khattak et al., 2023). The learning rate for the view prompt is 1e-2, while for the caption prompt is 1e-3. For all settings, multi-label testtime adaptation is performed on a single instance, i.e., the batch size is 1. The ratio for filtering confident views and captions is 0.1. The optimizer is Adam W (Loshchilov & Hutter, 2019) with a single update step, followed by immediate inference on the test instance. Following PVP (Wu et al., 2024), we collect 100k text descriptions for each dataset, resulting in a total size of 300k text description base. All experiments are evaluated by the mean Average Precision (m AP) metric, defined as m AP = 1 L PL i=1 APi, where L is the number of categories, and APi is the area under the Precision-Recall curve for the i-th category. 4.2 COMPARISONS WITH STATE-OF-THE-ART To our knowledge, our work is the first to investigate the feasibility of traditional entropy minimization in the multi-label setting. Therefore, in this section, we select the original CLIP model and other SOTA methods for multi-class scenarios as baselines, including episdoic methods that do not require retaining historical knowledge (TPT Shu et al. (2022), Diff TPT Feng et al. (2023), RCLF Zhao et al. (2024a)) and online methods that do (DMN Zhang et al. (2024b), TDA Karmanov et al. (2024)). Results on different architectures. Table 1 compares ML TTA with both online and episdoic TTA methods on different CLIP (Radford et al., 2021) architectures, demonstrating the superior performance across various multi-label datasets. Specifically, for the RN50 and RN101 architectures on COCO2014/2017 (Lin et al., 2014) datasets, ML TTA achieves 4 5% improvement in m AP over the original CLIP (Radford et al., 2021) model, whereas TPT (Shu et al., 2022) and Diff TPT (Feng et al., 2023) yield only 1% enhancement. Despite introducing dual-memory network knowledge from historical samples, DMN (Zhang et al., 2024b) and TDA Karmanov et al. (2024) present a slight performance decline, due to intensifying the optimization bias towards top-1 label. Notably, RLCF (Zhao et al., 2024a) employs a reinforcement learning-based knowledge distillation and more adaptation steps, resulting in a catastrophic degradation in the multi-label adaptation performance for smaller models due to excessive optimizations of top-1 label. On the VOC2012/2017 (Everingham et al., 2010) datasets, TPT and Diff TPT also show 1 2% decrease in performance compared to CLIP, whereas ML TTA still maintains 2 3% performance improvement, indicating the robustness of ML TTA in multi-label adaptation across various model architectures and datasets. Published as a conference paper at ICLR 2025 Table 1: Comparison with CLIP and SOTAs on adapting multi-label instances with different architectures. Methods Epsdoic COCO2014 COCO2017 VOC2007 VOC2012 NUSWIDE Average CLIP [ICML 2022] 47.53 47.32 75.91 74.25 41.53 57.31 DMN [CVPR 2024] 44.54 44.18 74.87 74.13 41.32 55.81 TDA [CVPR 2024] 48.91 49.11 76.64 75.12 42.34 58.42 TPT [Neur IPS 2022] 48.52 48.51 75.54 73.92 41.97 57.69 DIff TPT [ICCV 2023] 48.56 48.67 75.89 74.13 41.33 57.72 RLCF [ICLR 2024] 36.87 36.73 65.75 64.73 29.83 46.78 ML TTA (Ours) 51.58 51.39 78.62 76.63 42.53 60.15 CLIP [ICML 2022] 48.83 48.15 76.72 74.21 41.93 57.97 DMN [CVPR 2024] 46.28 45.44 76.82 75.32 42.71 57.31 TDA [CVPR 2024] 50.19 49.78 78.12 77.13 43.13 59.67 TPT [Neur IPS 2022] 49.71 48.89 74.82 73.39 43.10 57.98 DIff TPT [ICCV 2023] 49.45 49.19 74.98 74.31 42.93 58.17 RLCF [ICLR 2024] 40.53 39.79 71.21 69.63 31.77 50.59 ML TTA (Ours) 52.92 52.24 78.72 78.13 43.62 61.13 CLIP [ICML 2022] 50.31 50.15 77.18 76.85 42.90 59.48 DMN [CVPR 2024] 49.32 48.13 77.42 76.60 43.41 58.98 TDA [CVPR 2024] 51.23 51.49 77.62 77.12 44.13 60.32 TPT [Neur IPS 2022] 48.12 48.63 74.21 71.93 43.63 57.30 DIff TPT [ICCV 2023] 48.73 49.19 74.50 72.98 43.42 57.76 RLCF [ICLR 2024] 50.28 49.59 77.12 76.83 43.29 59.42 ML TTA (Ours) 52.83 52.99 78.70 77.97 44.12 61.32 CLIP [ICML 2022] 54.42 54.13 79.58 79.25 45.65 62.61 DMN [CVPR 2024] 52.52 52.37 79.83 79.67 46.27 62.13 DART [AAAI 2024] 54.73 54.68 79.91 78.56 45.91 62.76 TDA [CVPR 2024] 55.21 55.46 80.12 79.92 46.72 63.49 TPT [Neur IPS 2022] 53.32 54.20 77.54 77.39 46.15 61.72 DIff TPT [ICCV 2023] 53.91 54.15 77.93 77.24 46.13 61.87 RLCF [ICLR 2024] 54.21 54.43 79.29 79.26 43.18 62.07 ML TTA (Ours) 57.52 57.49 81.28 81.13 46.55 64.80 For the vision transformer series architectures, compared to CLIP (Radford et al., 2021), ML TTA consistently achieves 2 4% m AP improvement on the COCO2014/2017 (Lin et al., 2014) and VOC2007/2012 (Everingham et al., 2010) datasets. However, most TTA methods, except TDA (Karmanov et al., 2024) and DART (Liu et al., 2024b), exhibit a slight performance decrement, particularly among episodic methods. Additionally, we observed an intriguing observation: all TTA methods, excluding RLCF (Zhao et al., 2024a), fail to substantially enhance the m AP performance of CLIP (Radford et al., 2021) on the NUSWIDE (Chua et al., 2009) dataset, with an improvement of merely about 1%. This may be attributed to the low image resolution of NUSWIDE dataset, where random data augmentation struggles to preserve sufficient visual information. Consequently, adapting to multi-labels for small targets may become a research topic in the future. Results on different prompt initialization. For this comparison, we adopt the learned prompt from Co Op (Zhou et al., 2022b) and Maple (Khattak et al., 2023) to initialize the prompt weights, replacing the template a photo of a [CLS] employed in the original CLIP model. As shown in Table 2, the application of both Co Op and Maple prompt weights in our ML TTA framework results in a significant enhancement of over 4% in m AP on the COCO2014/2017 datasets. For instance, the m AP increases from 47.53% to 51.58% and from 47.32% to 51.39% on COCO2014/2017 with Co Op prompt initialization, whereas other episdoic methods, TPT (Shu et al., 2022) and Diff TPT (Feng et al., 2023), yield improvements of no more than 1.5%. Moreover, ML TTA also surpasses Published as a conference paper at ICLR 2025 Table 2: Comparison with SOTAs on adapting multi-label instances with different prompt initialization. Methods Epsdoic COCO2014 COCO2017 VOC2007 VOC2012 NUSWIDE Average Co Op [IJCV2022] 56.12 56.35 79.14 77.85 46.74 63.24 TDA [CVPR 2024] 56.93 57.15 80.20 78.58 47.82 64.13 TPT [Neur IPS 2022] 55.35 55.23 79.72 77.85 47.27 63.08 DIff TPT [ICCV 2023] 55.30 55.47 79.86 77.61 47.13 63.07 RLCF [ICLR 2024] 56.72 56.18 80.15 78.24 47.62 63.78 ML TTA (Ours) 59.68 59.33 83.17 81.36 48.12 66.33 Maple [CVPR2023] 62.18 62.35 85.34 84.79 48.42 68.62 TDA [CVPR 2024] 63.25 63.64 85.76 84.15 49.55 69.27 TPT [Neur IPS 2022] 63.36 63.75 85.04 83.92 48.90 69.01 DIff TPT [ICCV 2023] 62.93 63.14 85.15 83.78 48.81 68.76 RLCF [ICLR 2024] 62.84 62.90 85.35 85.28 49.37 69.15 ML TTA (Ours) 64.75 64.86 86.40 85.69 50.21 70.38 TDA (Karmanov et al., 2024), which is designed by dynamically employing the historical sample knowledge, on both Co Op and Maple prompt initialization across all datasets. Table 3: Results on different label counts. Methods {1,2} {3,4} {5,6,7} {>=8} CLIP [ICML 2022] 62.76 55.41 49.89 41.07 TPT [Neur IPS 2022] 62.88 53.05 45.57 37.43 Diff TPT [ICCV 2023] 61.97 52.67 44.32 36.89 RLCF [ICLR 2024] 66.01 51.65 43.32 35.08 ML TTA (Ours) 67.14 57.59 51.68 41.32 Results on different label counts. Apart from the analysis of architecture and prompt initialization weights, we explore a more challenging scenario in Table 3, where the COCO2014 dataset is divided into subsets with incrementally increasing numbers of image labels per part, e.g., {1,2} represents the number of labels L per image is either 1 or 2. When L {1, 2}, TPT achieves only a negligible improvement compared to CLIP and shows large considerable performance degradation in other situations as well as Diff TPT. RLCF improves significantly when L {1, 2}, but its performance sharply declines as L increases. In contrast, our ML TTA framework outperforms CLIP across all situations, demonstrating that ML TTA not only can address the distribution shifts during testing but also effectively handle varying numbers of labels in testing instances. Table 4: Results on adaptation complexity. Methods TPT Diff TPT RLCF ML-TTA Adapting Time 0.21s 0.41s 0.45s 0.24s m AP 48.52 48.56 36.87 51.58 Results on adaptation complexity. Furthermore, we analyze adapting time per test instance with methods that also do not require retaining historical knowledge. Table 4 shows that ML-TTA presents a significant advantage compared to Diff TPT, which involves generating multiple pseudo-images via a diffusion model, and RLCF, which requires distillation from a teacher model along with more gradient update steps. Compared to the benchmark TPT, ML-TTA increases adapting time due to simultaneous optimizing view and caption prompts. 4.3 ABLTION STUDIES. Table 5: Ablation studies of different components. Methods RN50 Vi T-B/16 COCO2014 VOC2007 COCO2014 VOC2007 VP (i.e., TPT) 48.51 75.52 53.32 77.57 VP+BEM 48.96 76.31 53.58 77.89 CP 49.12 76.16 55.14 78.93 CP+BEM 49.54 76.75 55.64 79.58 VP+CP 51.22 77.98 57.14 80.85 VP+CP+BEM 51.58 78.62 57.52 81.28 Different components. In Table 5, we discuss the effectiveness of different components within our proposed ML TTA framework on the COCO2014 and VOC2007 datasets, including view prompt (VP, i.e., TPT (Shu et al., 2022)), caption prompt (CP), and Bound Entropy Minimization (BEM). Across both RN50 and Vi T-B/16 architectures, BEM consistently enhances the m AP performance of VP, CP, and VP+CP, which indicates the reasonable effectiveness of our proposed Bound Entropy Minimization objective. Furthermore, we observe that CP and Published as a conference paper at ICLR 2025 CP+BEM always achieve superior performance compared to VP and VP+BEM in all settings. Such phenomenon shows treating text as a pseudo-image with a known label set to adapt multi-label test instance is more reliable than augmented views, as the positive label set of views is pseudo. 4.4 FURTHER ANALYSIS Table 6: Comparison with binary cross-entropy loss. Methods RN50 Vi T-B/16 COCO2014 VOC2007 COCO2014 VOC2007 CLIP 47.53 75.91 54.42 79.58 VP+CP+BCE 48.39 75.75 54.51 78.59 VP+CP+BEM 51.58 78.62 57.52 81.28 8 16 32 64 128 Number of augmented views TPT RLCF ML-TTA TPT RLCF ML-TTA Figure 3: Results on different number of views. Loss functions. Here, we conduct a comparison between Bound Entropy Minimization (BEM) and the conventional binary cross-entropy (BCE) loss function in multi-label classification tasks. Specifically, we regarded the weak label set of confident views as hard labels for those views and the strong label set of confident captions as hard labels for those captions, then using BCE loss to optimize the view and caption prompts. The results are shown in Table 6. Compared to CLIP, the m AP improvement using BCE loss on the COCO2014 is less than 1%. In contrast, our BEM objective surpasses BCE loss by 3 4% in m AP across all benchmarks, which demonstrates BEM is not only more effective than vanilla entropy minimization but also more robust compared to binary cross-entropy loss. BCE loss is not suitable for optimizing a single test instance. Number of augmented views. Following TPT (Shu et al., 2022), we conduct parameter experiments of different numbers of augmented views on the COCO2014 dataset using Vi T-B/16 architecture. As depicted in Figure 3, as the number of views increases from 1 to 128, the m AP performance of RLCF and ML TTA both show an upward trend and begin to stabilize at 64 views. Surprisingly, the performance curve of TPT does not have any regularity, which implies that vanilla entropy minimization, by focusing only on the label with the highest probability, leads to unstable adaptation for multi-label instances. Table 7: Results on different numbers of retrieved captions. Datasets CLIP TPT 1 2 4 8 16 32 64 128 COCO2014 47.53 48.52 51.35 51.37 51.41 51.49 51.58 51.59 51.55 51.48 VOC2007 75.91 75.54 78.29 78.33 78.48 78.54 78.61 78.59 78.53 78.42 COCO2014 54.42 53.32 57.23 57.33 57.41 57.48 57.49 57.52 57.55 57.58 VOC2007 79.58 77.54 81.06 81.12 81.21 81.24 81.28 81.19 81.15 80.98 Number of retrieved captions. We also investigate the impact of allocating different numbers of retrieved captions for each augmented view on the performance of ML TTA. As shown in Table 7, when only one caption is allocated to each view, ML TTA outperforms CLIP or TPT by 3 4%. As the number of captions increases, the performance of ML TTA gradually improves until it stabilizes. For the VOC2007 dataset, too many captions can lead to a slight decrease in performance, as captions that are not highly similar to the view may introduce noisy positive labels that do not exist in the corresponding view. 5 CONCLUSION In this paper, we investigate a test-time adaptation framework (ML TTA) designed for multi-label data without making any presumptions about the distribution of the test instances. The proposed Bound Entropy Minimization (BEM) objective overcomes the limitation of the vanilla entropy loss, which only optimizes the most confident class. By conceptualizing paired captions as pseudo-views with a known label set, ML TTA employs BEM to adapt to multi-label test instances by allocating weak label set to each augmented view and strong label set to each paired caption, binding the topk predicted labels with the highest probabilities. Extensive experiments on the MSCOCO, VOC, and NUSWIDE datasets demonstrate that ML TTA framework outperforms the source model CLIP and other state-of-the-art test-time adaptation methods, across various model architectures, prompt initialization, and varying label scenarios. Published as a conference paper at ICLR 2025 ACKNOWLEDGEMENTS This work was supported in part by Key Laboratory of Target Cognition and Application Technology (2023-CXPT-LC-005). Zhixiang Chi, Li Gu, Tao Zhong, Huan Liu, Yuanhao Yu, Konstantinos N. Plataniotis, and Yang Wang. Adapting to distribution shift by visual domain prompt generation. In ICLR. Open Review.net, 2024. Tat-Seng Chua, Jinhui Tang, Richang Hong, Haojie Li, Zhiping Luo, and Yantao Zheng. Nus-wide: a real-world web image database from national university of singapore. In CIVR, 2009. Mark Everingham, Luc Van Gool, Christopher K. I. Williams, John M. Winn, and Andrew Zisserman. The pascal visual object classes voc challenge. IJCV, 88(2):303 338, 2010. Chun-Mei Feng, Kai Yu, Yong Liu, Salman Khan, and Wangmeng Zuo. Diverse data augmentation with diffusions for effective test-time prompt tuning. In ICCV, pp. 2704 2714, 2023. Zhongtian Fu, Kefei Song, Luping Zhou, and Yang Yang. Noise-aware image captioning with progressively exploring mismatched words. In AAAI, pp. 12091 12099, 2024. Junyu Gao, Xuan Yao, and Changsheng Xu. Fast-slow test-time adaptation for online vision-andlanguage navigation. In ICML. Open Review.net, 2024. Zixian Guo, Bowen Dong, Zhilong Ji, Jinfeng Bai, Yiwen Guo, and Wangmeng Zuo. Texts as images in prompt tuning for multi-label image recognition. In CVPR, pp. 2808 2817, 2023. Ping Hu, Ximeng Sun, Stan Sclaroff, and Kate Saenko. Dualcoop++: Fast and effective adaptation to multi-label recognition with limited annotations. TPAMI, 2023. Longfei Huang, Xiangyu Wu, Jingyuan Wang, Weili Guo, and Yang Yang. Refining Visual Perception for Decoration Display: A self-enhanced deep captioning model. In ACML, volume 260, pp. 527 542, 05 08 Dec 2024. Adilbek Karmanov, Dayan Guan, Shijian Lu, Abdulmotaleb El Saddik, and Eric Xing. Efficient test-time adaptation of vision-language models. In CVPR, pp. 14162 14171, 2024. Muhammad Uzair Khattak, Hanoona Abdul Rasheed, Muhammad Maaz, Salman H. Khan, and Fahad Shahbaz Khan. Maple: Multi-modal prompt learning. In CVPR, pp. 19113 19122, 2023. Jae-Hong Lee and Joon-Hyuk Chang. Stationary latent weight inference for unreliable observations from online test-time adaptation. In ICML, 2024. Jonghyun Lee, Dahuin Jung, Saehyung Lee, Junsung Park, Juhyeon Shin, Uiwon Hwang, and Sungroh Yoon. Entropy is not enough for test-time adaptation: From the perspective of disentangled factors. In ICLR, 2024. Junnan Li, Ramprasaath R. Selvaraju, Akhilesh Gotmare, Shafiq R. Joty, Caiming Xiong, and Steven Chu-Hong Hoi. Align before fuse: Vision and language representation learning with momentum distillation. In Neur IPS, pp. 9694 9705, 2021. Junnan Li, Dongxu Li, Silvio Savarese, and Steven C. H. Hoi. Blip-2: Bootstrapping languageimage pre-training with frozen image encoders and large language models. In ICML, volume 202, pp. 19730 19742, 2023. Yiming Li, Xiangdong Wang, and Hong Liu. Audio-free prompt tuning for language-audio models. In ICASSP, pp. 491 495, 2024a. Zheng Li, Xiang Li, Xinyi Fu, Xin Zhang, Weiqiang Wang, Shuo Chen, and Jian Yang. Promptkd: Unsupervised prompt distillation for vision-language models. In CVPR, pp. 26617 26626, 2024b. Published as a conference paper at ICLR 2025 Tsung-Yi Lin, Michael Maire, Serge J. Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll ar, and C. Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, volume 8693, pp. 740 755, 2014. Jiaming Liu, Senqiao Yang, Peidong Jia, Renrui Zhang, Ming Lu, Yandong Guo, Wei Xue, and Shanghang Zhang. Vida: Homeostatic visual domain adapter for continual test time adaptation. In ICLR, 2024a. Zichen Liu, Hongbo Sun, Yuxin Peng, and Jiahuan Zhou. Dart:dual-modal adaptive online prompting and knowledge retention for test-time adaptation. In Artificial Intelligence, pp. 14106 14114, 2024b. Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In ICLR, 2019. Jiawei Ma, Po-Yao Huang, Saining Xie, Shang-Wen Li, Luke Zettlemoyer, Shih-Fu Chang, Wen Tau Yih, and Hu Xu. Mode: Clip data experts via clustering. In CVPR, pp. 26354 26363, 2024. Xiaosong Ma, Jie Zhang, Song Guo, and Wenchao Xu. Swapprompt: Test-time prompt adaptation for vision-language models. In Neur IPS, 2023. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. Learning transferable visual models from natural language supervision. In ICML, volume 139, pp. 8748 8763, 2021. Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, et al. Laion-5b: An open large-scale dataset for training next generation image-text models. In Neur IPS, volume 35, pp. 25278 25294, 2022. Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut. Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. In ACL, pp. 2556 2565, 2018. Manli Shu, Weili Nie, De-An Huang, Zhiding Yu, Tom Goldstein, Anima Anandkumar, and Chaowei Xiao. Test-time prompt tuning for zero-shot generalization in vision-language models. In Neur IPS, 2022. Junha Song, Jungsoo Lee, In So Kweon, and Sungha Choi. Ecotta: Memory-efficient continual test-time adaptation via self-distilled regularization. In CVPR, pp. 11920 11929, 2023. Ximeng Sun, Ping Hu, and Kate Saenko. Dualcoop: Fast adaptation to multi-label recognition with limited annotations. In Neur IPS, volume 35, pp. 30569 30582, 2022. Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton-Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing Ellen Tan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aur elien Rodriguez, Robert Stojnic, Sergey Edunov, and Thomas Scialom. Llama 2: Open foundation and fine-tuned chat models. ar Xiv, 2023. A aron van den Oord, Oriol Vinyals, and Koray Kavukcuoglu. Neural discrete representation learning. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp. 6306 6315, 2017. Published as a conference paper at ICLR 2025 Fengqiang Wan, Xiangyu Wu, Zhihao Guan, and Yang Yang. Covlr: Coordinating cross-modal consistency and intra-modal relations for vision-language retrieval. In ICME, pp. 1 6, 2024. Dequan Wang, Evan Shelhamer, Shaoteng Liu, Bruno A. Olshausen, and Trevor Darrell. Tent: Fully test-time adaptation by entropy minimization. In ICLR, 2021. Jiaheng Wei, Harikrishna Narasimhan, Ehsan Amid, Wen-Sheng Chu, Yang Liu, and Abhishek Kumar. Distributionally robust post-hoc classifiers under prior shifts. In ICLR, 2023. Tong Wei, Zhen Mao, Zi-Hao Zhou, Yuanyu Wan, and Min-Ling Zhang. Learning label shift correction for test-agnostic long-tailed recognition. In ICML, 2024. Xiangyu Wu, Jianfeng Lu, Zhuanfeng Li, and Fengchao Xiong. Ques-to-visual guided visual question answering. In ICIP, pp. 4193 4197, 2022. Xiangyu Wu, Qing-Yuan Jiang, Yang Yang, Yi-Feng Wu, Qing-Guo Chen, and Jianfeng Lu. Tai++:text as image for multi-label image classification by co-learning transferable prompt. In IJCAI, 2024. Xin Xing, Zhexiao Xiong, Abby Stylianou, Srikumar Sastry, Liyu Gong, and Nathan Jacobs. Visionlanguage pseudo-labels for single-positive multi-label learning. In CVPR, pp. 7799 7808, 2024. Yang Yang, Fengqiang Wan, Qing-Yuan Jiang, and Yi Xu. Facilitating multimodal classification via dynamically learning modality gap. In Neur IPS, 2024a. Yang Yang, Wenjuan Xi, Luping Zhou, and Jinhui Tang. Rebalanced vision-language retrieval considering structure-aware distillation. TIP, 33:6881 6892, 2024b. Hee Suk Yoon, Eunseop Yoon, Joshua Tian Jin Tee, Mark A. Hasegawa-Johnson, Yingzhen Li, and Chang D. Yoo. C-tpt: Calibrated test-time prompt tuning for vision-language models via text feature dispersion. In ICLR. Open Review.net, 2024. Yan Zeng, Xinsong Zhang, Hang Li, Jiawei Wang, Jipeng Zhang, and Wangchunshu Zhou. X2-vlm: All-in-one pre-trained model for vision-language tasks. TPAMI, 46(5):3156 3168, 2024. Ji Zhang, Shihan Wu, Lianli Gao, Heng Tao Shen, and Jingkuan Song. Dept: Decoupled prompt tuning. In CVPR, pp. 12924 12933, 2024a. Marvin Zhang, Sergey Levine, and Chelsea Finn. Memo: Test time robustness via adaptation and augmentation. In Neur IPS, 2022. Yabin Zhang, Wenjie Zhu, Hui Tang, Zhiyuan Ma, Kaiyang Zhou, and Lei Zhang. Dual memory networks: A versatile adaptation approach for vision-language models. In CVPR, pp. 28718 28728, 2024b. Bowen Zhao, Chen Chen, and Shu-Tao Xia. Delta: Degradation-free fully test-time adaptation. In ICLR, 2023. Shuai Zhao, Xiaohan Wang, Linchao Zhu, and Yi Yang. Test-time adaptation with clip reward for zero-shot generalization in vision-language models. In ICLR, 2024a. Xiongjun Zhao, Zheng-Yu Liu, Fen Liu, Guanting Li, Yutao Dou, and Shaoliang Peng. Reportconcept textual-prompt learning for enhancing x-ray diagnosis. In ACM MM, 2024b. Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu. Conditional prompt learning for vision-language models. In CVPR, pp. 16795 16804, 2022a. Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu. Learning to prompt for visionlanguage models. IJCV, 130(9):2337 2348, 2022b. Published as a conference paper at ICLR 2025 Appendix for Multi-Label Test-Time Adaptation with Bound Entropy Minimization A.1 PROOF OF PROPOSITION 1 Proposition 1. Consider a model s output logits of a selected view x, denoted as s = (s1, s2, . . . , s L), where without loss of generality, we assume s1 > s2 > > s L. It follows that the entropy loss H = H(p( |x)) decreases as s1 increases, and H increases as the sum of the remaining logits, Srest = PL i=2 si, decreases. Formally, this can be written as: H s1 < 0 and H Srest > 0. (9) Proof. We denote the predicted probability p(y = l|x) = exp sl PL i=1 exp si as pl for simplicity, where si is the logit of the i-th category. We first calculate the partial derivative of si with respect to pl: exp sl PL j=1 exp sj = δl,i exp sl PL j=1 exp sj exp sl exp si PL j=1 exp sj 2 = δl,ipl plpi where δi,j = 1 only if i = j, else is δi,j = 0. We can now directly calculate the partial derivative of H for si. l=1 pl log pl si log pl + pl 1 pl l=1 (δl,ipl log pl plpi log pl + δl,ipl plpi) = pi log pi + pi l=1 ( plpi log pl plpi) = (pi log pi + pi) l=1 ( plpi log pl plpi) l=1 (plpi log pi plpi log pl + plpi plpi) l=1 plpi log pi where the fourth equivalent uses the property of δi,j and fifth equivalent uses PL l=1 pl = 1. Since we assume s1 > s2 > > s L, then the probabilities have the same order p1 > p2 > > p L, therefor: l=1 plp1 log p1 l=2 plp1 log p1 pl < 0 (12) Published as a conference paper at ICLR 2025 pl > 0 for all l > 1, therefor we proof the first inequality in proposition 1. To prove the second inequality, we first calculate the sum of the partial derivative of H for all logits. l=1 plpi log pi l=1 (plpi log pi plpi log pl) l=1 (plpi log pi pipl log pi) where we change the position of index i and l for the second term in the double summation to get the third equivalent. Now the second inequality is easy to get: A.2 PROOF OF PROPOSITION 2 proposition 2. Consider a model s output logits of a selected view x, denoted as s = (s1, s2, . . . , s L), where without loss of generality, we assume s1 > s2 > > s L. We define the modified logits as s = (s 1, s 2, . . . , s L), where s i = ai + si for i k with ai = s1 si and s i = si for i > k. Here, ai is a constant that does not participate in differentiation, resulting in s i = s1 for all i k. Let Srest = PL i=k+1 si. For the modified logits s , we define the modified probability p = Softmax(s ), and the modified entropy as H = PL i=1 p i log p i. It follows that: si < 0, i k and H Srest > 0. (15) Proof. With the assumption s1 > s2 > > s L and the definition of s , we have s 1 = s 2 = = s k > > s L. Use the conclusion in proposition 1, for i k, we have: ds i dsi = H l=1 p lp i log p i p l = l=k+1 p lp i log p i p l < 0 where p i > p l for i k and l > k as s 1 = s 2 = = s k > > s L. Similar to the proof of proposition 1, we use the conclusion of PL i=1 H s i = 0, which has been proved in Equation. 13 to Published as a conference paper at ICLR 2025 prove the second inequality. B DETAILED LABEL BINDING PROCESS. In this section, we present a certain example to showcase the calculation of Label Binding 3.3.2. Label binding refers to making the top-k predicted logits equal, as expressed below: sxtest ij = ((mxtest i sxtest ij )+sxtest ij ) I(Rank (sxtest ij , sxtest i ) kxtest i )+sxtest ij I(Rank (sxtest ij , sxtest i ) > kxtest i ), (18) Since label binding (making ... equal) is non-differentiable, we employ the stop-gradient operation in VQ-VAE van den Oord et al. (2017) for backpropagation, i.e. ((mxtest i sxtest ij ) + sxtest ij ) to perform label binding. Taking a 3-class classification task as an example with class labels of (1, 2, 3), assuming kxtest i is 2, and the label binding process is s = [0.9, 0.7, 0.3] s = [0.9, 0.9, 0.3]. sxtest ij represents the logit of the j-th class in the i-th augmented view after label binding, e.g., sxtest i2 changes from 0.7 0.9. mxtest i denotes the maximum value of s, which is 0.9. I( ) is the indicator function. Rank (a, b) indicates the descending rank of a within b, e.g., Rank (0.7, s) = 2. The process for computing the bound logit for each class is as follows: sxtest i1 = ((0.9 0.9) + 0.9) I(Rank (0.9, s) 2) + 0.9 I(Rank (0.9, s) > 2) = 0.9 I(1 2) + 0.9 I(1 > 2) = 0.9, sxtest i2 = ((0.9 0.7) + 0.7) I(Rank (0.7, s) 2) + 0.7 I(Rank (0.7, s) > 2) = 0.9 I(2 2) + 0.7 I(2 > 2) = 0.9, sxtest i3 = ((0.9 0.3) + 0.3) I(Rank (0.3, s) 2) + 0.3 I(Rank (0.3, s) > 2) = 0.9 I(3 2) + 0.3 I(3 > 2) = 0.3, label binding process changes the logits from [0.9, 0.7, 0.3] [0.9, 0.9, 0.3]. C TEXT DESCRIPTION BASE CONSTRUCTION Here, we present the construction of the text description base using Large language models (LLMs). Initially, for a set of labels, denoted as L = {l1, l2, ..., l L}, where L represents the total number of labels across all multi-label datasets. Following PVP (Wu et al., 2024), we define a prompt template to instruct LLama-2-7B (Touvron et al., 2023), generating descriptions that describe a nature scene, which is as follows: PROMPT: Make a sentence to describe a photo. Requirements: Each sentence should be less than 15 words and include keywords: {li1, li2, . . . , lij}, where {li1, li2, . . . , lij} is a subset of L with i 5. We randomly sample j categories from L and input these categories along with the prompt template into LLMs to automatically generate text Published as a conference paper at ICLR 2025 descriptions. After obtaining generated descriptions, we employ the nouns filtering strategy used in PVP to extract textual labels for each description. Some examples are illustrated below: 1. A hot dog toaster is positioned next to a stop sign. 2. A group of girls enjoying a game of frisbee while sitting on chairs. 3. The little boy dreams of becoming a pilot as he falls asleep with his aeroplane. 4. Remotes control the TV, allowing people to enjoy their favorite shows. 5. A motorbike speeds past a man wearing a tie, as he holds a wine glass in one hand. where the underlined words indicate the textual labels extracted from the corresponding description. However, due to the uncontrollable quality and relevance of the paired captions generated by LLMs, these captions may not always accurately represent the image contents. In real-world scenarios, besides adopting a confident-based filtering strategy to filter out views and captions with high entropy (e.g., low confidence), we can also explore more robust strategies to retrieve paired captions, such as, constructing high-quality and content-rich text description databases, ensembling label sets from multiple captions, or improving the similarity retrieval strategy, thereby reducing the impact of noise on the model s adaptation.