# episodic_multitask_learning_with_heterogeneous_neural_processes__b39b8c02.pdf Episodic Multi-Task Learning with Heterogeneous Neural Processes Jiayi Shen1, Xiantong Zhen1,2 , Qi (Cheems) Wang3, Marcel Worring1 1University of Amsterdam, Netherlands, {j.shen, m.worring}@uva.nl 2 Inception Institute of Artificial Intelligence, Abu Dhabi, UAE, zhenxt@gmail.com 3 Kaiyuan Mathematical Sciences Institute, Changsha, China, hhq123go@gmail.com This paper focuses on the data-insufficiency problem in multi-task learning within an episodic training setup. Specifically, we explore the potential of heterogeneous information across tasks and meta-knowledge among episodes to effectively tackle each task with limited data. Existing meta-learning methods often fail to take advantage of crucial heterogeneous information in a single episode, while multi-task learning models neglect reusing experience from earlier episodes. To address the problem of insufficient data, we develop Heterogeneous Neural Processes (HNPs) for the episodic multi-task setup. Within the framework of hierarchical Bayes, HNPs effectively capitalize on prior experiences as meta-knowledge and capture task-relatedness among heterogeneous tasks, mitigating data-insufficiency. Meanwhile, transformer-structured inference modules are designed to enable efficient inferences toward meta-knowledge and task-relatedness. In this way, HNPs can learn more powerful functional priors for adapting to novel heterogeneous tasks in each meta-test episode. Experimental results show the superior performance of the proposed HNPs over typical baselines, and ablation studies verify the effectiveness of the designed inference modules. 1 Introduction Deep learning models have made remarkable progress with the help of the exponential increase in the amount of available training data [1]. However, many practical scenarios only have access to limited labeled data [2]. Such data-insufficiency sharply degrades the model s performance [2, 3]. Both meta-learning and multi-task learning have the potential to alleviate the data-insufficiency issue. Meta-learning can extract meta-knowledge from past episodes and thus enables rapid adaptation to new episodes with a few examples only [4 7]. Meanwhile, multi-task learning exploits the correlation among several tasks and results in more accurate learners for all tasks simultaneously [8 11]. However, the integration of meta-learning and multi-task learning in overcoming the datainsufficiency problem is rarely investigated. In episodic training [4], existing meta-learning methods [4 7, 12, 13] in every meta-training or meta-test episode learn a single-task. In this paper, we refer to this conventional setting as episodic single-task learning. This setting restricts the potential for these models to explore task-relatedness within each episode, leaving the learning of multiple heterogeneous tasks in a single episode underexplored. We consider multiple tasks in each episode as episodic multi-task learning. The crux of episodic multi-task learning is to generalize the ability of exploring task-relatedness from metatraining to meta-test episodes. The differences between episodic single-task learning and episodic multi-task learning are illustrated in Figure 1. To be specific, we restrict the scope of the problem Currently with United Imaging Healthcare, Co., Ltd., China. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). Figure 1: Illustration of episodic multi-task learning. Each row corresponds to a meta-training or meta-test episode. Different colors represent different label spaces among episodes; the same color with different shades represents different categories in the same task. Compared with episodic single-task learning, episodic multi-task learning simultaneously handles several related tasks in a single episode. setup to the case where tasks in each meta-training or meta-test episode are heterogeneous but also relate to each other by sharing the same target space. The neural process (NP) family [12, 13], as typical meta-learning probabilistic models [14], efficiently quantifies predictive uncertainty with limited data, making it in principle well-suited for tackling the problem of data-insufficiency. However, in practice, it is challenging for vanilla NPs [12] with a global latent variable to encode beneficial heterogeneous information in each episode. This issue is also known as the expressiveness bottleneck [15, 16], which weakens the model s capacity to handle insufficient data, especially when faced with diverse heterogeneous tasks. To better resolve the data-insufficiency problem, we develop Heterogeneous Neural Processes (HNPs) for episodic multi-task learning. As a new member of the NP family, HNPs improve the expressiveness of vanilla NPs by introducing a hierarchical functional space with global and local latent variables. The remainder of this work is structured as follows: We introduce our method in Section (2). Related work is overviewed in Section (3). We report experimental results with analysis in Section (4), after which we conclude with a technical discussion, existing limitations, and future extensions. In detail, our technical contributions are two-fold: Built on the hierarchical Bayes framework, our developed HNPs can simultaneously generalize meta-knowledge from past episodes to new episodes and exploit task-relatedness across heterogeneous tasks in every single episode. This mechanism makes HNPs more powerful when encoding complex conditions into functional priors. We design transformer-structured inference modules to infer the hierarchical latent variables, capture task-relatedness, and learn a set of tokens as meta-knowledge. The designed modules can fuse the meta-knowledge and heterogeneous information from context samples in a unified manner, boosting the generalization capability of HNPs across tasks and episodes. Experimental results show that the proposed HNPs together with transformer-structured inference modules, can exhibit superior performance on regression and classification tasks under the episodic multi-task setup. 2 Methodology Notations 2. We will now formally define episodic multi-task learning. For a single episode τ, we consider M heterogeneous but related tasks I1:M τ = {Im τ }M m=1. Notably, the subscript denotes an episode, while superscripts are used to distinguish tasks in this episode. In the episodic multitask setup, tasks in a single episode are heterogeneous since they are sampled from different task distributions {p(Im)}M m=1, but are related at the same time as they share the target space Yτ. 2 For ease of presentation, we abbreviate a set {( )m}M m=1 as ( )1:M, where M is a positive integer. Likewise, {( )o}O o=1 is abbreviated as ( )1:O. For convenience, the notation table is provided in Appendix B. To clearly relate to the modeling of vanilla neural processes [12], this paper follows its nomenclature to define each task. Note that in vanilla neural processes context and target are often respectively called support and query in conventional meta-learning [4, 5]. Each task Im τ contains a context set with limited training data Cm τ = { xm τ,i, ym τ,i}NC i=1 and a target set T m τ = {xm τ,j, ym τ,j}NT j=1, where NC and NT are the numbers of context samples and target samples, respectively. xm τ,i and xm τ,j represent features of context and target samples; while ym τ,i, ym τ,j Yτ are their corresponding targets, where i = 1, 2, ..., NC; j = 1, 2, .., NT ; m = 1, 2, ..., M. For simplicity, we denote the set of target samples and their corresponding ground-truths by xm τ = {xm τ,j}NT j=1, ym τ = {ym τ,j}NT j=1. For an episode τ, episodic multi-task learning aims to perform simultaneously well on each corresponding target set T m τ , m = 1, 2.., M, given the collection of context sets C1:M τ . For classification, this paper follows the protocol of meta models [4, 5, 17], such as O-way K-shot setup, clearly suffering from the data-insufficiency problem. Thus, episodic multi-task classification can be cast as a M-task O-way K-shot supervised learning problem. An episode has M related classification tasks, and each of them has a context set with K different instances from each of the O classes [5]. It is worth mentioning that the target spaces of meta-training episodes do not overlap with any categories in those of meta-test episodes. 2.1 Modeling and Inference of Heterogeneous Neural Processes We now present the proposed heterogeneous neural process. The proposed model inherits the advantages of multi-task learning and meta-learning, which can exploit task-relatedness among heterogeneous tasks and extract meta-knowledge from previous episodes. Next, we characterize the generative process, clarify the modeling within the hierarchical Bayes framework, and derive the approximate evidence lower bound (ELBO) in optimization. Generative Processes. To get to our proposed method HNPs, we extend the distribution over a single function p(fτ) as used in vanilla NPs to a joint distribution of multiple functions p(f 1:M τ ) for all heterogeneous tasks in a single episode τ. In detail, the underlying multi-task function distribution p(f 1:M τ ) is inferred from a collection of context sets C1:M τ and learnable meta-knowledge ω, ν1:M. Note that ω represents the shared meta-knowledge for all tasks, and νm denotes the task-specific meta-knowledge corresponding to the task distribution p(Im). Hence, we can formulate the predictive distribution for every single episode as follows: p(T 1:M τ |C1:M τ ; ω, ν1:M) = Z p(y1:M τ |x1:M τ , f 1:M τ )p(f 1:M τ |C1:M τ ; ω, ν1:M)df 1:M τ , (1) where p(f 1:M τ |C1:M τ ; ω, νm) denotes the data-dependent functional prior for multiple tasks of the episode τ. The functional prior encodes context sets from all heterogeneous tasks and quantifies uncertainty in the functional space. Nevertheless, it is less optimal to characterize multi-task function generative processes with vanilla NPs, since the single latent variable limits the capacity of the latent space to specify the complicated functional priors. This expressiveness bottleneck in vanilla NPs is particularly severe for our episodic multi-task learning since each episode has diverse heterogeneous tasks with insufficient data. Figure 2: Graphical model of the proposed HNPs in a single episode. Filled shapes indicate observations. Probabilistic and deterministic variables are indicated by unfilled circles and diamonds, respectively. Modeling within the Hierarchical Bayes Framework. To mitigate the expressiveness bottleneck of vanilla NPs, we model HNPs by parameterizing each task-specific function within a hierarchical Bayes framework. As illustrated in Figure 2, HNPs integrate a global latent representation zm τ and a set of local latent parameters wm τ,1:O to model each task-specific function f m τ . Specifically, the latent variables are introduced at different levels: zm τ encodes task-specific context information from Cm τ and νm in the representation level. wm τ,1:O encode prediction-aware information for a task-specific decoder from C1:M τ and ω in the parameter level, where O is the dimension of the decoder. For example, the dimension is the size of the target space when performing classification tasks. Notably, each local latent parameter is conditioned on the global latent representation, which controls access to all context sets in the episode for the corresponding task. Our method differs from previous hierarchical architectures [16, 18 20] in the NP family since the local latent parameters of our HNPs are prediction-aware and explicitly constitute a decoder for the subsequent inference processes. In practice, we assume that distributions of each task-specific function are conditionally independent. Thus, with the introduced hierarchical latent variables for each task in the episode, we can factorize the prior distribution over multiple functions in Eq. (1) into: p(f 1:M τ |C1:M τ ; ω, ν1:M) = m=1 p(zm τ |Cm τ ; νm)p(wm τ,1:O|zm τ , C1:M τ ; ω), (2) where p(zm τ |Cm τ ; νm) and p(wm τ,1:O|zm τ , C1:M τ ; ω) are prior distributions of the global latent representation and the local latent parameters to induce the task-specific function distribution. By integrating Eq. (2) into Eq. (1), we rewrite the modeling of HNPs in the following form: p(T 1:M τ |C1:M τ ; ω, ν1:M) = Z n Z p(ym τ |xm τ , wm τ,1:O) p(wm τ,1:O|zm τ , C1:M τ ; ω)dwm τ,1:O o p(zm τ |Cm τ ; νm)dzm τ , where p(ym τ |xm τ , wm τ,1:O) is the function distribution for the task Im τ in HNPs. This distribution is obtained by the matrix multiplication of xm τ and all local latent parameters wm τ,1:O. Compared with most NP models [12, 16, 18, 19] employing only latent representations, HNPs infer both latent representations and parameters in the hierarchical architecture from multiple heterogeneous context sets and learnable meta-knowledge. Our model specifies a richer and more intricate functional space by leveraging the hierarchical uncertainty inherent in the context sets and meta-knowledge. This theoretically yields more powerful functional priors to induce multi-task function distributions. Moreover, we claim that the developed model constitutes an exchangeable stochastic process and demonstrate this via Kolmogorov Extension Theorem [21]. Please refer to Appendix B for the proof. Approximate ELBO. Since both exact functional posteriors and priors are intractable, we apply variational inference to the proposed HNPs in Eq. (3). This results in the approximate ELBO: LHNPs(ω, ν1:M, θ, ϕ) = Eqθ(zm τ |T m τ ;νm) n Eqϕ(wm τ,1:O|zm τ ,T 1:M τ ;ω)[log p(ym τ |xm τ , wm τ,1:O)] DKL[qϕ(wm τ,1:O|zm τ , T 1:M τ ; ω)||pϕ(wm τ,1:O|zm τ , C1:M τ ; ω)] o DKL[qθ(zm τ |T m τ ; νm)||pθ(zm τ |Cm τ ; νm)] , (4) where qθ(zm τ |T m τ ; νm) and qϕ(wm τ,1:O|zm τ , T 1:M τ ; ω) are variational posteriors of their corresponding latent variables. θ and ϕ are parameters of inference modules for zm τ and wm τ,1:O, respectively. Following the protocol of vanilla NPs [12], the priors use the same inference modules as variational posteriors for tractable optimization. In this way, the KL-divergence terms in Eq. (4) encourage all latent variables inferred from the context sets to stay close to those inferred from the target sets, enabling effective function generation with few examples. Details on the derivation of the approximate ELBO and its tractable optimization are attached in Appendix C. 2.2 Transformer-Structured Inference Module In order to infer the prior and variational posterior distributions in Eq. (4), it is essential to develop welldesigned approximate inference modules. This is non-trivial and closely related to the performance of HNPs. Here we adopt a transformer structure as the inference module to better exploit taskrelatedness from the meta-knowledge and the context sets in the episode. More specifically, the previously mentioned meta-knowledge ω = ω1:O and ν1:M are instantiated as learnable tokens to induce the distributions of hierarchical latent variables in the proposed model. Without loss of generality, in the next, we provide an example of transformer-structured inference modules for prior distributions in classification scenarios. Details of the inference modules in regression scenarios can be found in Appendix D. In Figure 3, a diagram of the transformer-structured Figure 3: A diagram of transformer-structured inference modules of HNPs for the first metatraining episode in Figure 1 under the 3-task 5-way 1-shot setting. For clarity, we display the inference process of the local latent parameters specific to the third task in the episode. inference modules is displayed under the 3-task 5-way 1-shot setting. In this case, the number of context samples is the same as the size of the target space, and thus we have Cm τ = { xm τ,o, ym τ,o}O o=1, where O is set as 5. In episodic training, labels in context sets are always available during inference. Transformer-Structured Inference Module {θ, νm} for zm τ . In the proposed HNPs, each global latent representation encodes task-specific information relevant to the considered task in the episode as pθ(zm τ |Cm τ ; νm). The learnable token νm preserves the meta-knowledge from previous episodes for specific tasks, which are sampled from the corresponding task distribution p(Im). The role of νm is to help the model adapt efficiently to such specific tasks in meta-test episodes. In detail, we set the dimension of the learnable token νm to the same as that of the features xm τ,1:O. Then the transformer-structured inference module θ fuses them in a unified manner by taking [ xm τ,1:O; νm] as the input. The module θ outputs the mean and variance of the corresponding prior distribution. The inference steps for the global latent representation zm τ are: [exm τ,1:O; eνm] = MSA(LN([ xm τ,1:O; νm])) + [ xm τ,1:O; νm], (5) [bxm τ,1:O; bνm] = MLP(LN([exm τ,1:O; eνm])) + [exm τ,1:O; eνm], (6) pθ(zm τ |Cm τ ; νm) = N(zm τ ; µzm τ , σzm τ ), (7) where µzm τ = MLP(bνm), σzm τ = Softplus(MLP(bνm)). The transformer-structured inference module includes a multi-headed self-attention (MSA) and three multi-layer perceptrons (MLP). The layer normalization (LN) is "pre-norm" as done in [22]. Softplus is the activation function to output the appropriate value as the variance of the prior distribution [23]. Transformer-Structured Inference Module {ϕ, ω1:O} for wm τ,1:O. Likewise, each learnable token ωo corresponds to a local latent parameter wm τ,o. With the learnable tokens ω1:O, we reformulate the prior distribution of local latent parameters as pϕ(wm τ,1:O|zm τ , C1:M τ ; ω1:O). In this way, we learn the shared knowledge, inductive biases across all tasks, and their distribution at a parameter level, which in practical settings can capture epistemic uncertainty. To be specific, the prior distribution can be factorized as QO o=1 pϕ(wm τ,o|zm τ , C1:M τ ; ωo), where all local latent parameters are assumed to be conditionally independent. For each local latent parameter wm τ,o, the transformer-structured inference module ϕ takes [ x1:M τ,o , ωo] as input and outputs the corresponding prior distribution, where x1:M τ,o are deep features from the same class o in the episode and ωo is the corresponding learnable token. Here the inference steps for wm τ,o are as follows: [ex1:M τ,o ; eωo] = MSA(LN([ x1:M τ,o ; ωo])) + [ x1:M τ,o ; ωo], (8) [bx1:M τ,o ; bωo] = MLP(LN([ex1:M τ,o ; eωo])) + [ex1:M τ,o ; eωo], (9) pϕ(wm τ,o|zm τ , C1:M τ ; ωo) = N(wm τ,o; µwm τ,o, σwm τ,o), (10) where µwm τ,o = MLP(bωo, zm τ (i)), σwm τ,o = Softplus(MLP(bωo, zm τ (i))). zm τ (i) is a Monte Carlo sample from the variational posterior of the corresponding global latent representation during meta-training. Both transformer-structured inference modules use the refined tokens bνm and bωo to obtain a global latent representation and a local latent parameter, respectively. The introduced tokens preserve the specific meta-knowledge for each latent variable during inference. Compared with the θ-parameterised inference module exploring the intra-task relationships, the ϕ-parameterised inference module enables the exploitation of the inter-task relationships to reason over each local latent parameter. Thus, the introduced tokens can be refined with relevant information from the heterogeneous context sets. By integrating meta-knowledge and heterogeneous context sets, HNPs can reduce the negative transfer of task-specific knowledge among heterogeneous tasks in each episode. Please refer to Appendix E for algorithms. 3 Related Work Multi-Task Learning. Multi-task learning can operate in various settings [9]. Here we roughly separate the settings of MTL into two branches: (1) Single-input multi-output (SIMO) [24 30], where tasks are defined by different supervision information for the same input. (2) Multi-input multi-output (MIMO) [11, 10, 31 34], where heterogeneous tasks follow different data distributions. This work considers the MIMO setup of multi-task learning with episodic training. In terms of modeling methods, from a processing perspective, existing MTL methods can be roughly categorized into two groups: (1) Probabilistic MTL methods [11, 19, 35 41], which employ the Bayes framework to characterize probabilistic dependencies among tasks. (2) Deep MTL models [10, 32, 24 26, 42 48], which directly utilize deep neural networks to discover information-sharing mechanisms across tasks. However, deep MTL models rely on large amounts of training data and tend to overfit when encountering the data-insufficiency problem. Meanwhile, previous probabilistic MTL methods consider a small number of tasks that occur at the same time, limiting their applicability in real-world systems. Meta-Learning. Meta-learning aims to find strategies to quickly adapt to unseen tasks with a few examples [49, 4, 5, 50]. There exist a couple of branches in meta-learning methods, such as metrics-based methods [6, 51 57] and optimization-based methods [5, 58 68]. Our paper focuses on a probabilistic meta-learning method, namely neural processes, that can quantify predictive uncertainty. Models in this family [7, 12, 13, 15, 16, 18, 69 73] can approximate stochastic processes in neural networks. Vanilla NPs [12] usually encounter the expressiveness bottleneck because their functional priors are not rich enough to generate complicated functions [15, 16]. [7] introduces deterministic variables to model predictive distributions for meta-learning scenarios directly. Most NP-based methods only focus on a single task during inference [7, 12, 15, 16, 14], which leaves task-relatedness between heterogeneous tasks in a single episode an open problem. This paper combines multi-task learning and meta-learning paradigms to tackle the data-insufficiency problem. Our work shares the high-level goal of exploiting task-relatedness in an episode with [19, 74, 75]. Concerning the multi-task scenarios, the main differences are: [19, 74, 75] handles multiple attributes and multi-sensor data under the SIMO setting, while our work performs for the MIMO setting where tasks are heterogeneous and distribution shifts exist. Moreover, [76] theoretically addresses the conclusion that MTL methods are powerful and efficient alternatives to gradient-based meta-learning algorithms. However, our method inherits the advantages of multi-task learning and meta-learning: simultaneously generalizing meta-knowledge from past to new episodes and exploiting task-relatedness across heterogeneous tasks in every single episode. Thus, our method is more suitable for solving the data-insufficiency problem. Intuitive comparisons with related paradigms such as cross-domain few-shot learning [77 82], multimodal meta-learning [83 87, 56] and cross-modality few-shot learning [88 90] are provided in Appendix A. 4 Experiments We evaluate the proposed HNPs and baselines on three benchmark datasets under the episodic multi-task setup. Sec. 4.1 and Sec. 4.2 provide experimental results for regression and classification, respectively. Ablation studies are in Sec. 4.3. More comparisons with recent works on extra datasets are provided in Appendix F. Additional results under the convectional MIMO setup without episodic training can be found in Appendix G & H. 4.1 Episodic Multi-Task Regression Dataset and Settings. To evaluate the benefit of HNPs over typical NP baselines in uncertainty quantification, we conduct experiments in several 1D regression tasks. The baselines include conditional neural processes (CNPs [13]), vanilla neural processes (NPs [12]), and attentive neural processes (ANPs [15]). As a toy example, we construct multiple tasks with different task distributions: each task s input set is defined on separate intervals without overlap. Figure 4: Performance comparisons on the episodic multi-task 1-D function regression using 5 context points (black dots) for each task. Black curves are ground truth, and blue ones are predicted results. The shadow regions are 3 standard derivations from the mean [18]. Given four different tasks in an episode, their input sets are x1:4 τ . Each input set contains a few instances, drawn uniformly at random from separate intervals, such as x1 τ [ 4, 2), x2 τ [ 2, 0), x3 τ [0, 2), and x4 τ [2, 4). All tasks in an episode are related by sharing the same ground truth function. Following [12, 18], function-fitting tasks are generated with Gaussian processes (GPs). Here a zero mean Gaussian process y(0) GP(0, k( , )) is used to produce y1:4 τ for the inputs from all tasks x1:4 τ . A radial basis kernel k(x, x ) = σ2 exp( (x x )2)/2l2), with l = 0.4 and σ = 1.0 is used. Results and Discussions. As shown in Figure 4, CNPs, ANPs, and our HNPs exhibit more reasonable uncertainty than NPs in Figure 4: lower variances are predicted around observed (context) points with higher variances around unobserved points. Furthermore, NPs and ANPs detrimentally impact the smoothness of the predicted curves, whereas HNPs yield smoother predictive curves with reliable uncertainty estimation. These observations suggest that integrating correlation information across related tasks and meta-knowledge in HNPs can improve uncertainty quantification in multi-task regression. Table 1: Average negative log-likelihoods over target points from all tasks. Methods CNPs NPs ANPs HNPs Avg. NLL 0.0935 0.8649 -0.1165 -0.5207 To quantify uncertainty we use the average negative log-likelihood (the lower, the better). As shown in Table 1, our HNPs achieve a lower average negative log-likelihood than baselines, demonstrating our method s effectiveness in uncertainty estimation. 4.2 Episodic Multi-task Classification Datasets and Settings. We use Office-Home [91] and Domain Net [92] as episodic multi-task classification datasets. Office-Home contains images from four domains: Artistic (A), Clipart (C), Product (P) and Real-world (R). Each domain contains images from 65 categories collected from office and home environments. Note that all domains share the whole target space. The numbers of meta-training classes and meta-test classes are 40 and 25, respectively. There are about 15, 500 images in total. Domain Net has six distinct domains: Clipart, Infograph, Painting, Quickdraw, Real and Sketch. It includes approximately 0.6 million images distributed over 345 categories. The Table 2: Comparative results (95% confidence interval) for episodic multi-task classification on Office-Home and Domain Net. Best results are indicated in bold. Office-Home Domain Net 4-task 5-way 4-task 20-way 6-task 5-way 6-task 20-way Method 1-shot 5-shot 1-shot 5-shot 1-shot 5-shot 1-shot 5-shot ERM [93] 66.04 0.61 73.62 0.55 39.25 0.24 47.14 0.18 59.95 0.52 68.52 0.44 38.62 0.22 47.85 0.20 VMTL [11] 49.71 0.48 65.75 0.47 27.50 0.14 42.82 0.13 42.24 0.39 57.37 0.43 18.05 0.11 31.38 0.15 MAML [5] 60.58 0.60 75.29 0.53 34.29 0.19 48.39 0.20 53.21 0.46 65.24 0.47 17.10 0.12 20.35 0.14 Proto. Net. [6] 57.19 0.53 74.97 0.46 32.72 0.18 49.75 0.16 53.71 0.48 68.80 0.42 31.90 0.19 47.59 0.18 DGPs [94] 65.89 0.53 79.96 0.38 31.48 0.18 49.46 0.18 50.93 0.42 63.32 0.38 25.46 0.15 38.63 0.17 CNPs [13] 43.33 0.56 55.07 0.63 10.57 0.10 12.02 0.11 37.90 0.45 40.53 0.44 5.12 0.10 5.14 0.10 NPs [12] 33.66 0.48 53.99 0.60 5.25 0.16 11.40 0.11 20.58 0.51 20.53 0.53 5.12 0.09 5.11 0.09 TNP-D [95] 65.49 0.53 78.94 0.43 41.61 0.22 59.19 0.21 49.10 0.42 67.39 0.40 28.83 0.17 47.69 0.18 HNPs 76.29 0.51 80.80 0.42 51.82 0.23 59.97 0.18 62.36 0.53 69.38 0.42 39.32 0.23 48.56 0.19 numbers of meta-training classes and meta-test classes are 276 and 69, respectively. Here one domain corresponds to a specific task distribution in the episodic multi-task setting. When it comes to the episodic multi-task classification, we compare HNPs with the following three branches: (1) Multi-task learning methods: ERM [93] directly expands the training set of the current task with samples of related tasks. VMTL [11] is one of the state-of-the-art under the MIMO setting of multi-task learning. (2) Meta-learning methods: MAML [5], Proto.Net [6] and DGPs [94] address each task separately with no mechanism to leverage task-relatedness in a single episode. (3) Methods from the NP family: CNPs [13] and NPs [12] are established methods in the NP family. TNP-D [95] is recent NP work in sequential decision-making for a single task in each episode. Results and Discussions. The experimental results for episodic multi-task classification on Office-Home and Domain Net are reported in Table 2. We use the average accuracy across all task distributions as the evaluation metric. It can be seen that HNPs consistently outperform all baseline methods, demonstrating the effectiveness of HNPs in handling each task with limited data under the episodic multi-task classification setup. NPs and CNPs do not work well under all episodic multi-task classification cases. This can be attributed to their limited expressiveness of the global representation and the weak capability to extract discriminative information from multiple contexts. In contrast, HNPs explicitly abstract discriminative information for each task in the episode with the help of local latent parameters, enhancing the expressiveness of the functional prior. We also find that HNPs significantly surpass other baselines on 1-shot Office-Home and Domain Net, both under the 4/6-task 5-way and 4/6-task 20-way settings. This further implies that HNPs can circumvent the effect of the problem of data-insufficiency by simultaneously exploiting task-relatedness across heterogeneous tasks and meta-knowledge among episodes. 4.3 Ablation Studies Influence of Hierarchical Latent Variables. We first investigate the roles of the global latent representation zm τ and the local latent parameters wm τ,1:O by leaving out individual inference modules. These experiments are performed on Office-home under the 4-task 5-way 1-shot setting. We report the detailed performance for tasks sampled from a single task distribution (A/C/P/R) and the average accuracy across all task distributions (Avg.) in Table 3. The variants without specific latent variables are included in the comparison by removing the corresponding inference modules. Table 3: Effectiveness of global latent representations zm τ and local latent parameters wm τ,1:O in the model. and denote whether the variants of HNPs have the corresponding latent variable or not. zm τ wm τ,1:O A C P R Avg. 62.64 0.72 56.87 0.71 75.18 0.79 73.68 0.77 67.09 0.63 69.39 0.60 63.10 0.61 80.66 0.67 79.99 0.62 73.29 0.51 67.02 0.67 60.70 0.69 78.26 0.76 78.47 0.72 71.11 0.59 73.31 0.63 64.92 0.68 83.38 0.66 83.54 0.64 76.29 0.51 As shown in Table 3, both zm τ and wm τ,1:O benefit overall performance. Our method with hierarchical latent variables performs 9.20% better than the variant without both latent variables, 3.00% better than the variant without zm τ , and 5.18% better than the variant without wm τ,1:O. This indicates that latent variables of HNPs complement each other in representing con- text sets from multiple tasks and meta-knowledge. The variant without wm τ,1:O underperforms the variant without zm τ by 2.18%, in terms of the average accuracy. This demonstrates that zm τ suffers more from the expressiveness bottleneck than wm τ,1:O, weakening the models discriminative ability. For classification, local latent parameters are more crucial than a global latent representation in revealing the discriminating knowledge from multiple heterogeneous context sets. Influence of Transformer-Structured Inference Modules. To further understand our transformerstructured inference modules (Trans. w learnable tokens), we examine the performance against two other options: inference modules that solely utilize a multi-layer perceptron (MLP) and the variants that do not incorporate any learnable tokens (Trans. w/o learnable tokens). We also compare the probabilistic and deterministic versions of such inference modules. The deterministic variants consider the deterministic embedding for the hierarchical latent variables. Table 4: Performance comparisons between our transformer inference modules (Trans. w learnable tokens) and other alternatives. Inference networks 1-shot 5-shot Deterministic MLP 64.93 0.66 72.39 0.56 Trans. w/o learnable tokens 70.22 0.62 76.15 0.54 Trans. w learnable tokens 70.61 0.56 76.70 0.50 Probabilistic MLP 73.30 0.59 77.94 0.48 Trans. w/o learnable tokens 75.25 0.55 80.42 0.47 Trans. w learnable tokens 76.29 0.51 80.80 0.42 As shown in Table 4, our inference modules consistently outperform the variants, regardless of whether the inference network is probabilistic or deterministic. When using the probabilistic one, our inference modules respectively achieve 1.04% and 2.99% performance gains over Trans. w/o learnable tokens and MLP under the 4-task 5-way 1-shot setting. This implies the importance of learnable tokens and task-relatedness in formulating transformer-structured inference modules, which reduces negative transfer among heterogeneous tasks in each meta-test episode. Moreover, the variants with probabilistic inference modules consistently beat deterministic ones in performance, demonstrating the advantages of considering uncertainty during modeling and inference. Table 5: Performance comparisons of different implementations of generating each local latent parameter wm τ,o from the condition zm τ and C1:M τ . Methods A C P R Avg. Concat 65.69 0.59 58.64 0.61 77.54 0.68 77.10 0.64 69.74 0.51 Add 69.92 0.69 63.73 0.71 78.81 0.77 79.03 0.78 72.87 0.61 Ours 73.31 0.63 64.92 0.68 83.38 0.66 83.54 0.64 76.29 0.51 Effects of Different Ways to Generate Local Latent Parameters. We investigate the effects of different ways to generate each wm τ,o from the shared condition zm τ and C1:M τ . Given a Monte Carlo sample of global latent variables as zm τ (i), in Table 5, we compare with two alternatives: 1) Concat directly concatenates each context feature and zm τ (i), and takes the concatenation as inputs of the transformer-structured inference network ϕ. 2) Add sums up each context feature and zm τ (i) and takes the result as the input. 3) Ours incorporates zm τ (i) into the transformer-structured inference module by merging it with the refined learnable tokens in Eq. (10). As shown in Table 5, Ours consistently performs the best. This implies that incorporating the conditional variables into the inference module is more effective than the direct combinations of zm τ (i) and instance features. Effects of More "Shots" or "Classes". To investigate the effects of more "shots" or "classes" in the episodic multi-task classification setup, we conduct experiments by increasing K or O in the defined M-task O-way K-shot setup. Table 6: Performance comparisons on Office-Home under the 4-task 5-way K-shot setup. Methods 1-shot 5-shot 10-shot 20-shot TNP-D 65.49 0.53 78.94 0.43 80.81 0.32 81.12 0.68 HNPs 76.29 0.51 80.80 0.42 81.28 0.38 81.56 0.36 As shown in Table 6, the proposed HNPs have more advantages over the baseline method with the context data points below ten shots. With shots larger than ten, both methods will reach a performance bottleneck. Moreover, Table 7 shows that our method consistently outperforms the baseline method as the number of classes increases from 20 to 40 in step 5. However, the performance gap between them narrows slightly with more classes. The main reason could be that the setting with more classes suffers from less data insufficiency. Table 7: Performance comparisons on Domain Net under the 6-task O-way 1-shot setup. Methods 5-way 20-way 25-way 30-way 35-way 40-way TNP-D 49.10 0.42 28.83 0.17 25.93 0.14 24.08 0.12 22.62 0.11 21.64 0.53 HNPs 62.36 0.53 39.32 0.23 35.72 0.19 32.27 0.17 31.27 0.14 29.31 0.13 Figure 5: Average accuracy and runtime of HNPs with different numbers of Monte Carlo samples. Nz and Nw are sampling numbers of zm τ and wm τ,1:O, respectively. Sensitivity to the Number of Monte Carlo Samples. For the hierarchical latent variables in the HNPs, we investigate the model s sensitivity to the number of Monte Carlo samples. Specifically, the sampling number of the global latent representation zm τ and local latent parameters wm τ,1:O varies from 1 to 30. We examine on Office-Home under the 4-task 5-way 1-shot setting. In Figure 5, the runtime per iteration grows rapidly as the number of samples increases. However, there is no clear correlation between the performance and the number of Monte Carlo samples. There are two sweet spots in terms of average accuracy, one of which has favorable computation time. Hence, we set Nz and Nw to 5 and 10, respectively. Table 8: Inference time of different NP-based methods. Methods CNPs NPs TNP-D HNPs Inference time(s) 0.04 0.05 0.08 0.15 We also investigate the inference time of NP-based models per iteration on Office-Home under the 4task5way1shot setup. As shown in Table 8, our model needs more inference time than other NP-based methods for performance gains. The cost mainly comes from inferring the designed hierarchical latent variables; however, we consider this a worthwhile trade-off for the extra performance. 5 Conclusion Technical Discussion. This work develops heterogeneous neural processes by introducing hierarchical latent variables and transformer-structured inference modules for episodic multi-task learning. With the help of heterogeneous context information and meta-knowledge, the proposed model can exploit task-relatedness, reason about predictive function distributions, and efficiently distill past knowledge to unseen heterogeneous tasks with limited data. Limitation & Extension. Although the hierarchical probabilistic framework could mitigate the expressiveness bottleneck, the model needs more inference time than other NP-based methods for performance gains. Besides, the proposed method requires the target space to be the same across all tasks in a single episode. This requirement could limit the method s applicability in realistic scenarios where target spaces may differ across tasks. Our work could be extended to the new case without the shared target spaces, where the model should construct higher-order task-relatedness to improve knowledge sharing among tasks. Our code 3 is provided to facilitate such extensions. Acknowledgment This work is financially supported by the Inception Institute of Artificial Intelligence, the University of Amsterdam and the allowance Top consortia for Knowledge and Innovation (TKIs) from the Netherlands Ministry of Economic Affairs and Climate Policy. 3 https://github.com/autumn9999/HNPs.git [1] Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Deep Learning. MIT Press, 2016. http: //www.deeplearningbook.org. [2] Shichao Xu, Lixu Wang, Yixuan Wang, and Qi Zhu. Weak adaptation learning: Addressing cross-domain data insufficiency with weak annotator. In IEEE Conference on Computer Vision and Pattern Recognition, 2021. [3] Justin M Johnson and Taghi M Khoshgoftaar. Survey on deep learning with class imbalance. Journal of Big Data, 6(1):1 54, 2019. [4] Oriol Vinyals, Charles Blundell, Timothy Lillicrap, Daan Wierstra, et al. Matching networks for one shot learning. In Advances in Neural Information Processing Systems, 2016. [5] Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In International Conference on Machine Learning, 2017. [6] Jake Snell, Kevin Swersky, and Richard Zemel. Prototypical networks for few-shot learning. In Advances in Neural Information Processing Systems, 2017. [7] James Requeima, Jonathan Gordon, John Bronskill, Sebastian Nowozin, and Richard E Turner. Fast and flexible multi-task classification using conditional neural adaptive processes. In Advances in Neural Information Processing Systems, 2019. [8] Rich Caruana. Multitask learning. Machine learning, 28(1):41 75, 1997. [9] Yu Zhang and Qiang Yang. A survey on multi-task learning. IEEE Transactions on Knowledge and Data Engineering, 2021. [10] Mingsheng Long, Zhangjie Cao, Jianmin Wang, and S Yu Philip. Learning multiple tasks with multilinear relationship networks. In Advances in neural information processing systems, 2017. [11] Jiayi Shen, Xiantong Zhen, Marcel Worring, and Ling Shao. Variational multi-task learning with gumbelsoftmax priors. In Advances in Neural Information Processing Systems, 2021. [12] Marta Garnelo, Jonathan Schwarz, Dan Rosenbaum, Fabio Viola, Danilo J Rezende, SM Eslami, and Yee Whye Teh. Neural processes. ar Xiv preprint ar Xiv:1807.01622, 2018. [13] Marta Garnelo, Dan Rosenbaum, Christopher Maddison, Tiago Ramalho, David Saxton, Murray Shanahan, Yee Whye Teh, Danilo Rezende, and SM Ali Eslami. Conditional neural processes. In International Conference on Machine Learning, 2018. [14] Wessel Bruinsma, Stratis Markou, James Requeima, Andrew Y. K. Foong, Tom Andersson, Anna Vaughan, Anthony Buonomo, Scott Hosking, and Richard E Turner. Autoregressive conditional neural processes. In International Conference on Learning Representations, 2023. [15] Hyunjik Kim, Andriy Mnih, Jonathan Schwarz, Marta Garnelo, Ali Eslami, Dan Rosenbaum, Oriol Vinyals, and Yee Whye Teh. Attentive neural processes. In International Conference on Learning Representations, 2019. [16] Qi Wang and Herke van Hoof. Learning expressive meta-representations with mixture of expert neural processes. In Advances in Neural Information Processing Systems, 2022. [17] Sachin Ravi and Hugo Larochelle. Optimization as a model for few-shot learning. In International Conference on Learning Representations, 2017. [18] Qi Wang and Herke Van Hoof. Doubly stochastic variational inference for neural processes with hierarchical latent variables. In International Conference on Machine Learning, 2020. [19] Donggyun Kim, Seongwoong Cho, Wonkwang Lee, and Seunghoon Hong. Multi-task processes. ar Xiv preprint ar Xiv:2110.14953, 2021. [20] Zongyu Guo, Cuiling Lan, Zhizheng Zhang, Yan Lu, and Zhibo Chen. Versatile neural processes for learning implicit neural representations. In International Conference on Learning Representations, 2023. [21] Achim Klenke. Probability theory: a comprehensive course. Springer Science & Business Media, 2013. [22] Wonjae Kim, Bokyung Son, and Ildoo Kim. Vilt: Vision-and-language transformer without convolution or region supervision. In International Conference on Machine Learning, 2021. [23] Casper Kaae Sønderby, Tapani Raiko, Lars Maaløe, Søren Kaae Sønderby, and Ole Winther. Ladder variational autoencoders. In Advances in Neural Information Processing Systems, 2016. [24] Alex Kendall, Yarin Gal, and Roberto Cipolla. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In IEEE Conference on Computer Vision and Pattern Recognition, 2018. [25] Shikun Liu, Edward Johns, and Andrew J Davison. End-to-end multi-task learning with attention. In IEEE Conference on Computer Vision and Pattern Recognition, 2019. [26] Ishan Misra, Abhinav Shrivastava, Abhinav Gupta, and Martial Hebert. Cross-stitch networks for multi-task learning. In IEEE Conference on Computer Vision and Pattern Recognition, 2016. [27] Liyang Liu, Yi Li, Zhanghui Kuang, Jing-Hao Xue, Yimin Chen, Wenming Yang, Qingmin Liao, and Wayne Zhang. Towards impartial multi-task learning. In International Conference on Learning Representations, 2020. [28] Ozan Sener and Vladlen Koltun. Multi-task learning as multi-objective optimization. In Advances in Neural Information Processing Systems, 2018. [29] Amir R Zamir, Alexander Sax, William Shen, Leonidas J Guibas, Jitendra Malik, and Silvio Savarese. Taskonomy: Disentangling task transfer learning. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018. [30] Deblina Bhattacharjee, Tong Zhang, Sabine Süsstrunk, and Mathieu Salzmann. Mult: an end-to-end multitask learning transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022. [31] Yi Zhang, Yu Zhang, and Wei Wang. Multi-task learning via generalized tensor trace norm. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, 2021. [32] Jiayi Shen, Zehao Xiao, Xiantong Zhen, Cees GM Snoek, and Marcel Worring. Association graph learning for multi-task classification with category shifts. ar Xiv preprint ar Xiv:2210.04637, 2022. [33] Yunlong Liang, Fandong Meng, Jinan Xu, Yufeng Chen, and Jie Zhou. Scheduled multi-task learning for neural chat translation. ar Xiv preprint ar Xiv:2205.03766, 2022. [34] Yi Zhang, Yu Zhang, and Wei Wang. Learning linear and nonlinear low-rank structure in multi-task learning. IEEE Transactions on Knowledge and Data Engineering, 2022. [35] Bart Bakker and Tom Heskes. Task clustering and gating for bayesian multitask learning. Journal of Machine Learning Research, 2003. [36] Kai Yu, Volker Tresp, and Anton Schwaighofer. Learning gaussian processes from multiple tasks. In International Conference on Machine Learning, 2005. [37] Michalis K Titsias and Miguel Lázaro-Gredilla. Spike and slab variational inference for multi-task and multiple kernel learning. In Advances in neural information processing systems, 2011. [38] Neil D Lawrence and John C Platt. Learning to learn with the informative vector machine. In International Conference on Machine Learning, 2004. [39] Fariba Yousefi, Michael Thomas Smith, and Mauricio A Álvarez. Multi-task learning for aggregated data using gaussian processes. ar Xiv preprint ar Xiv:1906.09412, 2019. [40] Diane Oyen and Terran Lane. Leveraging domain knowledge in multitask bayesian network structure learning. In Proceedings of the AAAI Conference on Artificial Intelligence, 2012. [41] Weizhu Qian, Bowei Chen, Yichao Zhang, Guanghui Wen, and Franck Gechter. Multi-task variational information bottleneck. ar Xiv preprint ar Xiv:2007.00339, 2020. [42] Gjorgji Strezoski, Nanne van Noord, and Marcel Worring. Learning task relatedness in multi-task learning for images in context. In International Conference on Multimedia Retrieval, 2019. [43] Ximeng Sun, Rameswar Panda, Rogerio Feris, and Kate Saenko. Adashare: Learning what to share for efficient deep multi-task learning. ar Xiv preprint ar Xiv:1911.12423, 2019. [44] Gjorgji Strezoski, Nanne van Noord, and Marcel Worring. Many task learning with task routing. In IEEE International Conference on Computer Vision, 2019. [45] Tianhe Yu, Saurabh Kumar, Abhishek Gupta, Sergey Levine, Karol Hausman, and Chelsea Finn. Gradient surgery for multi-task learning. ar Xiv preprint ar Xiv:2001.06782, 2020. [46] Bo Liu, Xingchao Liu, Xiaojie Jin, Peter Stone, and Qiang Liu. Conflict-averse gradient descent for multi-task learning. In Advances in Neural Information Processing Systems, 2021. [47] Chris Fifty, Ehsan Amid, Zhe Zhao, Tianhe Yu, Rohan Anil, and Chelsea Finn. Efficiently identifying task groupings for multi-task learning. In Advances in Neural Information Processing Systems, 2021. [48] Pengsheng Guo, Chen-Yu Lee, and Daniel Ulbricht. Learning to branch for multi-task learning. In International Conference on Machine Learning, 2020. [49] Sebastian Thrun and Lorien Pratt. Learning to learn: Introduction and overview. In Learning to learn, pages 3 17. Springer, 1998. [50] Timothy M Hospedales, Antreas Antoniou, Paul Micaelli, and Amos J Storkey. Meta-learning in neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021. [51] Kelsey R Allen, Evan Shelhamer, Hanul Shin, and Joshua B Tenenbaum. Infinite mixture prototypes for few-shot learning. In International Conference on Machine Learning, 2019. [52] Boris Oreshkin, Pau Rodríguez López, and Alexandre Lacoste. Tadam: Task dependent adaptive metric for improved few-shot learning. In Advances in neural information processing systems, 2018. [53] Sung Whan Yoon, Jun Seo, and Jaekyun Moon. Tapnet: Neural network augmented with task-adaptive projection for few-shot learning. In International Conference on Machine Learning, 2019. [54] Victor Garcia and Joan Bruna. Few-shot learning with graph neural networks. In International Conference on Learning Representations, 2018. [55] Tianshi Cao, Marc Law, and Sanja Fidler. A theoretical analysis of the number of shots in few-shot learning. ar Xiv preprint ar Xiv:1909.11722, 2019. [56] Eleni Triantafillou, Tyler Zhu, Vincent Dumoulin, Pascal Lamblin, Utku Evci, Kelvin Xu, Ross Goroshin, Carles Gelada, Kevin Swersky, Pierre-Antoine Manzagol, and Hugo Larochelle. Meta-dataset: A dataset of datasets for learning to learn from few examples. ar Xiv preprint ar Xiv:1903.03096, 2019. [57] Flood Sung, Yongxin Yang, Li Zhang, Tao Xiang, Philip HS Torr, and Timothy M Hospedales. Learning to compare: Relation network for few-shot learning. In IEEE Conference on Computer Vision and Pattern Recognition, 2018. [58] Alex Nichol, Joshua Achiam, and John Schulman. On first-order meta-learning algorithms. ar Xiv preprint ar Xiv:1803.02999, 2018. [59] Andrei A Rusu, Dushyant Rao, Jakub Sygnowski, Oriol Vinyals, Razvan Pascanu, Simon Osindero, and Raia Hadsell. Meta-learning with latent embedding optimization. In International Conference on Learning Representations, 2019. [60] Flood Sung, Li Zhang, Tao Xiang, Timothy Hospedales, and Yongxin Yang. Learning to learn: Meta-critic networks for sample efficient learning. ar Xiv preprint ar Xiv:1706.09529, 2017. [61] Zhenguo Li, Fengwei Zhou, Fei Chen, and Hang Li. Meta-sgd: Learning to learn quickly for few-shot learning. ar Xiv preprint ar Xiv:1707.09835, 2017. [62] Jonathan Gordon, John Bronskill, Matthias Bauer, Sebastian Nowozin, and Richard E Turner. Metalearning probabilistic inference for prediction. In International Conference on Learning Representations, 2019. [63] Harrison Edwards and Amos Storkey. Towards a neural statistician. ar Xiv preprint ar Xiv:1606.02185, 2016. [64] Chelsea Finn, Kelvin Xu, and Sergey Levine. Probabilistic model-agnostic meta-learning. In Advances in Neural Information Processing Systems, 2018. [65] Steindór Sæmundsson, Katja Hofmann, and Marc Peter Deisenroth. Meta reinforcement learning with latent variable gaussian processes. ar Xiv preprint ar Xiv:1803.07551, 2018. [66] Sungyong Baik, Junghoon Oh, Seokil Hong, and Kyoung Mu Lee. Learning to forget for meta-learning via task-and-layer-wise attenuation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021. [67] Myungsub Choi, Janghoon Choi, Sungyong Baik, Tae Hyun Kim, and Kyoung Mu Lee. Test-time adaptation for video frame interpolation via meta-learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021. [68] Aniruddh Raghu, Maithra Raghu, Samy Bengio, and Oriol Vinyals. Rapid learning or feature reuse? towards understanding the effectiveness of maml. ar Xiv preprint ar Xiv:1909.09157, 2019. [69] Ying Wei, Peilin Zhao, and Junzhou Huang. Meta-learning hyperparameter performance prediction with neural processes. In International Conference on Machine Learning, 2021. [70] Stratis Markou, James Requeima, Wessel P Bruinsma, Anna Vaughan, and Richard E Turner. Practical conditional neural processes via tractable dependent predictions. ar Xiv preprint ar Xiv:2203.08775, 2022. [71] Zesheng Ye and Lina Yao. Contrastive conditional neural processes. In IEEE Conference on Computer Vision and Pattern Recognition, 2022. [72] Mingyu Kim, Kyeongryeol Go, and Se-Young Yun. Neural processes with stochastic attention: Paying more attention to the context dataset. ar Xiv preprint ar Xiv:2204.05449, 2022. [73] Qi Wang, Marco Federici, and Herke van Hoof. Bridge the inference gaps of neural processes via expectation maximization. In International Conference on Learning Representations, 2023. [74] Xiaozhuang Song, Shun Zheng, Wei Cao, James Yu, and Jiang Bian. Efficient and effective multi-task grouping via meta learning on task combinations. Advances in Neural Information Processing Systems, 2022. [75] Richa Upadhyay, Prakash Chandra Chhipa, Ronald Phlypo, Rajkumar Saini, and Marcus Liwicki. Multitask meta learning: learn how to adapt to unseen tasks. In 2023 International Joint Conference on Neural Networks (IJCNN). IEEE, 2023. [76] Haoxiang Wang, Han Zhao, and Bo Li. Bridging multi-task learning and meta-learning: Towards efficient training and effective adaptation. In International Conference on Machine Learning, 2021. [77] Hung-Yu Tseng, Hsin-Ying Lee, Jia-Bin Huang, and Ming-Hsuan Yang. Cross-domain few-shot classification via learned feature-wise transformation. In International Conference on Learning Representations, 2020. [78] Wei-Yu Chen, Yen-Cheng Liu, Zsolt Kira, Yu-Chiang Frank Wang, and Jia-Bin Huang. A closer look at few-shot classification. ar Xiv preprint ar Xiv:1904.04232, 2019. [79] Yunhui Guo, Noel C Codella, Leonid Karlinsky, James V Codella, John R Smith, Kate Saenko, Tajana Rosing, and Rogerio Feris. A broader study of cross-domain few-shot learning. In European Conference on Computer Vision, 2020. [80] Yingjun Du, Xiantong Zhen, Ling Shao, and Cees G M Snoek. Meta Norm: Learning to normalize few-shot batches across domains. In International Conference on Learning Representations, 2021. [81] Debasmit Das, Sungrack Yun, and Fatih Porikli. Confess: A framework for single source cross-domain few-shot learning. In International Conference on Learning Representations, 2022. [82] Wei-Hong Li, Xialei Liu, and Hakan Bilen. Cross-domain few-shot learning with task-specific adapters. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7161 7170, 2022. [83] Risto Vuorio, Shao-Hua Sun, Hexiang Hu, and Joseph J Lim. Multimodal model-agnostic meta-learning via task-aware modulation. In Advances in neural information processing systems, 2019. [84] Risto Vuorio, Shao-Hua Sun, Hexiang Hu, and Joseph J Lim. Toward multimodal model-agnostic metalearning. In ar Xiv preprint ar Xiv:1812.07172, 2018. [85] Huaxiu Yao, Ying Wei, Junzhou Huang, and Zhenhui Li. Hierarchically structured meta-learning. In International Conference on Machine Learning, 2019. [86] Milad Abdollahzadeh, Touba Malekzadeh, and Ngai-Man Man Cheung. Revisit multimodal meta-learning through the lens of multi-task learning. In Advances in Neural Information Processing Systems, 2021. [87] Jiayi Chen and Aidong Zhang. Hetmaml: Task-heterogeneous model-agnostic meta-learning for few-shot learning across modalities. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, 2021. [88] Chen Xing, Negar Rostamzadeh, Boris Oreshkin, and Pedro O O Pinheiro. Adaptive cross-modal few-shot learning. In Advances in Neural Information Processing Systems, 2019. [89] Frederik Pahde, Patrick Jähnichen, Tassilo Klein, and Moin Nabi. Cross-modal hallucination for few-shot fine-grained recognition. ar Xiv preprint ar Xiv:1806.05147, 2018. [90] Frederik Pahde, Mihai Puscas, Tassilo Klein, and Moin Nabi. Multimodal prototypical networks for few-shot learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021. [91] Hemanth Venkateswara, Jose Eusebio, Shayok Chakraborty, and Sethuraman Panchanathan. Deep hashing network for unsupervised domain adaptation. In IEEE Conference on Computer Vision and Pattern Recognition, 2017. [92] Xingchao Peng, Qinxun Bai, Xide Xia, Zijun Huang, Kate Saenko, and Bo Wang. Moment matching for multi-source domain adaptation. In IEEE International Conference on Computer Vision, 2019. [93] Ishaan Gulrajani and David Lopez-Paz. In search of lost domain generalization. ar Xiv preprint ar Xiv:2007.01434, 2020. [94] Ze Wang, Zichen Miao, Xiantong Zhen, and Qiang Qiu. Learning to learn dense gaussian processes for few-shot learning. In Advances in Neural Information Processing Systems, 2021. [95] Tung Nguyen and Aditya Grover. Transformer neural processes: Uncertainty-aware meta learning via sequence modeling. ar Xiv preprint ar Xiv:2207.04179, 2022.