# texttti2moe_interpretable_multimodal_interactionaware_mixtureofexperts__565aa521.pdf I2Mo E: Interpretable Multimodal Interaction-aware Mixture-of-Experts Jiayi Xin 1 Sukwon Yun 2 Jie Peng 3 Inyoung Choi 1 Jenna L. Ballard 1 Tianlong Chen 2 Qi Long 1 Modality fusion is a cornerstone of multimodal learning, enabling information integration from diverse data sources. However, vanilla fusion methods are limited by (1) inability to account for heterogeneous interactions between modalities and (2) lack of interpretability in uncovering the multimodal interactions inherent in the data. To this end, we propose I2Mo E (Interpretable Multimodal Interaction-aware Mixture of Experts), an end-to-end Mo E framework designed to enhance modality fusion by explicitly modeling diverse multimodal interactions, as well as providing interpretation on a local and global level. First, I2Mo E utilizes different interaction experts with weakly supervised interaction losses to learn multimodal interactions in a data-driven way. Second, I2Mo E deploys a reweighting model that assigns importance scores for the output of each interaction expert, which offers sample-level and dataset-level interpretation. Extensive evaluation of medical and general multimodal datasets shows that I2Mo E is flexible enough to be combined with different fusion techniques, consistently improves task performance, and provides interpretation across various real-world scenarios. Code is available at https://github. com/Raina-Xin/I2Mo E. 1. Introduction A core challenge in multimodal learning is modality fusion the integration of information from multiple modalities to improve predictive performance (Baltruˇsaitis et al., 2019; Barnum et al., 2020; Lv et al., 2021). By leverag- 1University of Pennsylvania, PA, USA 2University of North Carolina at Chapel Hill, NC, USA 3University of Science and Technology of China, Anhui, China. Correspondence to: Qi Long , Tianlong Chen , Jiayi Xin . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). Genres "Drama", "Fantasy", "Horror", "Romance" "A poor student rescues a beautiful countess and soon becomes obsessed with her. A sorcerer makes a deal with the young man to give him fabulous wealth and anything he wants, if he will sign his name to a contract. The student hurriedly signs the contract, but doesn't know what he's in for." Image Modality Language Modality Uniqueness of Language Figure 1. An illustrative example of modality interaction. The poster and plot are taken from the IMDB dataset. ing diverse data sources such as text, images, audio, and sensor data, modality fusion enables the capture of intricate relationships across modalities, which is especially crucial in fields like healthcare, where accurate decisionmaking relies on multimodal insights (Liang et al., 2022b; Kline et al., 2022; Teoh et al., 2024). Although recent advancements in neural architectures, such as transformers (Vaswani et al., 2017; Tsai et al., 2019) and sparse mixtureof-experts (Shazeer et al., 2017; Fedus et al., 2022; Jin et al., 2024), have significantly improved the modeling of modality interactions, an important yet underexplored area is the systematic understanding of how modalities influence one another whether they provide complementary, supplementary, or even conflicting information (Baltruˇsaitis et al., 2019; Liang et al., 2022b; 2023). Understanding modality interaction is essential for advancing multimodal machine learning (Baltruˇsaitis et al., 2019; Liang et al., 2022b). An information-theoretic framework called Partial Information Decomposition (PID) (Wollstadt et al., 2023; Liang et al., 2023) offers a theoretical foundation for understanding modality interactions. PID decomposes information into four distinct types: uniqueness for the first modality (information specific to modality 1), uniqueness for the second modality (information specific to modality 2), synergy (emergent information arising from the combination of two modalities), and redundancy (shared information across two modalities). Figure 1 illustrates the importance of carefully modeling different types of multimodal interactions. For instance, the unique information provided by the image modality Interpretable Multimodal Interaction-aware Mixture-of-Experts (mimg) contributes to predicting the Horror genre through distinct visual cues absent in the language modality (mlang) while the unique information from the language modality offers critical textual context for identifying the Romance genre. Redundant information refers to shared information present in both modalities, such as recognizing the Fantasy genre through the blurry figure in the poster and mentioning a sorcerer in the plot. Accurately classifying the movie as Drama, however, requires modeling synergistic information between the two modalities: visual elements such as clothing and facial expressions in mimg, complement the narrative details from mlang. From this example, systematic modeling of multimodal interactions is needed to make accurate predictions. While the PID framework provides valuable theoretical insights into the proportions of different modality interactions within a dataset, its practical application is limited, lacking integration into end-to-end and interpretable deep learning frameworks. Most existing multimodal fusion methods do not explicitly model multimodal interactions (Liu et al., 2018; Tsai et al., 2019; Xue & Marculescu, 2023). Notable efforts to address this gap, such as (W ortwein et al., 2022; Yu et al., 2024; Dufumier et al., 2024), exhibit key limitations: they either focus exclusively on pairwise modality interactions (W ortwein et al., 2022), require separate estimates for each interaction type (Yu et al., 2024), or lack sufficient interpretability (Dufumier et al., 2024). The opportunity to directly leverage PID for improving both task performance and model interpretability within multimodal fusion frameworks remains largely unexplored. In contrast to earlier works, we propose I2Mo E, an endto-end mixture-of-experts (Mo E) framework designed to enhance task performance while improving interpretability. Our approach incorporates separate parameters and weaklysupervised interaction losses, enabling the mixture of interaction experts to effectively model diverse interactions between modalities. To further enhance interpretability, we introduce a re-weighting model that assigns importance scores to each interaction expert, providing insights into decision-making at both local (sample-level) and global (dataset-level) scales. I2Mo E is backbone-agnostic and can be seamlessly integrated with any modality fusion approach. We evaluate the effectiveness of I2Mo E on two medical datasets and three real-world multimodal datasets, demonstrating its ability to consistently improve performance while offering interpretable insights into the model s decision-making process for individual samples. Our contributions are summarized as follows: We introduce I2Mo E, a novel mixture-of-experts framework designed to explicitly model diverse modality interactions through specialized parameters and weakly-supervised interaction losses, enabling a more nuanced understanding of multimodal data. We enhance interpretability by providing both samplelevel and dataset-level insights into model decisions, offering a deeper understanding of how interaction experts contribute to predictions. I2Mo E is highly flexible and can be seamlessly integrated with existing modality fusion methods, demonstrating its versatility in improving vanilla multimodal fusion backbones. Extensive experiments on five diverse real-world multimodal datasets validate the efficacy of I2Mo E, showcasing significant performance improvements (up to 5.5% in accuracy) and interpretability benefits over vanilla modality fusion methods. 2. Related Work Modality Interaction is theoretically grounded in the Partial Information Decomposition (PID) framework (Liang et al., 2023), which analyzes heterogeneous interactions but lacks an end-to-end learning framework. Prior works attempt to model interactions but are either restricted to specific interaction types (Zhang et al., 2023; Kim et al.), fail to quantify interactions in the data (W ortwein et al., 2024; Liang et al., 2024; Long et al., 2024; Dufumier et al., 2024), or are limited to only two modalities (W ortwein et al., 2022; Fan et al., 2024). Our approach bridges this gap by directly modeling and quantifying modality interactions within a unified Mo E-based fusion architecture, enabling effective and interpretable multimodal learning. Multimodal Fusion integrates data from multiple sources to enhance prediction tasks. Existing methods often rely on concatenating input modalities using off-the-shelf architectures (Liu et al., 2018; Tsai et al., 2019; Xue & Marculescu, 2023; Shazeer et al., 2017; Fedus et al., 2022). Mixture-of-Experts (Mo E) offers a natural architecture for modeling interactions via expert specialization (Jacobs et al., 1991; Chen et al., 1999; Yuksel et al., 2012). Several recent works (Mustafa et al., 2022; Lin et al., 2024; Yu et al., 2024) explore Mo E for multimodal learning. Among them, only MMo E (Yu et al., 2024) explicitly models different types of modality interactions by using a mixture of interaction experts on sentiment analysis. However, MMo E treats modality interaction modeling as a preprocessing step rather than integrating it into an end-to-end learning framework, limiting flexibility and interpretability. Multimodal Interpretation has gained traction as researchers seek to explain decision-making in multimodal AI systems. Prior studies either focus on isolating the effect of individual modalities while overlooking inter-modal interactions (Ismail et al., 2022; Ghosh et al., 2023; Swamy et al., 2024b), provide human-interpretable rationales but Interpretable Multimodal Interaction-aware Mixture-of-Experts fail to quantify interaction contributions (Park et al., 2018; Zadeh et al., 2018; Dominici et al., 2023), or lack explicit categorization of interaction types (Tsai et al., 2020; Chefer et al., 2021; Lyu et al., 2022; Liang et al., 2022a; Wenderoth et al., 2024). As no prior work has explored interpretation from a modality interaction perspective, our contribution is to systematically quantify multimodal interactions while maintaining interpretability. 3. Interpretable Multimodal Interaction-aware Mixture-of-Experts 3.1. Preliminary and Notation Problem Setup. Let M = {m1, m2, . . . , mn} denote a set of n input data modalities, and let y represent the target variable for a given task. For classification tasks, y is expressed as a one-hot encoded vector corresponding to the class label. For regression tasks, y is a real-valued scalar. The objective is twofold: (1) to improve the performance of predicting the ground truth target y by effectively modeling the interactions between modalities in M, and (2) to provide meaningful interpretations of these multimodal interactions. Vanilla multimodal fusion (Figure 2(a)) utilizes modalityspecific encoders E = {E1, E2, . . . , En} to process M and obtain latent embeddings L = {e1, e2, . . . , en}, where each embedding is computed as ei = Ei(mi) for i {1, . . . , n}. We define the fusion method as F, which operates on the set of latent embeddings L and produces a fused embedding x, expressed as: F(L) = x. A prediction head H maps the fused embedding to the final prediction, expressed as: H(x) = ˆy. However, this naive modality fusion approach does not explicitly account for the heterogeneous interactions present between M. 3.2. Algorithm Overview of I2Mo E Framework I2Mo E is a mixture of interaction experts, where each expert specializes in modeling a specific type of multimodal interaction. The predictions from individual interaction experts are weighted by a re-weighting model to produce the final prediction. During the training phase, we first perform a forward pass using the intact input of all modalities to estimate the multimodal prediction. Next, additional forward passes are conducted, where one modality is replaced by a random vector in each pass. These perturbed inputs serve as weak supervision signals to help train the interaction experts to specialize in different types of modality interactions. We designed a dual-objective loss, encouraging the interaction experts to specialize effectively without degrading task performance. The task loss is calculated using the re-weighted output from the interaction experts with the complete modality input, while the interaction loss is computed from the outputs generated with the perturbed modality inputs. During inference, a single forward pass is performed using the complete modality input. The final output is a weighted sum of the interaction expert prediction with the weights produced by the re-weighting model (Equation 1). We provide a detailed explanation of I2Mo E with two input modalities in Section 3.3, describe its extension to a higher number of modalities in Section 3.4, and explain how to obtain multimodal interaction interpretation in Section 3.5. 3.3. I2Mo E with Two Input Modalities 3.3.1. I2MOE ARCHITECTURE Figure 2(b) illustrates the I2Mo E architecture for modeling different types of modality interactions in two input modalities. We employ a Mo E comprising four fusion models, referred to as interaction experts: Funi1, Funi2, Fsyn, and Fred. Each interaction expert specializes in capturing a specific type of interaction: Funi1 models the unique information contained in modality m1; Funi2 models the unique information contained in modality m2; Fsyn captures the synergistic information between m1 and m2; and Fred models the redundant information between m1 and m2. Each interaction expert processes the latent embeddings of the two modalities, e1 and e2, and produces fused embeddings, represented as xi = Fi(e1, e2), where i {uni1, uni2, syn, red}. These fused embeddings are then passed through a prediction head within each interaction expert, generating predictions for the corresponding interaction type as ˆyi = Hi(xi), where i {uni1, uni2, syn, red}. To combine the predictions from the four interaction experts, we introduce a re-weighting model W, which assigns importance scores to the predictions of each expert. The model W takes the latent embeddings e1 and e2 as inputs and outputs a set of soft weights: W(e1, e2) = [wuni1, wuni2, wsyn, wred]. The final prediction is obtained by combining the predictions from all experts using these weights, expressed as: i wi ˆyi, i {uni1, uni2, syn, red}. (1) 3.3.2. I2MOE LEARNING OBJECTIVE The loss function consists of two components. The first component is the task loss, which encourages the predictions of I2Mo E, ˆy, to closely match the ground truth target y. The second component, termed the interaction loss, ensures that the initially identical fusion models within I2Mo E specialize into interaction experts by capturing diverse interactions in the dataset. Following Yu et al. (2024), we characterize interaction types by comparing unimodal and multimodal predictions: predictions made using only the first modality (y1), predictions Interpretable Multimodal Interaction-aware Mixture-of-Experts (b) Ours: I2Mo E Random Vector Random Vector Model Diverse Interactions Provide Interpretation Rdncy Uni-1 Final Prediction Fusion Model Prediction Head (a) Vanilla Fusion No Interaction Modeling No Interpretation Final Prediction Genres Genres Experts Uni-1 Expert Uni-2 Expert Syngy Expert Rdncy Expert Prediction Interaction ... ... ... ... Genres Genres Fusion Model Prediction Head Fusion Model Prediction Head Fusion Model Prediction Head Fusion Model Prediction Head Encoder Encoder Encoder Encoder Figure 2. Comparison between vanilla modality fusion and I2Mo E in the case of movie genre classification with two input modalities. Left: Existing modality fusion approaches typically use the same parameters to model all types of interactions between the two modalities. Right: In contrast, we design a mixture-of-experts framework that employs four different interaction experts and a re-weighting model to explicitly capture heterogeneous interactions between the two input modalities. made using only the second modality (y2), and predictions made using both modalities (y12). For interactions emphasizing the uniqueness of the first modality, the relationships are defined as y12 = y1 and y12 = y2. Similarly, for interactions emphasizing the uniqueness of the second modality, we have y12 = y2 and y12 = y1. For synergistic interactions, the condition is y12 = y1 and y12 = y2. For redundant interactions, the relationship is y12 = y1 = y2. To approximate the interaction loss, we simulate the unimodal scenario by replacing one of the modalities with a random vector. For each interaction expert, a unimodal prediction using only the first modality can be obtained by replacing the latent embedding of the second modality with a random vector r, represented as: ˆy 2,i = Hi Fi(E1(x1), r) , (2) where i {uni1, uni2, syn, red}. Similarly, a unimodal prediction using only the second modality can be generated by replacing the latent embedding of the first modality with r, expressed as: ˆy 1,i = Hi Fi(r, E2(x2)) , (3) where i {uni1, uni2, syn, red}. We designed a general framework to approximate different types of modality interactions. In all cases, the output using the complete multimodal input, ˆy12, serves as the anchor. For the Funi1, the output with modality 2 masked, ˆy 2, is treated as a positive example, while the output with modality 1 masked, ˆy 1, is treated as a negative example. The objective is to encourage ˆy12 to be maximally similar to ˆy 2 and maximally different from ˆy 1, since Funi1 models the uniqueness information presented in m1. For the Funi2, ˆy 2 is treated as a negative example, while ˆy 1 is treated as a positive example. Here, the objective is to encourage ˆy12 to be maximally similar to ˆy 1 and maximally different from ˆy 2, since Funi2 models the uniqueness information presented in m2. For the Fsyn, ˆy 1 and ˆy 2 are both treated as negative examples. The objective is to ensure that ˆy12 is maximally different from both ˆy 2 and ˆy 1, capturing interactions that require the combination of both modalities. For the Fred, ˆy 1, and ˆy 2 are treated as positive examples. The goal is to encourage ˆy12, ˆy 2, and ˆy 1 to be as similar as possible, modeling information shared between the modalities. We discuss the connection between the proposed interaction loss and the PID formulation in Appendix A and present empirical evidence supporting the design choice of random vector masking,in Appendix B. 3.4. Extend I2Mo E to Higher Number of Modalities Increase Uniqueness Interaction Experts. To extend I2Mo E to support more than two input modalities, we increase the number of interaction experts to the |M| + 2. Instead of a combinatorial explosion in the number of interaction experts, as the number of input modalities grows, we define m uniqueness interaction experts, one for each input Interpretable Multimodal Interaction-aware Mixture-of-Experts Algorithm 1 Training and Inference of I2Mo E Require: Modalities X1, . . . , Xn, label T Require: Modality-specific Encoders {Enci}n i=1 Require: Experts {Fi}E i=1, reweighting module W Require: Expert loss functions {Interaction Lossi}E i=1 // Training with masked modality input 1: Encode modalities: Zi Enci(Xi) for i = 1, . . . , n 2: for i = 1 to E do 3: [ˆy(0) i , . . . , ˆy(n) i ] F multi i (Z1, . . . , Zn) 4: Li int Interaction Lossi(ˆy(0) i , ˆy(1:n) i ) 5: end for 6: [w1, . . . , w E] W(Z1, . . . , Zn) 7: ˆy PE i=1 wi ˆy(0) i 8: Ltask ℓ(ˆy, T) 9: Ltotal Ltask + λint E PE i=1 Li int 10: Update model parameters to minimize Ltotal 11: procedure INFERENCE 12: Encode modalities: Zi Enci(Xi) for i = 1, . . . , n 13: ˆy(0) i Fi(Z1, . . . , Zn) for i = 1, . . . , E 14: [w1, . . . , w E] W(Z1, . . . , Zn) 15: ˆy PE i=1 wi ˆy(0) i 16: Store {ˆyi}, w, and prediction ˆy 17: end procedure modality, along with a single synergy expert and a single redundancy expert. Each uniqueness expert, Funi,i, is responsible for capturing the unique information specific to its corresponding modality, mi M, where i {1, . . . , n}. The synergy expert, Fsyn, focuses on modeling global synergistic interactions across all modalities, while the redundancy expert, Fred, captures globally redundant information shared among the modalities. Modify Interaction loss. For uniqueness expert i, we consider the output of the complete modality as the anchor. The masked modality i serves as a negative example, while all other perturbed inputs are treated as positive examples. This is because the unique information of modality i is lost when the modality embedding is replaced by random vectors. For the synergy interaction loss, we treat all the output of the perturbed modality as negative examples, as input modality perturbations damage the synergistic information. For the redundancy interaction loss, we consider the output of the perturbed modality as a positive example because redundant information remains available even when one modality is masked. For classification tasks, we employed Triplet Margin Loss to model uniqueness interactions. For synergy and redundancy interactions, we utilized Cosine Similarity to capture the relationships between modality outputs. For regression tasks, we used the Mean Squared Error (MSE) Loss to measure differences in predictions. I2Mo E Algorithm and Complete Objective. We present the training and inference pipeline of I2Mo E in Algorithm 1. The complete learning objective is provided in Appendix C. We analyze computational overhead and scalability in Appendix D. 3.5. Local and Global Interpretation from I2Mo E Local interpretation provides insight into the extent to which different interactions contribute to the final prediction for each individual sample, while global interpretation highlights the average trends of interaction importance across the entire dataset. For I2Mo E, decisions are made locally for each specific input sample by analyzing the prediction, ˆyi, from each interaction expert Fi, and the importance coefficients, wi, assigned by the MLP-based re-weighting model W. Global interpretation for I2Mo E can be achieved by calculating the statistics of the importance weights wi assigned to each interaction expert across all samples in the test set, thereby capturing the overall trends in feature contributions. 4. Experiment Setup Data Collection and Datasets. We evaluate our method on five multimodal datasets, using all available modalities while discarding samples with missing data. Two Medical Multimodal Datasets: ADNI (Weiner et al., 2010; 2017) consists of 2,380 samples for Alzheimer s Disease classification (Dementia, Cognitively Normal, or Mild Cognitive Impairment). It includes four modalities: Image (I), Genetic (G), Clinical (C), and Biospecimen (B). MIMIC-IV (Johnson et al., 2023) is a critical care dataset with 9,003 patient records for oneyear mortality prediction (binary classification), utilizing three modalities: Lab (L), Notes (N), and Code (C). Three General Multimodal Datasets: IMDB (Arevalo et al., 2017) includes 25,959 movies for multi-label genre classification across 23 genres, leveraging Image (I) and Language (L) modalities. MOSI (Zadeh et al., 2016) comprises 2,199 annotated You Tube clips for sentiment analysis (regression with scores [-3,3] and then map to binary classification), incorporating Vision (V), Audio (A), and Text (T ) modalities. ENRICO (Leiva et al., 2020) contains 1,460 Android app screens for UI design classification into 20 categories, featuring two modalities: Screenshot (S) and Wireframe (W). Detailed dataset preprocessing is provided in Appendix E. Modality-Specific Encoders and Prediction Heads. The primary objective of our experiments is to evaluate whether the proposed mixture-of-experts framework improves modality fusion. To ensure a fair comparison, we control for variations in modality-specific encoders (E) and prediction models (H) by using the same E and H for both vanilla multimodal fusion and I2Mo E. For further details on the encoder and classification head configurations, please refer to Appendix F. Interpretable Multimodal Interaction-aware Mixture-of-Experts Baseline Fusion Methods. To validate the effectiveness of I2Mo E in enhancing multimodal learning, we compare it to various widely used fusion techniques. We begin with fundamental approaches, including early fusion (EF) (Baltruˇsaitis et al., 2019), late fusion (LF) (Baltruˇsaitis et al., 2019), low-rank multimodal fusion (LRMF) (Liu et al., 2018), and multimodal transformers (Mul T) (Tsai et al., 2019). We then implement more advanced fusion methods, including interpretable conditional computation (Interpret CC) (Swamy et al., 2024a), the Switch Transformer (Switch Gate) (Fedus et al., 2022), and sparse mixture-of-experts (Mo E++) (Jin et al., 2024). In both Switch Gate and Mo E++, the MLP layer in Mul T is replaced with a sparse Mo E layer that incorporates the respective routing function. Implementations. The dataset is partitioned into training, validation, and testing sets, with 70% allocated for training, 15% for validation, and the remaining 15% for testing. Each experiment is run three times with different random seeds and the results are averaged. To ensure a fair comparison with other baselines, we utilize the optimal hyperparameter settings provided in the original studies. If a dataset does not have reported optimal parameters, we perform a grid search over the key hyperparameters of the baseline methods. The re-weighting model (W) is implemented as a multilayer perceptron (MLP). For a detailed description of the hyperparameter settings, we refer the reader to Appendix G. 5. Performance and Interpretability of I2Mo E 5.1. I2Mo E Demonstrates Superior Task Performance In Table 1, we compared the performance of I2Mo E combining with Mul T (I2Mo E-Mul T) with other vanilla fusion methods across five datasets: ❶Compared to vanilla Mul T, I2Mo E yields a significant accuracy improvement of 5.5% for ADNI and 3% for MOSI, demonstrating its ability to enhance the performance of existing transformers. ❷Across all datasets, I2Mo E outperforms advanced baselines such as Switch Gate and Mo E++, with a notable gain of 2.5% accuracy, 1.5% AUROC on ADNI, and 1.4% improvement in Macro F1 for IMDB. These results illustrate the benefit of I2Mo E in tackling the challenges of modality interaction to achieve superior task performance. 5.2. Generalization Across Different Fusion Methods To evaluate the generalizability of I2Mo E across various fusion backbones, we integrate it with three fusion architectures, including Mo E++, Switch Gate, and Interpret-CC, and assess the combined models on all datasets (Table 2): ❶For the ADNI dataset, I2Mo E yields significant performance gains, with up to 5.23% improvement in accuracy and 2.12% in AUROC when combined with Switch Gate. ❷ On the MIMIC dataset, I2Mo E achieves notable AUROC improvements of 4.43% when combined with Interpret-CC, highlighting its ability to capture complex interaction in multimodal patient data. However, accuracy decreases (- 0.56% to -11.82%) are observed, which can be attributed to dataset imbalance. In such cases, the model becomes less overfitted to the majority class, leading to a decrease in accuracy but a corresponding increase in AUROC, reflecting improved performance in distinguishing between classes overall. ❸I2Mo E consistently enhances multimodal learning, achieving improvements in Micro F1 on IMDB (2.45%), sentiment analysis accuracy on MOSI (4.76%), and design classification accuracy on ENRICO (5.14%) when integrated with Mo E++ and Switch Gate. Results with different fusion backbones emphasize the generalizability and effectiveness of I2Mo E . 5.3. I2Mo E Offers Local Interpretation To illustrate the interpretability provided by I2Mo E on the individual sample level, we present a qualitative example from the IMDB test set where I2Mo E-Mul T makes a correct prediction (Figure 3). This example showcases how different interaction experts contribute to the final prediction through visualized logits and assigned weights, offering a clear decomposition of the decision-making process. The ground truth genres of this movie include Animation. In Figure 3(a), the logits produced by each interaction expert are shown. Notably, the uniqueness expert for the image modality and the redundancy expert generate positive logits, while the synergy expert yields a negative logit. This aligns with the visual content of the image, which features cartoon characters uniquely contributing to the prediction in Figure 3(d). Figure 3(b) depicts the weights assigned by the reweighting mechanism. Higher weights are given to the uniqueness expert for the image modality and the redundancy expert. As shown in Figure 3(c), the final weighted logits for the Animation genre become positive, enabling the correct prediction. This example demonstrates how I2Mo E leverages different interaction patterns to make accurate predictions. We provide human evaluation of local interpretation in Appendix H and additional qualitative examples in Appendix I. 5.4. I2Mo E Enables Global Interpretation We analyze the weight assigned by the reweighting model to each interaction expert across all test samples. Figure 4 illustrates the weight variation across datasets, offering insights into dataset-level interaction patterns. The reweighting model demonstrates the ability to adaptively assign distinct weights to interaction experts, reflecting its capacity to capture dataset-specific nuances. In the ADNI dataset, weights are relatively uniform, with a subtle bias toward certain experts, indicating balanced contributions from all inter- Interpretable Multimodal Interaction-aware Mixture-of-Experts Table 1. Comparison of Accuracy, AUROC, and F1 scores across different fusion methods and datasets. The upper panel lists vanilla fusion methods, while the last row presents the proposed I2Mo E framework combined with Mul T fusion method. Dataset ADNI MIMIC IMDB MOSI ENRICO Metrics Accuracy AUROC Accuracy AUROC Micro F1 Macro F1 Accuracy Accuracy EF 52.01 0.92 65.69 1.81 67.63 1.66 67.75 0.93 56.10 0.27 41.12 1.08 72.16 0.66 42.35 0.81 LF 50.79 3.11 68.60 3.77 67.11 1.06 67.58 0.88 56.22 0.03 45.27 0.64 70.51 1.14 44.20 1.64 LRMF 50.79 2.20 69.37 3.13 70.17 1.79 65.45 6.31 56.22 0.03 45.27 0.64 76.63 0.18 46.12 1.06 Interpret CC 54.53 3.43 72.18 1.70 72.34 4.48 61.93 2.53 58.00 0.23 48.68 0.11 75.85 0.07 47.60 1.56 Switch Gate 62.28 1.17 79.70 0.20 70.98 0.83 68.26 3.25 55.92 0.07 47.33 0.47 72.35 0.27 43.95 2.83 Mo E++ 58.08 2.52 75.18 1.95 72.51 2.09 68.50 2.13 58.15 0.32 50.49 0.25 70.85 0.83 47.83 1.86 Mul T 59.57 0.66 77.21 0.51 72.42 2.53 68.79 3.34 59.68 0.19 51.41 0.04 68.80 0.78 47.37 1.82 I2Mo E-Mul T 65.08 1.52 81.09 0.02 69.78 0.91 68.81 0.99 61.00 0.44 52.38 0.48 71.91 2.20 48.22 1.61 Table 2. Comparison of metrics across datasets using different fusion methods for I2Mo E. Performance improvements are indicated in blue, and decreases are indicated in red. Dataset i2Mo ESwitch Gate Interpret CC Mo E++ ADNI Accuracy 67.51 (5.23) 56.02 (1.49) 59.01 (0.93) AUROC 81.82 (2.12) 73.36 (1.18) 75.69 (0.51) MIMIC Accuracy 70.42 (-0.56) 69.85 (-2.49) 60.69 (-11.82) AUROC 69.08 (0.82) 66.36 (4.43) 69.15 (0.65) IMDB Micro F1 57.43 (1.51) 58.32 (0.32) 60.60 (2.45) Macro F1 47.77 (0.44) 49.21 (0.53) 50.73 (0.24) MOSI Accuracy 73.86 (1.51) 76.14 (0.29) 75.61 (4.76) ENRICO Accuracy 49.09 (5.14) 49.09 (1.49) 47.83 (0) When a green ogre named Shrek discovers his swamp has been 'swamped' with all sorts of fairytale creatures by the scheming Lord Farquaad, Shrek sets out with a very loud donkey by his side to 'persuade' Farquaad to give Shrek his (c) (d) Figure 3. Qualitative example of local interpretation on the IMDB dataset provided by I2Mo E-Mul T. Ground truth labels are Comedy, Adventure, Fantasy, Family, and Animation. (a) Logits output by different interaction experts. (b) Weighting assigned by the reweighting model. (c) Contribution of each interaction expert to the final weighted logit. (d) Raw image and language modalities used for prediction. action experts to the model s performance. Conversely, the MIMIC dataset displays pronounced variability in weight assignments, emphasizing I2Mo E s reliance on reweighting model to address variance among individual patients. For the IMDB dataset, the weight variation is less pronounced compared to MIMIC, aligning with its more homogeneous characteristics. The MOSI dataset shows evenly distributed weights, reflecting equal contributions from all interaction experts. Finally, the ENRICO dataset demonstrates a concentrated weight distribution with dominant experts for the screenshot modality. 6. In-depth Analysis of I2Mo E 6.1. Accuracy of Individual Experts To further analyze the effectiveness of I2Mo E, we compare its task performance against individual interaction experts across different datasets, as shown in Figure 5. The results highlight the following insights: ❶Across all datasets, the overall performance of I2Mo E-Mul T (red horizontal line) consistently surpasses that of any individual interaction expert expert, with performance gains of 2.2%, 1.3%, 7.1%, 0.6%, and 2.6% for ADNI, MIMIC, IMDB, MOSI, and ENRICO, respectively. This underscores the advantage of leveraging a mixture-of-experts approach over single-expert methods. ❷The proposed method exhibits the largest performance gains in datasets with high interaction importance distribution variability, such as MIMIC and ENRICO. While for more uniform datasets like MOSI, the performance of individual experts is closer to that of the overall model, indicating that the ensemble effect may be less pronounced in these cases. This suggests that the fusion of multiple experts becomes particularly beneficial in datasets with complex and heterogeneous multimodal interactions. 6.2. Interaction Expert Diversification To analyze the diversification of different interaction experts, we evaluate the ratio of expert agreement to disagreement and assess the corresponding accuracy of I2Mo E. A Interpretable Multimodal Interaction-aware Mixture-of-Experts Figure 4. Visualization of interaction weight distributions across all test samples for five datasets. Black bars indicate the median, mean, and extreme values. Figure 5. Comparison between the task performance of I2Mo E-Mul T (red horizontal line) and each individual interaction expert across different datasets. high proportion of disagreement among experts indicates greater diversity, which is essential for capturing distinct interaction patterns. Furthermore, when experts disagree, we expect I2Mo E to still maintain a high level of accuracy, demonstrating its ability to leverage diverse expert opinions effectively. Table 3 presents the proportion of cases where experts disagree or agree, along with the corresponding accuracy of I2Mo E across five datasets: ❶For ADNI and MIMIC datasets, the proportion of disagreement among experts is relatively high (81% and 85%, respectively), while I2Mo E achieves correct predictions in a substantial portion of these cases. ❷On the IMDB and ENRICO datasets, the proportion of disagreement is very high (99.99% and 98%), yet I2Mo E achieves significantly fewer correct predictions when experts disagree (15.85% Correct, 84.14% Wrong and 46.85% Correct, 51.44% Wrong). ❸For the MOSI dataset, the disagreement proportions (59%) highlight moderate diversity among experts. Notably, I2Mo E maintains relatively high accuracy when experts disagree (37.80% Correct for MOSI. These results indicate a potential need for better handling of disagreement in complex datasets, and how dataset characteristics influence the diversification and effectiveness of interaction experts. 7. Ablation Studies To validate the effectiveness of I2Mo E, we perform extensive ablation studies by systematically removing or modifying key components of the model. Each variant is designed to assess the contribution of specific design choices to the overall performance: (1) No-Interaction: The Table 3. Interaction experts agreement analysis on test set for all datasets. Disagree or Agree indicates whether all expert prediction is the same. ( Correct ) or ( Incorrect ) refers to the correctness of I2Mo E s prediction. % of Data ADNI MIMIC IMDB MOSI ENRICO Disagree, 48.74 63.51 15.85 37.80 46.85 Disagree, 32.40 21.39 84.14 21.97 51.44 Agree, 16.34 6.37 0.00 34.11 1.37 Agree, 2.52 8.73 0.01 6.12 0.34 interaction loss is removed, resulting in a simple mixture-ofexperts model without explicit encouragement for learning diverse multimodal interaction among experts. (2) Latent-Contrastive: The interaction loss is applied directly to the latent embeddings produced by each interaction expert instead of their outputs. (3) Simple-Weight: The MLP-based reweighting model is replaced by a shared, learnable global weight that does not adapt to individual samples. (4) Less-Forward: Perturbation is reduced by randomly masking only two modalities per sample instead of perturbing all modalities. (5) Synergy-Redundancy: Only synergy and redundancy experts are included, omitting uniqueness experts. From Table 7: ❶No-Interaction: Removing the interaction loss results in significant performance degradation across all datasets (e.g., -6.35% accuracy on ADNI and -3.99% AUROC), confirming that explicitly encouraging diversity among experts is crucial for capturing complementary modality interactions. ❷Latent-Contrastive: Applying the interaction loss to latent embeddings instead of expert outputs causes a noticeable performance drop (e.g., -6.91% accuracy on ADNI). This highlights the importance of applying the interaction loss at the output level to di- Interpretable Multimodal Interaction-aware Mixture-of-Experts Table 4. Ablation study results on three datasets (ADNI, MOSI, ENRICO), showing the impact of removing or modifying key components of I2Mo E. Each row corresponds to a variant of the model with a specific component ablated. Performance drops (in red) are reported relative to the full model. Dataset ADNI MOSI ENRICO Ablation Accuracy AUROC Accuracy Accuracy (1) 58.73 (-6.35) 77.10 (-3.99) 69.49 (-2.42) 47.63 (-0.59) (2) 58.17 (-6.91) 75.40 (-5.69) 69.68 (-2.23) 47.50 (-0.72) (3) 59.29 (-5.79) 74.55 (-6.54) 68.46 (-3.45) 47.49 (-0.73) (4) 59.76 (-5.32) 76.81 (-4.28) 69.89 (-2.02) 46.92 (-1.30) (5) 56.77 (-8.31) 74.30 (-6.79) 70.12 (-1.79) 47.49 (-0.73) rectly guide expert specialization. ❸Simple-Weight: Replacing the sample-specific reweighting model with a global weight reduces performance (e.g., -5.32% accuracy on ADNI and -1.30% on ENRICO), demonstrating the value of adaptive reweighting for leveraging diverse expert outputs effectively. ❹Less-Forward: Reducing modality perturbations leads to reduced accuracy (e.g., -5.79% on ADNI and -3.45% on MOSI). This suggests that generating sufficient negative examples through extensive perturbation is essential for capturing diverse interactions. ❺ Synergy-Redundancy: Limiting the experts to only synergy and redundancy results in the largest performance drop (e.g., -8.31% accuracy on ADNI). This emphasizes the importance of uniqueness experts in modeling comprehensive modality interactions. The ablation study demonstrates that each component of I2Mo E is vital for its success. 8. Conclusion We introduced I2Mo E, a novel Mo E framework designed to enhance multimodal task performance and interpretability by explicitly capturing heterogeneous modality interactions. Extensive experiments on five real-world datasets demonstrated the superiority of I2Mo E in improving performance across diverse multimodal scenarios. By leveraging a mixture-of-experts design with adaptive reweighting and specialized interaction losses, our approach systematically models and quantifies modality interactions. Additionally, we analyzed the distribution of interaction weights, providing meaningful insights at both the sample and dataset levels, which enhances the interpretability of the model s predictions. We also conducted ablation studies to evaluate the impact of each design component and demonstrated the flexibility of I2Mo E to generalize across various fusion methods. For future work, alternative forms of interaction loss could be explored to further improve performance. Additionally, integrating feature attribution methods to analyze the contributions of individual features within interaction experts can offer deeper interpretable insights. Acknowledgements This work was supported in part by NIH grants, RF1AG063481, R01AG071174, and U01CA274576. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH. We would like to thank the anonymous reviewers for their insightful feedback. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none of which we feel must be specifically highlighted here. Arevalo, J., Solorio, T., Montes-y G omez, M., and Gonz alez, F. A. Gated multimodal units for information fusion. ar Xiv preprint ar Xiv:1702.01992, 2017. Baltruˇsaitis, T., Ahuja, C., and Morency, L.-P. Multimodal machine learning: A survey and taxonomy. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41(2):423 443, 2019. URL https://ieeexplore. ieee.org/document/8269806. Barnum, G., Talukder, S., and Yue, Y. On the benefits of early fusion in multimodal representation learning, 2020. URL https://arxiv.org/abs/2011.07191. Bertschinger, N., Rauh, J., Olbrich, E., Jost, J., and Ay, N. Quantifying unique information. Entropy, 16(4):2161 2183, 2014. Chefer, H., Gur, S., and Wolf, L. Generic attention-model explainability for interpreting bi-modal and encoderdecoder transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 397 406, 2021. Chen, K., Xu, L., and Chi, H. Improved learning algorithms for mixture of experts in multiclass classification. Neural networks, 12(9):1229 1252, 1999. Dominici, G., Barbiero, P., Magister, L. C., Li o, P., and Simidjievski, N. Sharcs: Shared concept space for explainable multimodal learning. ar Xiv preprint ar Xiv:2307.00316, 2023. Doshi, J., Erus, G., Ou, Y., Resnick, S. M., Gur, R. C., Gur, R. E., Satterthwaite, T. D., Furth, S., Davatzikos, C., Initiative, A. N., et al. Muse: Multi-atlas region segmentation utilizing ensembles of registration algorithms and parameters, and locally optimal atlas selection. Neuroimage, 127:186 195, 2016. Interpretable Multimodal Interaction-aware Mixture-of-Experts Dufumier, B., Castillo-Navarro, J., Tuia, D., and Thiran, J.-P. What to align in multimodal contrastive learning? ar Xiv preprint ar Xiv:2409.07402, 2024. Esmaeilzadeh, S., Belivanis, D. I., Pohl, K. M., and Adeli, E. End-to-end alzheimer s disease diagnosis and biomarker identification. In Machine Learning in Medical Imaging: 9th International Workshop, MLMI 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Proceedings 9, pp. 337 345. Springer, 2018. Fan, C., Zhu, K., Tao, J., Yi, G., Xue, J., and Lv, Z. Multilevel contrastive learning: Hierarchical alleviation of heterogeneity in multimodal sentiment analysis. IEEE Transactions on Affective Computing, 2024. Fedus, W., Zoph, B., and Shazeer, N. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. Journal of Machine Learning Research, 23(120):1 39, 2022. Ghosh, S., Yu, K., Arabshahi, F., and Batmanghelich, K. Dividing and conquering a blackbox to a mixture of interpretable models: route, interpret, repeat. In Proceedings of the... International Conference on Machine Learning. International Conference on Machine Learning, volume 202, pp. 11360, 2023. Goldberger, A. L., Amaral, L. A., Glass, L., Hausdorff, J. M., Ivanov, P. C., Mark, R. G., Mietus, J. E., Moody, G. B., Peng, C.-K., and Stanley, H. E. Physiobank, physiotoolkit, and physionet: components of a new research resource for complex physiologic signals. circulation, 101(23): e215 e220, 2000. Ismail, A. A., Arik, S. O., Yoon, J., Taly, A., Feizi, S., and Pfister, T. Interpretable mixture of experts for structured data. ar Xiv preprint ar Xiv:2206.02107, 2022. Jacobs, R. A., Jordan, M. I., Nowlan, S. J., and Hinton, G. E. Adaptive mixtures of local experts. Neural computation, 3(1):79 87, 1991. Jin, P., Zhu, B., Yuan, L., and Yan, S. Moe++: Accelerating mixture-of-experts methods with zero-computation experts. ar Xiv preprint ar Xiv:2410.07348, 2024. Johnson, A., Bulgarelli, L., Pollard, T., Gow, B., Moody, B., Horng, S., Celi, L., and Mark, R. Mimic-iv (version 3.1). physionet, 2024. Johnson, A. E., Bulgarelli, L., Shen, L., Gayles, A., Shammout, A., Horng, S., Pollard, T. J., Hao, S., Moody, B., Gow, B., et al. Mimic-iv, a freely accessible electronic health record dataset. Scientific data, 10(1):1, 2023. Kim, C., van der Schaar, M., and Lee, C. Discovering features with synergistic interactions in multiple views. In Forty-first International Conference on Machine Learning. Kline, A., Wang, H., Li, Y., Dennis, S., Hutch, M., Xu, Z., Wang, F., Cheng, F., and Luo, Y. Multimodal machine learning in precision health: A scoping review. npj Digital Medicine, 5(1):171, 2022. Leiva, L. A., Hota, A., and Oulasvirta, A. Enrico: A dataset for topic modeling of mobile ui designs. In 22nd International Conference on Human-Computer Interaction with Mobile Devices and Services, pp. 1 4, 2020. Liang, P. P., Lyu, Y., Fan, X., Wu, Z., Cheng, Y., Wu, J., Chen, L. Y., Wu, P., Lee, M. A., Zhu, Y., et al. Multibench: Multiscale benchmarks for multimodal representation learning. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1), 2021. Liang, P. P., Lyu, Y., Chhablani, G., Jain, N., Deng, Z., Wang, X., Morency, L.-P., and Salakhutdinov, R. Multiviz: Towards visualizing and understanding multimodal models. ar Xiv preprint ar Xiv:2207.00056, 2022a. Liang, P. P., Zadeh, A., and Morency, L.-P. Foundations and trends in multimodal machine learning: Principles, challenges, and open questions. ar Xiv preprint ar Xiv:2209.03430, 2022b. URL https://arxiv. org/abs/2209.03430. Liang, P. P., Cheng, Y., Fan, X., Ling, C. K., Nie, S., Chen, R., Deng, Z., Allen, N., Auerbach, R., Mahmood, F., Salakhutdinov, R., and Morency, L.-P. Quantifying & modeling multimodal interactions: An information decomposition framework. ar Xiv preprint ar Xiv:2302.12247, 2023. URL https://arxiv. org/abs/2302.12247. Liang, P. P., Deng, Z., Ma, M. Q., Zou, J. Y., Morency, L.-P., and Salakhutdinov, R. Factorized contrastive learning: Going beyond multi-view redundancy. Advances in Neural Information Processing Systems, 36, 2024. Lin, X. V., Shrivastava, A., Luo, L., Iyer, S., Lewis, M., Ghosh, G., Zettlemoyer, L., and Aghajanyan, A. Moma: Efficient early-fusion pre-training with mixture of modality-aware experts. ar Xiv preprint ar Xiv:2407.21770, 2024. Liu, Z., Shen, Y., Lakshminarasimhan, V. B., Liang, P. P., Zadeh, A., and Morency, L.-P. Efficient low-rank multimodal fusion with modality-specific factors. ar Xiv preprint ar Xiv:1806.00064, 2018. Long, L., Cui, J., Zeng, P., Li, Y., Liu, Y., and Wang, Y. Mugi: Multi-granularity interactions of heterogeneous biomedical data for survival prediction. In International Interpretable Multimodal Interaction-aware Mixture-of-Experts Conference on Medical Image Computing and Computer Assisted Intervention, pp. 490 500. Springer, 2024. Lv, F., Chen, X., Huang, Y., Duan, L., and Lin, G. Progressive modality reinforcement for human multimodal emotion recognition from unaligned multimodal sequences. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2554 2562, June 2021. Lyu, Y., Liang, P. P., Deng, Z., Salakhutdinov, R., and Morency, L.-P. Dime: Fine-grained interpretations of multimodal models via disentangled local explanations. In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society, pp. 455 467, 2022. Mustafa, B., Riquelme, C., Puigcerver, J., Jenatton, R., and Houlsby, N. Multimodal contrastive learning with limoe: the language-image mixture of experts. Advances in Neural Information Processing Systems, 35:9564 9576, 2022. Park, D. H., Hendricks, L. A., Akata, Z., Rohrbach, A., Schiele, B., Darrell, T., and Rohrbach, M. Multimodal explanations: Justifying decisions and pointing to the evidence. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8779 8788, 2018. Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le, Q., Hinton, G., and Dean, J. Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. ar Xiv preprint ar Xiv:1701.06538, 2017. Swamy, V., Montariol, S., Blackwell, J., Frej, J. A., Jaggi, M., and K aser, T. Interpretcc: Intrinsic user-centric interpretability through global mixture of experts. ar Xiv preprint ar Xiv:2402.02933, 2024a. Swamy, V., Satayeva, M., Frej, J., Bossy, T., Vogels, T., Jaggi, M., K aser, T., and Hartley, M.-A. Multimodn multimodal, multi-task, interpretable modular networks. Advances in Neural Information Processing Systems, 36, 2024b. Teoh, J. R., Dong, J., Zuo, X., Lai, K. W., Hasikin, K., and Wu, X. Advancing healthcare through multimodal data fusion: a comprehensive review of techniques and applications. Peer J Computer Science, 10:e2298, 2024. Tsai, Y.-H. H., Bai, S., Liang, P. P., Kolter, J. Z., Morency, L.-P., and Salakhutdinov, R. Multimodal transformer for unaligned multimodal language sequences. In Proceedings of the conference. Association for computational linguistics. Meeting, volume 2019, pp. 6558. NIH Public Access, 2019. Tsai, Y.-H. H., Ma, M. Q., Yang, M., Salakhutdinov, R., and Morency, L.-P. Multimodal routing: Improving local and global interpretability of multimodal language analysis. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing, volume 2020, pp. 1823. NIH Public Access, 2020. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., and Polosukhin, I. Attention is all you need. nips 17. In Proceedings of the 31st International Conference on Neural Information Processing Systems December, pp. 6000 6010, 2017. Weiner, M. W., Aisen, P. S., Jack Jr, C. R., Jagust, W. J., Trojanowski, J. Q., Shaw, L., Saykin, A. J., Morris, J. C., Cairns, N., Beckett, L. A., et al. The alzheimer s disease neuroimaging initiative: progress report and future plans. Alzheimer s & Dementia, 6(3):202 211, 2010. Weiner, M. W., Veitch, D. P., Aisen, P. S., Beckett, L. A., Cairns, N. J., Green, R. C., Harvey, D., Jack Jr, C. R., Jagust, W., Morris, J. C., et al. The alzheimer s disease neuroimaging initiative 3: Continued innovation for clinical trial improvement. Alzheimer s & Dementia, 13(5): 561 571, 2017. Wenderoth, L., Hemker, K., Simidjievski, N., and Jamnik, M. Measuring cross-modal interactions in multimodal models. ar Xiv preprint ar Xiv:2412.15828, 2024. Wibral, M., Priesemann, V., Kay, J. W., Lizier, J. T., and Phillips, W. A. Partial information decomposition as a unified approach to the specification of neural goal functions. Brain and cognition, 112:25 38, 2017. Williams, P. L. and Beer, R. D. Nonnegative decomposition of multivariate information. ar Xiv preprint ar Xiv:1004.2515, 2010. Wollstadt, P., Schmitt, S., and Wibral, M. A rigorous information-theoretic definition of redundancy and relevancy in feature selection based on (partial) information decomposition. Journal of Machine Learning Research, 24(131):1 44, 2023. W ortwein, T., Sheeber, L., Allen, N., Cohn, J., and Morency, L.-P. Beyond additive fusion: Learning non-additive multimodal interactions. In Findings of the Association for Computational Linguistics: EMNLP 2022, pp. 4681 4696, 2022. W ortwein, T., Allen, N. B., Cohn, J. F., and Morency, L.- P. Smurf: Statistical modality uniqueness and redundancy factorization. In Proceedings of the 26th International Conference on Multimodal Interaction, pp. 339 349, 2024. Interpretable Multimodal Interaction-aware Mixture-of-Experts Xue, Z. and Marculescu, R. Dynamic multimodal fusion, 2023. URL https://arxiv.org/abs/2204. 00102. Yu, H., Qi, Z., Jang, L., Salakhutdinov, R., Morency, L.-P., and Liang, P. P. Mmoe: Enhancing multimodal models with mixtures of multimodal interaction experts. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pp. 10006 10030, 2024. Yuksel, S. E., Wilson, J. N., and Gader, P. D. Twenty years of mixture of experts. IEEE Transactions on Neural Networks and Learning Systems, 23(8):1177 1193, 2012. doi: 10.1109/TNNLS.2012.2200299. Yun, S., Choi, I., Peng, J., Wu, Y., Bao, J., Zhang, Q., Xin, J., Long, Q., and Chen, T. Flex-moe: Modeling arbitrary modality combination via the flexible mixture-of-experts. ar Xiv preprint ar Xiv:2410.08245, 2024. Zadeh, A., Zellers, R., Pincus, E., and Morency, L.-P. Mosi: multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos. ar Xiv preprint ar Xiv:1606.06259, 2016. Zadeh, A. B., Liang, P. P., Poria, S., Cambria, E., and Morency, L.-P. Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2236 2246, 2018. Zhang, Y., Doughty, H., and Snoek, C. Learning unseen modality interaction. Advances in Neural Information Processing Systems, 36:54716 54726, 2023. Interpretable Multimodal Interaction-aware Mixture-of-Experts A. The Connection between Interaction Loss and PID We link our perturbation-based losses to components in Partial Information Decomposition (PID), following Bertschinger et al. (2014): I(T; X1, X2) = Red(T; X1, X2) + Unq(T; X1 \ X2) + Unq(T; X2 \ X1) + Syn(T; X1, X2) (4) In the two-modality scenario, our model learns four experts, each trained to specialize in a PID component using corrupted modality inputs. Unique Information. Experts Funi1 and Funi2 are trained on inputs where the other modality is replaced with noise: Luni1 = Funi1(X1, X2) T , Luni2 = Funi2( X1, X2) T (5) Assuming Xi contains no task-relevant information, these losses approximate: Luni1 Unq(T; X1 \ X2), Luni2 Unq(T; X2 \ X1) (6) This aligns with unique information as defined by conditional information under fixed marginals (Bertschinger et al., 2014; Wollstadt et al., 2023). Redundant Information. Expert Fred is trained to match predictions from either single-modality input: Fred(X1, X2) T + Fred( X1, X2) T (7) This loss encourages Fred to extract information shared by both X1 and X2, approximating: Lred Red(T; X1, X2) (8) It aligns with redundancy defined via shared informativeness (Williams & Beer, 2010; Wollstadt et al., 2023). Synergistic Information. Expert Fsyn is trained to rely on both modalities jointly. It is penalized for performing well on any partial view: Fsyn(X1, X2) T Fsyn( X1, X2) T Fsyn(X1, X2) T (9) This loss isolates information that emerges only through joint modality interaction: Lsyn Syn(T; X1, X2) (10) This formulation reflects the formal synergy component as defined in Williams & Beer (2010); Wibral et al. (2017). By explicitly constructing perturbed input views that suppress or preserve specific modality contributions, each expert is trained to model a distinct PID component. This forms a contrastive approximation to the constrained information projections discussed in prior work (Bertschinger et al., 2014; Williams & Beer, 2010). B. Empirical Evidence for the Random Vector Masking The use of random vector replacement for modality dropout may appear ad hoc. However, our design is motivated by the need to fully suppress information from the dropped modality during interaction supervision. In contrast, alternatives such as mean or zero vector replacement risk preserving residual signals, which can undermine disentanglement of unique and redundant information pathways. Interpretable Multimodal Interaction-aware Mixture-of-Experts This decision is further supported by findings from Co MM (Dufumier et al., 2024), which highlight the regularization benefits and improved robustness of full modality dropout. To assess this empirically, we conducted an ablation comparing three masking strategies random, mean, and zero vector replacements across five datasets. The results (Table 5) show that random vector masking consistently yields stronger performance on most metrics and tasks. Table 5. Performance comparison across different modality masking strategies (Random, Mean, Zero). Metrics: Accuracy (Acc), AUROC, Micro/Macro F1. Numbers are reported as mean standard deviation. Dataset ADNI MIMIC IMDB MOSI ENRICO Metric Acc (3) AUROC Acc (2) AUROC Micro F1 (23) Macro F1 (23) Acc (2) Acc (20) Random 65.08 1.52 81.09 0.02 69.78 0.91 68.81 0.99 61.00 0.44 52.38 0.48 71.91 2.20 48.22 1.61 Mean 59.85 3.52 76.40 2.84 70.00 1.27 67.96 1.43 59.36 0.14 50.82 0.46 68.95 2.37 50.00 1.94 Zero 59.48 1.61 77.06 0.60 69.80 0.97 64.62 1.39 60.57 0.07 51.16 0.76 70.41 0.66 48.63 1.28 These results support our use of random vector masking as a more effective strategy for isolating and supervising interactionspecific information flow in multimodal learning. C. Complete Training Objective Let {Fi}B i=1 denote the B = n+2 interaction experts: n uniqueness experts, one synergy expert, and one redundancy expert. For each expert Fi, we obtain outputs from (1 + n) forward passes (one full input and one for each modality replaced): [ˆy(0) i , ˆy(1) i , . . . , ˆy(n) i ] = Fi.forward multiple(X1, . . . , Xn) The main prediction is computed as: i=1 wi ˆy(0) i , where [w1, . . . , w B] = MLPRe Weight(X1, . . . , Xn) The task loss is defined as: Ltask = ℓ(ˆy, T) We define the expert-specific interaction losses as follows: Uniqueness loss for each Fi (i = 1, . . . , n): L(i) int = 1 n 1 j =i Triplet Loss ˆy(0) i , ˆy(j) i , ˆy(i) i Synergy loss (Fn+1): L(n+1) int = 1 j=1 Cos Sim normalize(ˆy(0) n+1), normalize(ˆy(j) n+1) Redundancy loss (Fn+2): L(n+2) int = 1 1 Cos Sim normalize(ˆy(0) n+2), normalize(ˆy(j) n+2) We then average the interaction loss over all experts: i=1 L(i) int Interpretable Multimodal Interaction-aware Mixture-of-Experts The final training objective is: Ltotal = Ltask + λint Lint Model parameters are updated to minimize Ltotal. D. Computational Overhead and Scalability In theory, I2Mo E scales linearly with the number of input modalities. Specifically, the fusion overhead increases by approximately (Numer of modalities +2) times, corresponding to one uniqueness expert per modality, plus one redundancy and one synergy expert. To quantify the overhead of our method, we compare I2Mo E-Mul T with the Mul T baseline across three key metrics: training time per epoch (in seconds), inference latency (in seconds), and parameter count. As shown in Table 6, I2Mo E introduces moderate increases in compute roughly proportional to the number of modalities plus two (accounting for synergy and redundancy experts). All experiments were run on a single NVIDIA A100 GPU. Despite this additional cost, the model yields consistent improvements in interpretability and predictive performance, justifying the added overhead. Table 6. Comparison of Mul T and I2Mo E-Mul T on training time, inference latency, and model size across datasets. Train / epoch (s) Inference (s) # Params Dataset Modalities Mul T I2Mo E-Mul T Mul T I2Mo E-Mul T Mul T I2Mo E-Mul T ADNI I, G, C, B 8.98 0.04 16.82 0.02 1.34 0.00 2.29 0.00 1,072,131 6,696,728 MIMIC L, N, C 2.24 0.01 33.67 0.67 0.15 0.00 0.91 0.00 268,034 1,390,095 IMDB L, I 3.62 0.00 44.20 0.59 0.53 0.00 3.23 0.00 1,068,567 4,423,008 MOSI V, A, T 0.70 0.00 4.47 0.01 0.09 0.00 0.48 0.00 134,402 673,935 ENRICO S, W 1.38 0.02 6.17 0.03 0.20 0.00 0.44 0.00 538,644 2,352,724 E. Details for Dataset Preprocessing We followed the same preprocessing procedure of the ADNI dataset and MIMIC dataset, as described in Flex-Mo E (Yun et al., 2024). E.1. Detailed Data Preprocessing in ADNI Imaging, Genetic, Biospecimen, Clinical Modalities. The Alzheimer s Disease Initiative (ADNI) is a longitudinal multicenter observational study containing multi-modal data from subjects diagnosed as cognitively normal (CN), mild cognitive impairment (MCI), and Alzheimer s dementia (AD) (Weiner et al., 2010; 2017). In our experiments, we utilized imaging, genetic, biospecimen, and clinical modalities. The imaging data consisted of magnetic resonance images (MRIs) which were preprocessed using field intensity inhomogeneity correction, gray tissue matter segmentation via MUSE (Multiatlas Region Segmentation Utilizing Ensembles of Registration Algorithms and Parameters) (Doshi et al., 2016), and voxel-wise volumetric mapping of tissue regions. The genetic data consisted of SNP (single nucleotide polymorphisms) data from the ADNI 1, GO/2, and 3 studies. These were preprocessed via alignment to a unified reference, followed by aligning strands based on the 1000 Genome Project phase 3, linkage disequilibrium (LD) pruning, and imputation. The resulting data consisted of 144, 746 SNPs. The biospecimen modality included CSF Aβ1-42 and Aβ1-40, Total Tau and Phosphorylated Tau, Plasma Neurofilament Light Chain, and Apo E genotype. Clinical data included medical history, neurological exams, patient demographics, medications, and vital signs. Data columns directly containing Alzheimer s Disease diagnosis information were excluded. For both biospecimen and clinical data, numerical data was scaled using a Min Max scaler to a range of -1 to 1, while categorical data was one-hot encoded. Missing values, were imputed using the mean for numerical fields and the mode for categorical fields. E.2. Detailed Data Preprocessing in MIMIC Lab, Notes, Codes Modalities. The MIMIC dataset was extracted from the Medical Information Mart for Intensive Care IV (MIMIC-IV) database, which contains de-identified health data for patients who were admitted to either the emergency Interpretable Multimodal Interaction-aware Mixture-of-Experts department or stayed in critical care units of the Beth Israel Deaconess Medical Center in Boston, Massachusetts (Johnson et al., 2024; 2023; Goldberger et al., 2000). MIMIC-IV excludes patients under 18 years of age. We take a subset of the MIMIC-IV data, where each patient has at least more than 1 visit in the dataset as this subset corresponds to patients who likely have more serious health conditions. For each datapoint, we extract ICD-9 codes, clinical text, and labs and vital values. Using this data, we perform binary classification on one-year mortality. We drop visits that occur at the same time as the patient s death. F. Details for Modality-specific Encoder and Classification Head ❶ADNI Dataset: For the image modality, we employed a customized 3D-CNN (Esmaeilzadeh et al., 2018) with a hidden dimension of 256 as the encoder. For the genomics, clinical, and biospecimen modalities, we used a one-hidden-layer MLP with a hidden dimension of 256 as the encoder. ❷MIMIC Dataset: For all lab, note, and code modalities, we utilized an LSTM with a hidden dimension of 256 as the encoder. ❸MOSI Dataset: A Gated Recurrent Unit (GRU) with a hidden dimension of 256 was used as the encoder for the vision, audio, and text modalities. ❹ENRICO Dataset: For both the screenshot image and wireframe image modalities, we used VGG11 from the torchvision library with a hidden dimension size of 16 as the encoder. ❺IMDB Dataset: For the image modality, a VGG-16 model was applied as the feature extractor. For the language modality, features were extracted using the pretrained Google Word2vec model. Additionally, we employed VGG11 from the torchvision library with a hidden dimension size of 16 as the encoder and used Maxout Linear unimodal encoders, following current work (Liang et al., 2021). Classification Head: For all models and all datasets, we use a linear classification head to output the corresponding prediction. G. Details for Hyperparameter Setting To improve reproducibility, the tables below provide a summary of the hyperparameters used in our experiments. For hyperparameters of other baseline fusion methods, please refer to the scripts in the Git Hub repository at https://github. com/Raina-Xin/I2Mo E/tree/main/scripts/train_scripts. Table 7. Hyperparameter Configuration for I2Mo E-Mul T on Different Datasets Hyperparameter ADNI MIMIC IMDB MOSI ENRICO Learning Rate (lr) 0.0001 0.0001 0.0001 0.0001 0.0001 Temperature for Reweighting (temperature rw) 1 2 2.0 2.0 2.0 Hidden Dimension for Reweighting (hidden dim rw) 256 128 256 256 256 Number of Layers in Reweighting (num layer rw) 2 2 3 3 3 Interaction Loss Weight (interaction loss weight) 0.5 0.01 0.5 0.005 0.5 Modality (modality) IGCB LNC LI TVA SW Training Epochs (train epochs) 50 30 40 30 50 Batch Size (batch size) 32 32 32 32 32 Number of Experts (num experts) 8 4 4 4 4 Number of Layers in Encoder (num layers enc) 1 1 1 1 2 Number of Layers in Fusion (num layers fus) 2 2 2 1 2 Number of Layers in Prediction (num layers pred) 2 2 2 1 2 Number of Attention Heads (num heads) 4 1 4 1 4 Hidden Dimension (hidden dim) 256 128 256 256 256 Number of Patches (num patches) 16 8 4 4 8 Interpretable Multimodal Interaction-aware Mixture-of-Experts Table 8. Hyperparameter Configuration for I2Mo E-Switch Gate on Different Datasets Hyperparameter ADNI MIMIC IMDB MOSI ENRICO Learning Rate (lr) 0.0001 0.0001 0.0001 0.0001 0.0001 Temperature for Reweighting (temperature rw) 2 2 2.0 2.0 1 Hidden Dimension for Reweighting (hidden dim rw) 256 256 256 128 128 Number of Layers in Reweighting (num layer rw) 2 2 2 1 3 Interaction Loss Weight (interaction loss weight) 0.01 0.5 0.5 0.001 0.01 Modality (modality) IGCB LNC LI TVA SW Training Epochs (train epochs) 30 30 40 50 30 Batch Size (batch size) 8 64 64 32 8 Number of Experts (num experts) 16 16 16 4 4 Number of Layers in Encoder (num layers enc) 2 2 2 1 1 Number of Layers in Fusion (num layers fus) 2 2 2 1 1 Number of Layers in Prediction (num layers pred) 2 2 2 1 1 Number of Attention Heads (num heads) 4 4 4 4 2 Hidden Dimension (hidden dim) 128 256 128 128 128 Number of Patches (num patches) 8 16 4 16 4 Table 9. Hyperparameter Configuration for I2Mo E-Interpret CC on Different Datasets Hyperparameter ADNI MIMIC IMDB MOSI ENRICO Learning Rate (lr) 0.0001 0.0001 0.0001 0.0001 0.0001 Temperature for Reweighting (temperature rw) 2 2 2.0 1.5 4.0 Hidden Dimension for Reweighting (hidden dim rw) 128 128 256 256 256 Number of Layers in Reweighting (num layer rw) 2 2 3 2 2 Interaction Loss Weight (interaction loss weight) 0.5 0.1 0.01 0.001 0.5 Modality (modality) IGCB LNC LI TVA SW Tau (τ) 1.0 0.7 1.0 1.0 0.5 Threshold (threshold) 0.5 0.5 0.6 0.6 0.4 Train Epochs (train epochs) 30 50 40 50 60 Batch Size (batch size) 32 128 32 32 64 Hidden Dimension (hidden dim) 128 256 256 128 256 Hard (hard) True True True True True Table 10. Hyperparameter Configuration for I2Mo E-Mo E++ on Different Datasets Hyperparameter ADNI MIMIC IMDB MOSI ENRICO Learning Rate (lr) 0.0001 0.0001 0.0001 0.0001 0.0001 Temperature for Reweighting (temperature rw) 2 1 1.0 2 1 Hidden Dimension for Reweighting (hidden dim rw) 256 256 256 128 256 Number of Layers in Reweighting (num layer rw) 3 2 2 2 2 Interaction Loss Weight (interaction loss weight) 0.5 0.5 0.5 0.001 0.5 Modality (modality) IGCB LNC LI TVA SW Training Epochs (train epochs) 50 30 40 50 50 Batch Size (batch size) 64 32 32 32 32 Number of Experts (num experts) 8 4 4 8 8 Number of Layers in Encoder (num layers enc) 2 2 2 2 2 Number of Layers in Fusion (num layers fus) 2 2 2 1 2 Number of Layers in Prediction (num layers pred) 2 2 2 2 2 Number of Attention Heads (num heads) 4 4 4 4 4 Hidden Dimension (hidden dim) 256 128 256 64 64 Number of Patches (num patches) 8 4 8 4 4 H. Human Evaluation for Local Interpretation To strengthen evidence for the local interpretability of our model, we conducted a human evaluation study involving 15 participants. Each participant was shown 20 movie examples, resulting in a total of 300 interaction expert weight evaluations. Participants were asked to assess how reasonable the model s assigned expert weights were, choosing from a 5-point Likert scale: Completely makes sense, Mostly makes sense, Neutral, Makes little sense, and Makes no sense at all. Interpretable Multimodal Interaction-aware Mixture-of-Experts Overall, 70.4% of responses were positive (i.e., Mostly makes sense or Completely makes sense ), while only 9% were negative. Notably, just 0.7% of ratings selected the lowest option. These results suggest that the model s expert weight assignments are broadly viewed as reasonable and interpretable by human evaluators. The questionnaire and de-identified responses are available at https://github.com/Raina-Xin/I2Mo E/tree/main/ assets/human_eval Table 11. Distribution of human ratings for local interaction expert weights (n = 300). Response Option Percentage of Responses Completely makes sense 19.4% Mostly makes sense 51.0% Neutral 19.7% Makes little sense 9.0% Makes no sense at all 0.7% I. More Qualitative Examples for Local Interpretation We present a comprehensive visualization of all 23 classes in the IMDB dataset, illustrating local interpretability for individual examples. All examples are correctly predicted by I2Mo E. The Care Bears live in a country high in the clouds, where they have a lot of fun together. But they also do care for the human children on Earth, who they watch through huge telescopes from the sky, and come to help whenever there is need. Nikolas, a magician's apprentice, is in danger of getting under the influence of a bad spirit, which resides in an ancient spell book. The siblings Kim and Jason don't trust anyone anymore after being disappointed once too often. The Care Bears take them into their wonderland where they experience exciting and dangerous adventures together and quickly become good friends. Figure 6. IMDB example (ID: 0088885). Interpretable Multimodal Interaction-aware Mixture-of-Experts The history of the first victim of modern artillery and its moving agony, amidst conspiracies and betrayals of the powerful. Life and death of Giovanni De' Medici, a young brave captain in the war of Charles V against the Pope, in the first half of 1500. Figure 7. IMDB example (ID: 0245276). When Haseem arrives at Aladdin's home pretending to be his lost uncle, he brings the boy to a magical place which hides the entrance to a dangerous underground cave. There, Haseem asks of Aladdin to find but one simple oil lamp which contains a genie. As stubborn as he is, Aladdin refuses to come outside of the cave with the lamp and eventually discovers its secret. A genie inside the lamp then helps his life change and marry his true love. Figure 8. IMDB example (ID: 0827990).