# learning_factorized_multimodal_representations__29c0aa97.pdf Published as a conference paper at ICLR 2019 LEARNING FACTORIZED MULTIMODAL REPRESENTATIONS Yao-Hung Hubert Tsai 1, Paul Pu Liang 1, Amir Zadeh2, Louis-Philippe Morency2, Ruslan Salakhutdinov1 {1Machine Learning Department, 2Language Technologies Institute}, Carnegie Mellon University {yaohungt,pliang,abagherz,morency,rsalakhu}@cs.cmu.edu Learning multimodal representations is a fundamentally complex research problem due to the presence of multiple heterogeneous sources of information. Although the presence of multiple modalities provides additional valuable information, there are two key challenges to address when learning from multimodal data: 1) models must learn the complex intra-modal and cross-modal interactions for prediction and 2) models must be robust to unexpected missing or noisy modalities during testing. In this paper, we propose to optimize for a joint generative-discriminative objective across multimodal data and labels. We introduce a model that factorizes representations into two sets of independent factors: multimodal discriminative and modality-specific generative factors. Multimodal discriminative factors are shared across all modalities and contain joint multimodal features required for discriminative tasks such as sentiment prediction. Modality-specific generative factors are unique for each modality and contain the information required for generating data. Experimental results show that our model is able to learn meaningful multimodal representations that achieve state-of-the-art or competitive performance on six multimodal datasets. Our model demonstrates flexible generative capabilities by conditioning on independent factors and can reconstruct missing modalities without significantly impacting performance. Lastly, we interpret our factorized representations to understand the interactions that influence multimodal learning. 1 INTRODUCTION Multimodal machine learning involves learning from data across multiple modalities (Baltrušaitis et al., 2017). It is a challenging yet crucial research area with real-world applications in robotics (Liu et al., 2017), dialogue systems (Pittermann et al., 2010), intelligent tutoring systems (Petrovica et al., 2017), and healthcare diagnosis (Frantzidis et al., 2010). At the heart of many multimodal modeling tasks lies the challenge of learning rich representations from multiple modalities. For example, analyzing multimedia content requires learning multimodal representations across the language, visual, and acoustic modalities (Cho et al., 2015). Although the presence of multiple modalities provides additional valuable information, there are two key challenges to address when learning from multimodal data: 1) models must learn the complex intra-modal and cross-modal interactions for prediction (Zadeh et al., 2017), and 2) trained models must be robust to unexpected missing or noisy modalities during testing (Ngiam et al., 2011). In this paper, we propose to optimize for a joint generative-discriminative objective across multimodal data and labels. The discriminative objective ensures that the representations learned are rich in intra-modal and cross-modal features useful towards predicting the label, while the generative objective allows the model to infer missing modalities at test time and deal with the presence of noisy modalities. To this end, we introduce the Multimodal Factorization Model (MFM in Figure 1) that factorizes multimodal representations into multimodal discriminative factors and modality-specific generative factors. Multimodal discriminative factors are shared across all modalities and contain joint multimodal features required for discriminative tasks. Modality-specific generative factors are unique for each modality and contain the information required for generating each modality. We believe that factorizing multimodal representations into different explanatory factors can help each factor focus on learning from a subset of the joint information across multimodal data and labels. This method is in contrast to jointly learning a single factor that summarizes all generative and equal contributions Published as a conference paper at ICLR 2019 (a) (c) MFM Neural Architecture MFM Generative Network (b) MFM Inference Network Inference Network Generative Network Figure 1: Illustration of the proposed Multimodal Factorization Model (MFM) with three modalities. MFM factorizes multimodal representations into multimodal discriminative factors Fy and modality-specific generative factors Fa{1 M}. (a) MFM Generative Network with latent variables {Zy, Za{1 M}}, factors {Fy, Fa{1 M}}, generated multimodal data ˆX1 3 and labels ˆY. (b) MFM Inference Network. (c) MFM Neural Architecture. Best viewed zoomed in and in color. discriminative information (Srivastava & Salakhutdinov, 2012). To sum up, MFM defines a joint distribution over multimodal data, and by the conditional independence assumptions in the assumed graphical model, both generative and discriminative aspects are taken into account. Our model design further provides interpretability of the factorized representations. Through an extensive set of experiments, we show that MFM learns improved multimodal representations with these characteristics: 1) The multimodal discriminative factors achieve state-of-the-art or competitive performance on six multimodal time series datasets. We also demonstrate that MFM can generalize by integrating it with other existing multimodal discriminative models. 2) MFM allows flexible generation concerning multimodal discriminative factors (labels) and modality-specific generative factors (styles). We further show that we can perform reconstruction of missing modalities from observed modalities without significantly impacting discriminative performance. Finally, we interpret our learned representations using information-based and gradient-based methods, allowing us to understand the contributions of individual factors towards multimodal prediction and generation. 2 MULTIMODAL FACTORIZATION MODEL Multimodal Factorization Model (MFM) is a latent variable model (Figure 1(a)) with conditional independence assumptions over multimodal discriminative factors and modality-specific generative factors. According to these assumptions, we propose a factorization over the joint distribution of multimodal data (Section 2.1). Since exact posterior inference on this factorized distribution can be intractable, we propose an approximate inference algorithm based on minimizing a joint-distribution Wasserstein distance over multimodal data (Section 2.2). Finally, we derive the MFM objective by approximating the joint-distribution Wasserstein distance via a generalized mean-field assumption. Notation: We define X1 M as the multimodal data from M modalities and Y as the labels, with joint distribution PX1 M,Y = P(X1 M,Y). Let ˆX1 M denote the generated multimodal data and ˆY denote the generated labels, with joint distribution P ˆ X1 M, ˆ Y = P(ˆX1 M, ˆY). 2.1 FACTORIZED MULTIMODAL REPRESENTATIONS To factorize multimodal representations into multimodal discriminative factors and modality-specific generative factors, MFM assumes a Bayesian network structure as shown in Figure 1(a). In this graphical model, factors Fy and Fa{1 M} are generated from mutually independent latent variables Z = [Zy,Za{1 M}] with prior PZ. In particular, Zy generates the multimodal discriminative factor Fy and Za{1 M} generate modality-specific generative factors Fa{1 M}. By construction, Fy contributes to the generation of ˆY while {Fy,Fai} both contribute to the generation of ˆXi. As a result, the joint distribution P(ˆX1 M, ˆY) can be factorized as follows: P ( ˆ X1 M, ˆ Y) = F,Z P ( ˆ X1 M, ˆ Y F)P (F Z)P (Z)d Fd Z = Fy,Fa{1 M} Zy,Za{1 M} (P ( ˆ Y Fy) M i=1 P ( ˆ Xi Fai, Fy))(P (Fy Zy) M i=1 P (Fai Zai))(P (Zy) M i=1 P (Zai))d Fd Z, (1) with d F = d Fy M i=1 d Fai and d Z = d Zy M i=1 d Zai. Exact posterior inference in Equation 1 may be analytically intractable due to the integration over Z. We therefore resort to using an approximate inference distribution Q(Z X1 M,Y) as detailed in the Published as a conference paper at ICLR 2019 following subsection. As a result, MFM can be viewed as an autoencoding structure that consists of encoder (inference) and decoder (generative) modules (Figure 1(c)). The encoder module for Q( ) allows us to easily sample Z from an approximate posterior. The decoder modules are parametrized according to the factorization of P(ˆX1 M, ˆY Z) as given by Equation 1 and Figure 1(a). 2.2 MINIMIZING JOINT-DISTRIBUTION WASSERSTEIN DISTANCE OVER MULTIMODAL DATA Two common choices for approximate inference in autoencoding structures are Variational Autoencoders (VAEs) (Kingma & Welling, 2013) and Wasserstein Autoencoders (WAEs) (Zhao et al., 2017; Tolstikhin et al., 2017). The former optimizes the evidence lower bound objective (ELBO), and the latter derives an approximation for the primal form of the Wasserstein distance. We consider the latter since it simultaneously results in better latent factor disentanglement (Zhao et al., 2017; Rubenstein et al., 2018) and better sample generation quality than its counterparts (Chen et al., 2016; Higgins et al., 2016; Kingma & Welling, 2013). However, WAEs are designed for unimodal data and do not consider factorized distributions over latent variables that generate multimodal data. Therefore, we propose a variant for handling factorized joint distributions over multimodal data. As suggested by Kingma & Welling (2013), we adopt the design of nonlinear mappings (i.e. neural network architectures) in the encoder and decoder (Figure 1 (c)). For the encoder Q(Z X1 M,Y), we learn a deterministic mapping Qenc X1 M,Y Z (Rubenstein et al., 2018; Tolstikhin et al., 2017). For the decoder, we define the generation process from latent variables as Gy Zy Fy, Ga{1 M} Za{1 M} Fa{1 M}, D Fy ˆY, and F1 M Fy,Fa{1 M} ˆX1 M, where Gy,Ga{1 M},D and F1 M are deterministic functions parametrized by neural networks. Let Wc(PX1 M,Y,P ˆ X1 M, ˆ Y) denote the joint-distribution Wasserstein distance over multimodal data under cost function c Xi and c Y . We choose the squared cost c(a,b) = a b 2 2, allowing us to minimize the 2-Wasserstein distance. The cost function can be defined not only on static data but also on time series data such as text, audio and videos. For example, given time series data X = [X1,X2, ,XT ] and ˆX = [ ˆX1, ˆX2, , ˆXT ], we define c(X, ˆX) = T t=1 Xt ˆXt 2 2. With conditional independence assumptions in Equation 1, we express Wc(PX1 M,Y,P ˆ X1 M, ˆ Y) as: Proposition 1. For any functions Gy Zy Fy, Ga{1 M} Za{1 M} Fa{1 M}, D Fy ˆY, and F1 M Fa{1 M},Fy ˆX1 M, we have Wc(PX1 M,Y,P ˆ X1 M, ˆ Y) = inf QZ=PZ EPX1 M ,YEQ(Z X1 M ,Y) M i=1 c Xi(Xi, Fi(Gai(Zai), Gy(Zy))) + c Y (Y, D(Gy(Zy))) , (2) where PZ is the prior over Z = [Zy,Za{1,M}] and QZ is the aggregated posterior of the proposed approximate inference distribution Q(Z X1 M,Y). Proof: The proof is adapted from Tolstikhin et al. (Tolstikhin et al., 2017). The two differences are: (1) we show that P(ˆX1 M, ˆY Z = z) are Dirac for all z Z, and (2) we use the fact that c((X1 M,Y),(ˆX1 M, ˆY)) = M i=1 c Xi(Xi, ˆXi) + c Y (Y, ˆY). Please refer to the supplementary material for proof details. The constraint on QZ = PZ in Proposition 1 is hard to satisfy. To obtain a numerical solution, we first relax the constraint by performing a generalized mean field assumption on Q according to the conditional independence as shown in the inference network of Figure 1 (b): Q(Z X1 M, Y) = Q(Z X1 M) = Q(Zy X1 M) M i=1 Q(Zai Xi). (3) The intuition here is based on our design that Zy generates the multimodal discriminative factor Fy and Za{1 M} generate modality-specific generative factors Fa{1 M}. Therefore, the inference for Zy should depend on all modalities X1 M and the inference for Zai should depend only on the specific modality Xi. Following this assumption, we define Q as a nonparametric set of all encoders that fulfill the factorization in Equation 3. A penalty term is added into our objective to find the Q(Z ) Q that is the closest to prior PZ, thereby approximately enforcing the constraint QZ = PZ: min F,Ga{1 M},Gy,D inf Q(Z ) Q EPX1 M ,YEQ(Za1 X1) EQ(Za M XM )EQ(Zy X1 M ) M i=1 c Xi(Xi, F(Gai(Zai), Gy(Zy))) + c Y (Y, D(Gy(Zy))) + λMMD(QZ, PZ), (4) Published as a conference paper at ICLR 2019 nine (label) Method UM(SVHN) UM(MNIST) MM MFM Acc. 91.84 99.01 99.20 99.36 (b) Modality 1 (SVHN) Modality 2 (MNIST) (c) Figure 2: (a) MFM generative network for multimodal image dataset SVHN+MNIST, (b) unimodal and multimodal classification accuracies, and (c) conditional generation for SVHN and MNIST digits. MFM shows improved capabilities in digit prediction as well as flexible generation of both images based on labels and styles. where λ is a hyper-parameter and MMD is the Maximum Mean Discrepancy (Gretton et al., 2012) as a divergence measure between QZ and PZ. The prior PZ is chosen as a centered isotropic Gaussian N(0,I), so that it implicitly enforces independence between the latent variables Z = [Zy,Za{1,M}] (Higgins et al., 2016; Kingma & Welling, 2013; Rubenstein et al., 2018). Equation 4 represents our hybrid generative-discriminative optimization objective over multimodal data: the first loss term M i=1 c Xi(Xi,F(Gai(Zai),Gy(Zy))) is the generative objective based on reconstruction of multimodal data and the second term c Y (Y,D(Gy(Zy))) is the discriminative objective. In practice we compute the expectations in Equation 4 using empirical estimates over the training data. The neural architecture of MFM is illustrated in Figure 1(c). 2.3 SURROGATE INFERENCE FOR MISSING MODALITIES A key challenge in multimodal learning involves dealing with missing modalities. A good multimodal model should be able to infer the missing modality conditioned on the observed modalities and perform predictions based only on the observed modalities. To achieve this objective, the inference process of MFM can be easily adapted using a surrogate inference network to reconstruct the missing modality given the observed modalities. Formally, let Φ denote the surrogate inference network. The generation of missing modality ˆX1 given the observed modalities X2 M can be formulated as Φ = argmin Φ EPX2 M , ˆ X1 ( log PΦ( ˆX1 X2 M)) with PΦ( ˆX1 X2 M) = P( ˆX1 Za1,Zy)QΦ(Za1 X2 M)QΦ(Zy X2 M)d Za1d Zy. (5) Similar to Section 2.2, we use deterministic mappings in QΦ( ) and QΦ(Zy ) is also used for prediction PΦ( ˆY X2 M) = P( ˆY Zy)QΦ(Zy X2 M)d Zy. Equation 5 suggests that in the presence of missing modalities, we only need to infer the latent codes rather than the entire modality. 2.4 ENCODER AND DECODER DESIGN We now discuss the implementation choices for the MFM neural architecture in Figure 1(c). The encoder Q(Zy X1 M) can be parametrized by any model that performs multimodal fusion (Morency et al., 2011; Zadeh et al., 2017). For multimodal image datasets, we adopt Convolutional Neural Networks (CNNs) and Fully-Connected Neural Networks (FCNNs) with late fusion (Nojavanasghari et al., 2016) as our encoder Q(Zy X1 M). The remaining functions in MFM are also parametrized by CNNs and FCNNs. For multimodal time series datasets, we choose the Memory Fusion Network (MFN) (Zadeh et al., 2018a) as our multimodal encoder Q(Zy X1 M). We use Long Short-term Memory (LSTM) networks (Hochreiter & Schmidhuber, 1997) for functions Q(Za{1 M} X1 M), decoder LSTM networks (Cho et al., 2014) for functions F1 M, and FCNNs for functions Gy, Ga{1 M} and D. Details are provided in the appendix and the code is available at . 3 EXPERIMENTS In order to show that MFM learns multimodal representations that are discriminative, generative and interpretable, we design the following experiments. We begin with a multimodal synthetic image dataset that allows us to examine whether MFM displays discriminative and generative capabilities from factorized latent variables. Utilizing image datasets allows us to clearly visualize the generative capabilities of MFM. We then transition to six more challenging real-world multimodal video datasets Published as a conference paper at ICLR 2019 Table 1: Results for multimodal speaker traits recognition on POM, multimodal sentiment analysis on CMUMOSI, ICT-MMMO, You Tube, MOUD, and multimodal emotion recognition on IEMOCAP. SOTA1 and SOTA2 refer to the previous best and second best state-of-the-art respectively, and SOT A shows improvement over SOTA1. Symbols depict the baseline giving the result: # MFN, MARN, TFN, BC-LSTM, MV-LSTM, EF-LSTM, DF, SVM, RF. For detailed tables with results for all models, please refer to the appendix. Dataset POM Personality Traits Task Con Pas Voi Dom Cre Viv Exp Ent Res Tru Rel Out Tho Ner Per Hum Metric r SOTA2 0.359 0.425 0.166 0.235 0.358 0.417 0.450 0.378 0.295 0.237 0.215 0.238 0.363 0.258 0.344 0.319 SOTA1 0.395# 0.428# 0.193# 0.313# 0.367# 0.431# 0.452# 0.395# 0.333# 0.296# 0.255# 0.259# 0.381# 0.318# 0.377# 0.386# MFM 0.431 0.450 0.197 0.411 0.380 0.448 0.467 0.452 0.368 0.212 0.309 0.333 0.404 0.333 0.334 0.408 SOT A 0.036 0.022 0.004 0.097 0.013 0.017 0.015 0.057 0.035 0.054 0.074 0.023 0.015 0.022 Dataset CMU-MOSI ICT-MMMO You Tube MOUD Task Sentiment Sentiment Sentiment Sentiment Metric Acc_7 Acc_2 F1 MAE r Acc_2 F1 Acc_3 F1 Acc_2 F1 SOTA2 34.1# 77.1 77.0 0.968 0.625 72.5 72.6 48.3 45.1 81.1# 80.4# SOTA1 34.7 77.4# 77.3# 0.965# 0.632# 73.8# 73.1# 51.7# 51.6# 81.1 81.2 MFM 36.2 78.1 78.1 0.951 0.662 81.3 79.2 53.3 52.4 82.1 81.7 SOT A 1.5 0.7 0.8 0.014 0.030 7.5 6.1 1.6 0.8 1.0 0.5 Dataset IEMOCAP Emotions Task Happy Sad Angry Frustrated Excited Neutral Metric Acc_2 F1 Acc_2 F1 Acc_2 F1 Acc_2 F1 Acc_2 F1 Acc_2 F1 SOTA2 86.7 84.2 83.4 81.7 85.1 84.5 79.5 76.6 89.6 86.3# 68.8 67.1 SOTA1 90.1# 85.3# 85.8# 82.8 87.0# 86.0# 80.3# 76.8# 89.8# 87.1 71.8# 68.5 MFM 90.2 85.8 88.4 86.1 87.5 86.7 80.4 74.5 90.0 87.1 72.1 68.1 SOT A 0.1 0.5 2.6 3.3 0.5 0.7 0.1 0.2 0.3 to 1) rigorously evaluate the discriminative capabilities of MFM in comparison with existing baselines, 2) analyze the importance of each design component through ablation studies, 3) assess the robustness of MFM s modality reconstruction and prediction capabilities to missing modalities, and 4) interpret the learned representations using information-based and gradient-based methods to understand the contributions of individual factors towards multimodal prediction and generation. 3.1 MULTIMODAL SYNTHETIC IMAGE DATASET In this section, we study MFM on a synthetic image dataset that considers SVHN (Netzer et al., 2011) and MNIST (Lecun et al., 1998) as the two modalities. SVHN and MNIST are images with different styles but the same labels (digits 0 9). We randomly pair 100,000 SVHN and MNIST images that have the same label, creating a multimodal dataset which we call SVHN+MNIST. 80,000 pairs are used for training and the rest for testing. To justify that MFM is able to learn improved multimodal representations, we show both classification and generation results on SVHN+MNIST in Figure 2. Prediction: We perform experiments on both unimodal and multimodal classification tasks. UM denotes a unimodal baseline that performs prediction given only one modality as input and MM denotes a multimodal discriminative baseline that performs prediction given both images (Nojavanasghari et al., 2016). We compare the results for UM(SVHN), UM(MNIST), MM and MFM on SVHN+MNIST in Figure 2(b). We achieve better classification performance from unimodal to multimodal which is not surprising since more information is given. More importantly, MFM outperforms MM, which suggests that MFM learns improved factorized representations for discriminative tasks. Generation: We generate images using the MFM generative network (Figure 2(a)). We fix one variable out of Z = [Za1, Za2, and Zy] and randomly sample the other two variables from prior PZ. From Figure 2(c), we observe that MFM shows flexible generation of SVHN and MNIST images based on labels and styles. This suggests that MFM is able to factorize multimodal representations into multimodal discriminative factors (labels) and modality-specific generative factors (styles). 3.2 MULTIMODAL TIME SERIES DATASETS In this section, we transition to more challenging multimodal time series datasets. All the datasets consist of monologue videos. Features are extracted from the language (Glo Ve word embeddings (Pen- Published as a conference paper at ICLR 2019 Model Multimodal Hybrid Factorized Mod.-Spec. CMU-MOSI Disc. Gen.-Disc. Gen.-Disc. Gen. ˆX Reconstruction ˆY Prediction Factor Objective Factors Factors MSE (ℓ) MSE (a) MSE (v) Acc_7 Acc_2 F1 MAE r MA no no - - - - - 33.2 75.2 75.2 1.020 0.616 MB yes no - - - - - 34.1 77.4 77.3 0.965 0.632 MC no yes no - 0.0413 0.0509 0.0220 34.8 75.9 76.0 0.979 0.640 MD yes yes no - 0.0413 0.0486 0.0223 35.0 77.4 77.2 0.960 0.649 ME yes yes yes no 0.0397 0.0452 0.0211 35.9 77.3 77.2 0.956 0.661 MFM yes yes yes yes 0.0391 0.0384 0.0183 36.2 78.1 78.1 0.951 0.662 Figure 3: Models used in the ablation studies of MFM. Each model removes a design component from our model. Modality reconstruction and sentiment prediction results are reported on CMU-MOSI with best results in bold. Factorizing multimodal representations into multimodal discriminative factors and modality-specific generative factors are crucial for improved performance. nington et al., 2014)), visual (Facet (i Motions, 2017)), and acoustic (COVAREP (Degottex et al., 2014)) modalities. For a detailed description of feature extraction, please refer to the appendix. We consider the following six datasets across three domains: 1) Multimodal Personality Trait Recognition: POM (Park et al., 2014) contains 903 movie review videos annotated for the following personality traits: confident (con), passionate (pas), voice pleasant (voi), dominant (dom), credible (cre), vivid (viv), expertise (exp), entertaining (ent), reserved (res), trusting (tru), relaxed (rel), outgoing (out), thorough (tho), nervous (ner), persuasive (per) and humorous (hum). The short form is indicated in parenthesis. 2) Multimodal Sentiment Analysis: CMU-MOSI (Zadeh et al., 2016) is a collection of 2199 monologue opinion video clips annotated with sentiment. ICT-MMMO (Wöllmer et al., 2013) consists of 340 online social review videos annotated for sentiment. You Tube (Morency et al., 2011) contains 269 product review and opinion video segments from You Tube each annotated for sentiment. MOUD (Perez-Rosas et al., 2013) consists of 79 product review videos in Spanish. Each video consists of multiple segments labeled as either positive, negative or neutral sentiment. 3) Multimodal Emotion Recognition: IEMOCAP (Busso et al., 2008) consists of 302 videos of recorded dyadic dialogues. The videos are divided into multiple segments each annotated for the presence of 6 discrete emotions (happy, sad, angry, frustrated, excited and neutral), resulting in a total of 7318 segments in the dataset. We report results using the following metrics: Acc_C = multiclass accuracy across C classes, F1 = F1 score, MAE = Mean Absolute Error, r = Pearson s correlation. Prediction: We first compare the performance of MFM with existing multimodal prediction methods. For a detailed description of the baselines, please refer to the appendix. From Table 1, we first observe that the best performing baseline results are achieved by different models across different datasets (most notably MFN, MARN, and TFN). On the other hand, MFM consistently achieves state-of-the-art or competitive results for all six multimodal datasets. We believe that the multimodal discriminative factor Fy in MFM has successfully learned more meaningful representations by distilling discriminative features. This highlights the benefit of learning factorized multimodal representations towards discriminative tasks. Furthermore, MFM is model-agnostic and can be applied to other multimodal encoders Q(Zy X1 M). We perform experiments to show consistent improvements in discriminative performance for several choices of the encoder: EF-LSTM (Morency et al., 2011) and TFN (Zadeh et al., 2017). For Acc_2 on CMU-MOSI, our factorization framework improves the performance of EF-LSTM from 74.3 to 75.2 and TFN from 74.6 to 75.5. Ablation Study: In Figure 3, we present the models M{A,B,C,D,E} used for ablation studies. These models are designed to analyze the effects of using a multimodal discriminative factor, a hybrid generative-discriminative objective, factorized generative-discriminative factors and modality-specific generative factors towards both modality reconstruction and label prediction. The simplest variant is MA which represents a purely discriminative model without a joint multimodal discriminative factor (i.e. early fusion (Morency et al., 2011)). MB models a joint multimodal discriminative factor which incorporates more general multimodal fusion encoders (Zadeh et al., 2018a). MC extends Published as a conference paper at ICLR 2019 Table 2: The effect of missing modalities on multimodal data reconstruction and sentiment prediction on CMU-MOSI. MFM with surrogate inference is able to better handle missing modalities during test time as compared to the purely generative (Seq2Seq) or purely discriminative baselines. Task ˆX Reconstruction ˆY Prediction Metric MSE (ℓ) MSE (a) MSE (v) Acc_7 Acc_2 F1 MAE r Purely Generative and Discriminative Baselines ℓ(anguage) missing 0.0411 - - 19.4 59.6 59.7 1.386 0.225 a(udio) missing - 0.0533 - 34.0 73.5 73.4 1.024 0.615 v(isual) missing - - 0.0220 33.7 75.4 75.4 0.996 0.634 Multimodal Factorization Model (MFM) ℓ(anguage) missing 0.0403 - - 21.7 62.0 61.7 1.313 0.236 a(udio) missing - 0.0468 - 35.4 74.3 74.3 1.011 0.603 v(isual) missing - - 0.0215 35.0 76.4 76.3 0.990 0.635 all present 0.0391 0.0384 0.0182 36.2 78.1 78.1 0.951 0.662 MA by optimizing a hybrid generative-discriminative objective over modality-specific factors. MD extends MB by optimizing a hybrid generative-discriminative objective over a joint multimodal factor (resembling Srivastava & Salakhutdinov (2012)). ME factorizes the representation into separate generative and discriminative factors. Finally, MFM is obtained from ME by using modality-specific generative factors instead of a joint multimodal generative factor. From the table in Figure 3, we observe the following general trends. For sentiment prediction, using 1) a multimodal discriminative factor outperforms modality-specific discriminative factors (MD > MC, MB > MA), and 2) adding generative capabilities to the model improves performance (MC > MA, ME > MB). For both sentiment prediction and modality reconstruction, 3) factorizing into separate generative and discriminative factors improves performance (ME > MD), and 4) using modality-specific generative factors outperforms multimodal generative factors (MFM > ME). These observations support our design decisions of factorizing multimodal representations into multimodal discriminative factors and modality-specific generative factors. Missing Modalities: We now evaluate the performance of MFM in the presence of missing modalities using the surrogate inference model as described in Subsection 2.3. We compare with two baselines: 1) a purely generative Seq2Seq model (Cho et al., 2014) ΦG from observed modalities to missing modalities by optimizing EPX1 M ( log PΦD(X1 X2 M)), and 2) a purely discriminative model ΦD from observed modalities to the label by optimizing EPX2 M ,Y ( log PΦG(Y X2 M)). Both models are modified from MFM by using only the two observed modalities as input and not explicitly accounting for missing modalities. We compare the reconstruction error of each modality (language, visual and acoustic) as well as the performance on sentiment prediction. Table 2 shows that MFM with missing modalities outperforms the generative (ΦG) or discriminative baselines (ΦD) in terms of modality reconstruction and sentiment prediction. Additionally, MFM with missing modalities performs close to MFM with all modalities observed. This fact indicates that MFM can learn representations that are relatively robust to missing modalities. In addition, discriminative performance is most affected when the language modality is missing, which is consistent with prior work which indicates that language is most informative in human multimodal language (Zadeh et al., 2017). On the other hand, sentiment prediction is more robust to missing acoustic and visual features. Finally, we observe that reconstructing the low-level acoustic and visual features is easier as compared to the high-dimensional language features that contain high-level semantic meaning. Interpretation of Multimodal Representations: We devise two methods to study how individual factors in MFM influence the dynamics of multimodal prediction and generation. These interpretation methods represent both overall trends and fine-grained analysis that could be useful towards deeper understandings of multimodal representation learning. For more details, please refer to the appendix. Firstly, an information-based interpretation method is chosen to summarize the contribution of each modality towards the multimodal representations. Since Fy is a common cause of ˆX1 M, we can compare MI(Fy, ˆX1), ,MI(Fy, ˆXM), where MI( , ) denotes the mutual information measure between Fy and generated modality ˆXi. Higher MI(Fy, ˆXi) indicates greater contribution from Fy to ˆXi. Figure 4 reports the ratios ri = MI(Fy, ˆXi)/MI(Fai, ˆXi) which measure a normalized version of the mutual information between Fai and ˆXi. We observe that on CMU-MOSI, the language modality is most informative towards sentiment prediction, followed by the acoustic modality. We Published as a conference paper at ICLR 2019 Ratio rℓ rv ra CMU-MOSI 0.307 0.030 0.107 Figure 4: Analyzing the multimodal representations learnt in MFM via information-based (entire dataset) and gradient-based interpretation methods (single video) on CMU-MOSI. Umm, in a way, a lot of the themes in never let me go , which were very profound and deep. (hesitancy) (emphasis) (neutral) (uninformative) (uninformative) (slight smile) believe that this result represents a prior over the expression of sentiment in human multimodal language and is closely related to the connections between language and speech (Kuhl, 2000). Secondly, a gradient-based interpretation method to used analyze the contribution of each modality for every time step in multimodal time series data. We measure the gradient of the generated modality with respect to the target factors (e.g., Fy). Let {x1,x2, ,x M} denote multimodal time series data where xi represents modality i, and ˆxi = [ˆx1 i , , ˆxt i, , ˆx T i ] denote generated modality i across time steps t [1,T]. The gradient fy(ˆxi) measures the extent to which changes in factor fy P(Fy X1 M = x1 M) influences the generation of sequence ˆxi. Figure 4 plots fy(ˆxi) for a video in CMU-MOSI. We observe that multimodal communicative behaviors that are indicative of speaker sentiment such as positive words (e.g. very profound and deep ) and informative acoustic features (e.g. hesitant and emphasized tone of voice) indeed correspond to increases in fy(ˆxi). 4 RELATED WORK The two main pillars of research in multimodal representation learning have considered the discriminative and generative objectives individually. Discriminative representation learning (Liang et al., 2018; Chen et al., 2017; Chaplot et al., 2017; Frome et al., 2013; Socher et al., 2013; Tsai et al., 2017) models the conditional distribution P(Y X1 M). Since these approaches are not concerned with modeling P(X1 M) explicitly, they use parameters more efficiently to model P(Y X1 M). For instance, recent works learn visual representations that are maximally dependent with linguistic attributes for improving one-shot image recognition (Tsai & Salakhutdinov, 2017) or introduce tensor product mechanisms to model interactions between the language, visual and acoustic modalities (Liu et al., 2018; Zadeh et al., 2017). On the other hand, generative representation learning captures the interactions between modalities by modeling the joint distribution P(X1, ,XM) using either undirected graphical models (Srivastava & Salakhutdinov, 2012), directed graphical models (Suzuki et al., 2016b), or neural networks (Sohn et al., 2014). Some generative approaches compress multimodal data into lower-dimensional feature vectors which can be used for discriminative tasks (Pham et al., 2018; Ngiam et al., 2011). To unify the advantages of both approaches, MFM factorizes multimodal representations into generative and discriminative components and optimizes for a joint objective. Factorized representation learning resembles learning disentangled data representations which have been shown to improve the performance on many tasks (Kulkarni et al., 2015; Lake et al., 2017; Higgins et al., 2016; Bengio et al., 2013). Several methods involve specifying a fixed set of latent attributes that individually control particular variations of data and performing supervised training (Cheung et al., 2014; Karaletsos et al., 2015; Yang et al., 2015; Reed et al., 2014; Zhu et al., 2014), assuming an isotropic Gaussian prior over latent variables to learn disentangled generative representations (Kingma & Welling, 2013; Rubenstein et al., 2018) and learning latent variables in charge of specific variations in the data by maximizing the mutual information between a subset of latent variables and the data (Chen et al., 2016). However, these methods study factorization of a single modality. MFM factorizes multimodal representations and demonstrates the importance of modality-specific and multimodal factors towards generation and prediction. A concurrent and parallel work that factorizes latent factors in multimodal data was proposed by Hsu & Glass (2018). They differ from us in the graphical model design, discriminative objective, prior matching criterion, and scale of experiments. We provide a detailed comparison with their model in the appendix. 5 CONCLUSION In this paper, we proposed the Multimodal Factorization Model (MFM) for multimodal representation learning. MFM factorizes the multimodal representations into two sets of independent factors: multimodal discriminative factors and modality-specific generative factors. The multimodal discriminative Published as a conference paper at ICLR 2019 factor achieves state-of-the-art or competitive results on six multimodal datasets. The modalityspecific generative factors allow us to generate data based on factorized variables, account for missing modalities, and have a deeper understanding of the interactions involved in multimodal learning. Our future work will explore extensions of MFM for video generation, semi-supervised learning, and unsupervised learning. We believe that MFM sheds light on the advantages of learning factorizing multimodal representations and potentially opens up new horizons for multimodal machine learning. ACKNOWLEDGEMENTS This work was supported in part by the DARPA grants D17AP00001 and FA875018C0150, Office of Naval Research, Apple, and Google focused award. We would also like to acknowledge NVIDIA s GPU support. This material is also based upon work partially supported by the National Science Foundation (Award 1750439). Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of National Science Foundation, and no official endorsement should be inferred. Tadas Baltrušaitis, Chaitanya Ahuja, and Louis-Philippe Morency. Multimodal machine learning: A survey and taxonomy. ar Xiv preprint ar Xiv:1705.09406, 2017. Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell., 35(8), August 2013. Leo Breiman. Random forests. Mach. Learn., 45(1):5 32, October 2001. ISSN 0885-6125. doi: 10.1023/A:1010933404324. URL https://doi.org/10.1023/A:1010933404324. Carlos Busso, Murtaza Bulut, Chi-Chun Lee, Abe Kazemzadeh, Emily Mower, Samuel Kim, Jeannette Chang, Sungbok Lee, and Shrikanth S. Narayanan. Iemocap: Interactive emotional dyadic motion capture database. Journal of Language Resources and Evaluation, 2008. Devendra Singh Chaplot, Kanthashree Mysore Sathyendra, Rama Kumar Pasumarthi, Dheeraj Rajagopal, and Ruslan Salakhutdinov. Gated-attention architectures for task-oriented language grounding. ar Xiv preprint ar Xiv:1706.07230, 2017. Minghai Chen, Sen Wang, Paul Pu Liang, Tadas Baltrušaitis, Amir Zadeh, and Louis-Philippe Morency. Multimodal sentiment analysis with word-level fusion and reinforcement learning. In Proceedings of the 19th ACM International Conference on Multimodal Interaction, ICMI, 2017. Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In Advances in Neural Information Processing Systems, pp. 2172 2180, 2016. Brian Cheung, Jesse A Livezey, Arjun K Bansal, and Bruno A Olshausen. Discovering hidden factors of variation in deep networks. ar Xiv preprint ar Xiv:1412.6583, 2014. Kyunghyun Cho, Bart van Merriënboer, Ça glar Gülçehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. Learning phrase representations using rnn encoder decoder for statistical machine translation. In EMNLP. ACL, 2014. Kyung Hyun Cho, Aaron C. Courville, and Yoshua Bengio. Describing multimedia content using attention-based encoder-decoder networks. Co RR, abs/1507.01053, 2015. URL http://arxiv. org/abs/1507.01053. Marco Cuturi, Jean-Philippe Vert, Oystein Birkenes, and Tomoko Matsui. A kernel for time series based on global alignments. In Acoustics, Speech and Signal Processing, ICASSP 2007., 2007. Gilles Degottex, John Kane, Thomas Drugman, Tuomo Raitio, and Stefan Scherer. Covarepa collaborative voice analysis repository for speech technologies. In ICASSP. IEEE, 2014. Paul Ekman. An argument for basic emotions. Cognition & emotion, 6(3-4):169 200, 1992. Published as a conference paper at ICLR 2019 Paul Ekman, Wallace V Freisen, and Sonia Ancoli. Facial signs of emotional experience. Journal of personality and social psychology, 39(6):1125, 1980. C. A. Frantzidis, C. Bratsas, M. A. Klados, E. Konstantinidis, C. D. Lithari, A. B. Vivas, C. L. Papadelis, E. Kaldoudi, C. Pappas, and P. D. Bamidis. On the classification of emotional biosignals evoked while viewing affective pictures: An integrated data-mining-based approach for healthcare applications. IEEE Transactions on Information Technology in Biomedicine, March 2010. Andrea Frome, Greg S Corrado, Jon Shlens, Samy Bengio, Jeff Dean, Tomas Mikolov, et al. Devise: A deep visual-semantic embedding model. In NIPS, 2013. A. Graves, A. r. Mohamed, and G. Hinton. Speech recognition with deep recurrent neural networks. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, May 2013. Arthur Gretton, Olivier Bousquet, Alex Smola, and Bernhard Schölkopf. Measuring statistical dependence with hilbert-schmidt norms. In International conference on algorithmic learning theory, pp. 63 77. Springer, 2005. Arthur Gretton, Karsten M Borgwardt, Malte J Rasch, Bernhard Schölkopf, and Alexander Smola. A kernel two-sample test. Journal of Machine Learning Research, 13(Mar):723 773, 2012. Sylvain Guimond and Wael Massrieh. Intricate correlation between body posture, personality trait and incidence of body pain: A cross-referential study report. PLOS ONE, 2012. Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. beta-vae: Learning basic visual concepts with a constrained variational framework. 2016. Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. beta-vae: Learning basic visual concepts with a constrained variational framework. In ICLR, 2017. Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8): 1735 1780, 1997. Wei-Ning Hsu and James Glass. Disentangling by partitioning: A representation learning framework for multimodal sensory data. ar Xiv preprint ar Xiv:1805.11264, 2018. i Motions. Facial expression analysis, 2017. URL goo.gl/1rh1JN. Theofanis Karaletsos, Serge Belongie, and Gunnar Rätsch. Bayesian representation learning with oracle constraints. ar Xiv preprint ar Xiv:1506.05011, 2015. Diederik P Kingma and Max Welling. Auto-encoding variational bayes. ar Xiv preprint ar Xiv:1312.6114, 2013. Patricia K. Kuhl. A new view of language acquisition. Proceedings of the National Academy of Sciences, 2000. doi: 10.1073/pnas.97.22.11850. Tejas D Kulkarni, William F Whitney, Pushmeet Kohli, and Josh Tenenbaum. Deep convolutional inverse graphics network. In Advances in Neural Information Processing Systems, 2015. Brenden M Lake, Tomer D Ullman, Joshua B Tenenbaum, and Samuel J Gershman. Building machines that learn and think like people. Behavioral and Brain Sciences, 40, 2017. Yann Lecun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. In Proceedings of the IEEE, pp. 2278 2324, 1998. Sergey Levine, Christian Theobalt, and Vladlen Koltun. Real-time prosody-driven synthesis of body language. ACM Trans. Graph., 28(5):172:1 172:10, December 2009. Paul Pu Liang, Amir Zadeh, and Louis-Philippe Morency. Multimodal local-global ranking fusion for emotion recognition. In Proceedings of the 20th ACM International Conference on Multimodal Interaction, ICMI, 2018. Published as a conference paper at ICLR 2019 Z. Liu, M. Wu, W. Cao, L. Chen, J. Xu, R. Zhang, M. Zhou, and J. Mao. A facial expression emotion recognition based human-robot interaction system. IEEE/CAA Journal of Automatica Sinica, 4(4): 668 676, 2017. ISSN 2329-9266. doi: 10.1109/JAS.2017.7510622. Zhun Liu, Ying Shen, Varun Bharadhwaj Lakshminarasimhan, Paul Pu Liang, Amir Ali Bagher Zadeh, and Louis-Philippe Morency. Efficient low-rank multimodal fusion with modality-specific factors. In ACL, 2018. Gelareh Mohammadi, Alessandro Vinciarelli, and Marcello Mortillaro. The voice of personality: Mapping nonverbal vocal behavior into trait attributions. In Proceedings of ACM Multimedia Workshop on Social Signal Processing, 0 2010. Louis-Philippe Morency, Ariadna Quattoni, and Trevor Darrell. Latent-dynamic discriminative models for continuous gesture recognition. In CVPR. IEEE, 2007. Louis-Philippe Morency, Rada Mihalcea, and Payal Doshi. Towards multimodal sentiment analysis: Harvesting opinions from the web. In ICMI, pp. 169 176. ACM, 2011. Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y. Ng. Reading digits in natural images with unsupervised feature learning. 2011. Jiquan Ngiam, Aditya Khosla, Mingyu Kim, Juhan Nam, Honglak Lee, and Andrew Y Ng. Multimodal deep learning. In (ICML-11), 2011. Behnaz Nojavanasghari, Deepak Gopinath, Jayanth Koushik, Tadas Baltrušaitis, and Louis-Philippe Morency. Deep multimodal fusion for persuasiveness prediction. In Proceedings of the 18th ACM International Conference on Multimodal Interaction, ICMI 2016. ACM, 2016. Sunghyun Park, Han Suk Shim, Moitreya Chatterjee, Kenji Sagae, and Louis-Philippe Morency. Computational analysis of persuasiveness in social multimedia: A novel dataset and multimodal prediction approach. ICMI 14, 2014. Jeffrey Pennington, Richard Socher, and Christopher D Manning. Glove: Global vectors for word representation. In EMNLP, volume 14, pp. 1532 1543, 2014. Veronica Perez-Rosas, Rada Mihalcea, and Louis-Philippe Morency. Utterance-Level Multimodal Sentiment Analysis. In Association for Computational Linguistics (ACL), August 2013. Sintija Petrovica, Alla Anohina-Naumeca, and HazÄ m Kemal Ekenel. Emotion recognition in affective tutoring systems: Collection of ground-truth data. Procedia Computer Science, 2017. Hai Pham, Thomas Manzini, Paul Pu Liang, and Barnabas Poczos. Seq2seq2sentiment: Multimodal sequence to sequence models for sentiment analysis. In Proceedings of Grand Challenge and Workshop on Human Multimodal Language (Challenge-HML). ACL, 2018. Johannes Pittermann, Angela Pittermann, and Wolfgang Minker. Emotion recognition and adaptation in spoken dialogue systems. International Journal of Speech Technology, 13(1):49 60, Mar 2010. Soujanya Poria, Erik Cambria, Devamanyu Hazarika, Navonil Majumder, Amir Zadeh, and Louis Philippe Morency. Context-dependent sentiment analysis in user-generated videos. In ACL, 2017. Ariadna Quattoni, Sybor Wang, Louis-Philippe Morency, Michael Collins, and Trevor Darrell. Hidden conditional random fields. IEEE Trans. Pattern Anal. Mach. Intell., 2007. Shyam Sundar Rajagopalan, Louis-Philippe Morency, Tadas Baltrušaitis, and Goecke Roland. Extending long short-term memory for multi-view structured learning. In ECCV, 2016. Scott Reed, Kihyuk Sohn, Yuting Zhang, and Honglak Lee. Learning to disentangle factors of variation with manifold interaction. In International Conference on Machine Learning, 2014. Paul K Rubenstein, Bernhard Schoelkopf, and Ilya Tolstikhin. On the latent space of wasserstein auto-encoders. ar Xiv preprint ar Xiv:1802.03761, 2018. Published as a conference paper at ICLR 2019 M. Schuster and K.K. Paliwal. Bidirectional recurrent neural networks. Trans. Sig. Proc., 45(11), November 1997. Richard Socher, Milind Ganjoo, Christopher D Manning, and Andrew Ng. Zero-shot learning through cross-modal transfer. In Advances in neural information processing systems, pp. 935 943, 2013. Kihyuk Sohn, Wenling Shang, and Honglak Lee. Improved multimodal deep learning with variation of information. In Advances in Neural Information Processing Systems, pp. 2141 2149, 2014. Yale Song, Louis-Philippe Morency, and Randall Davis. Multi-view latent variable discriminative models for action recognition. In CVPR, 2012. Yale Song, Louis-Philippe Morency, and Randall Davis. Action recognition by hierarchical sequence summarization. In CVPR, 2013. Nitish Srivastava and Ruslan R Salakhutdinov. Multimodal learning with deep boltzmann machines. In Advances in neural information processing systems, pp. 2222 2230, 2012. Masashi Sugiyama and Makoto Yamada. On kernel parameter selection in hilbert-schmidt independence criterion. IEICE TRANSACTIONS on Information and Systems, 2012. Masahiro Suzuki, Kotaro Nakayama, and Yutaka Matsuo. Joint multimodal learning with deep generative models. Co RR, abs/1611.01891, 2016a. Masahiro Suzuki, Kotaro Nakayama, and Yutaka Matsuo. Joint multimodal learning with deep generative models. ar Xiv preprint ar Xiv:1611.01891, 2016b. Ilya Tolstikhin, Olivier Bousquet, Sylvain Gelly, and Bernhard Schoelkopf. Wasserstein auto-encoders. ar Xiv preprint ar Xiv:1711.01558, 2017. Yao-Hung Hubert Tsai and Ruslan Salakhutdinov. Improving one-shot learning through fusing side information. ar Xiv preprint ar Xiv:1710.08347, 2017. Yao-Hung Hubert Tsai, Liang-Kang Huang, and Ruslan Salakhutdinov. Learning robust visualsemantic embeddings. ar Xiv preprint ar Xiv:1703.05908, 2017. Martin Wöllmer, Felix Weninger, Tobias Knaup, Björn Schuller, Congkai Sun, Kenji Sagae, and Louis-Philippe Morency. Youtube movie reviews: Sentiment analysis in an audio-visual context. IEEE Intelligent Systems, 28(3):46 53, 2013. Denny Wu, Yixiu Zhao, Yao-Hung Hubert Tsai, Makoto Yamada, and Ruslan Salakhutdinov. " dependency bottleneck" in auto-encoding architectures: an empirical study. ar Xiv preprint ar Xiv:1802.05408, 2018. Jimei Yang, Scott E Reed, Ming-Hsuan Yang, and Honglak Lee. Weakly-supervised disentangling with recurrent transformations for 3d view synthesis. In NIPS, 2015. Jiahong Yuan and Mark Liberman. Speaker identification on the scotus corpus. Journal of the Acoustical Society of America, 123(5):3878, 2008. Amir Zadeh, Rowan Zellers, Eli Pincus, and Louis-Philippe Morency. Multimodal sentiment intensity analysis in videos: Facial gestures and verbal messages. IEEE Intelligent Systems, 2016. Amir Zadeh, Minghai Chen, Soujanya Poria, Erik Cambria, and Louis-Philippe Morency. Tensor fusion network for multimodal sentiment analysis. In EMNLP, 2017. Amir Zadeh, Paul Pu Liang, Navonil Mazumder, Soujanya Poria, Erik Cambria, and Louis-Philippe Morency. Memory fusion network for multi-view sequential learning. AAAI, 2018a. Amir Zadeh, Paul Pu Liang, Soujanya Poria, Prateek Vij, Erik Cambria, and Louis-Philippe Morency. Multi-attention recurrent network for human communication comprehension. AAAI, 2018b. Matthew D. Zeiler, Dilip Krishnan, Graham W. Taylor, and Rob Fergus. Deconvolutional networks. 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 2528 2535, 2010. Published as a conference paper at ICLR 2019 Shengjia Zhao, Jiaming Song, and Stefano Ermon. Infovae: Information maximizing variational autoencoders. ar Xiv preprint ar Xiv:1706.02262, 2017. Zhenyao Zhu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Multi-view perceptron: a deep model for learning face identity and view representations. In NIPS, 2014. A PROOF OF PROPOSITION 1 To simplify the proof, we first prove it for the unimodal case by considering the Wasserstein distance between PX,Y and P ˆ X, ˆ Y. A.1 UNIMODAL JOINT-DISTRIBUTION WASSERSTEIN DISTANCE Proposition 2. For any functions Gy Zy Fy, Ga Za Fa, D Fy ˆY, and F Fa,Fy ˆX, we have Wc(PX,Y,P ˆ X, ˆ Y) = inf QZ=PZ EPX,YEQ(Z X) c X(X,F(Ga(Za),Gy(Zy))) + c Y (Y,D(Gy(Zy))) , where Wc is the Wasserstein distance under cost function c X and c Y , PZ is the prior over Z = [Za,Zy] and QZ is the aggregated posterior of the proposed inference distribution Q(Z X). Proof: See the following. To begin the proof, we abuse some notations as follows. By definition, the Wasserstein distance under cost function c between PX,Y and P ˆ X, ˆ Y is Wc(PX,Y, P ˆ X, ˆ Y) = inf Γ P((X,Y) PX,Y,( ˆ X, ˆ Y) P ˆ X, ˆ Y) E((X,Y),( ˆ X, ˆ Y)) Γ[c((X, Y), ( ˆX, ˆY))], (7) where c((X,Y),(ˆX, ˆY)) (X,Y) (X,Y) R+ is any measurable cost function. P((X,Y) PX,Y,(ˆX, ˆY) P ˆ X, ˆ Y) is the set of all joint distributions of ((X,Y),(ˆX, ˆY)) with marginals PX,Y and P ˆ X, ˆ Y, respectively. Note that c((X,Y),(ˆX, ˆY)) = c X(X, ˆX) + c Y (Y, ˆY). Next, we denote the set of all joint distributions of (X,Y, ˆX, ˆY,Z) such that (X,Y) PX,Y, (ˆX, ˆY,Z) P ˆ X, ˆ Y,Z, and ((X,Y) (ˆX, ˆY) Z) as PX,Y, ˆ X, ˆ Y,Z. PX,Y, ˆ X, ˆ Y and PX,Y,Z are the sets of the marginals (X,Y, ˆX, ˆY) and (X,Y,Z) induced by PX,Y, ˆ X, ˆ Y,Z. We now introduce two Lemmas to help the proof. Lemma 1. P(ˆX, ˆY Z = z) are Dirac for all z Z. Proof: First, we have ˆX = F(Ga(Za),Gy(Zy)) and ˆY = D(Gy(Zy)) with Z = {Za,Zy}. Since the functions F,Ga,Gy,D are all deterministic, then P(ˆX, ˆY Z) are Dirac measures. Lemma 2. P(PX,Y,P ˆ X, ˆ Y) = PX,Y, ˆ X, ˆ Y when P(ˆX, ˆY Z = z) are Dirac for all z Z. Proof: When ˆX, ˆY are deterministic functions of Z, for any A in the sigma-algebra induced by ˆX, ˆY, we have E[I[ ˆ X, ˆ Y A] X,Y,Z] = E[I[ ˆ X, ˆ Y A] Z]. Therefore, this implies that (X,Y) (ˆX, ˆY) Z which concludes the proof. A similar argument is made in Lemma 1 of (Tolstikhin et al., 2017). Published as a conference paper at ICLR 2019 Now, we use the fact that P(PX,Y,P ˆ X, ˆ Y) = PX,Y, ˆ X, ˆ Y (Lemma 1 + Lemma 2), c((X,Y),(ˆX, ˆY)) = c X(X, ˆX) + c Y (Y, ˆY), ˆX = F(Ga(Za),Gy(Zy)), and ˆY = D(Gy(Zy)), Eq. equation 7 becomes inf P PX,Y, ˆ X, ˆ Y EX,Y, ˆ X, ˆ Y P [c X(X, ˆX) + c Y (Y, ˆY)] = inf P PX,Y, ˆ X, ˆ Y,Z EX,Y, ˆ X, ˆ Y,Z P [c X(X, ˆX) + c Y (Y, ˆY)] = inf P PX,Y, ˆ X, ˆ Y,Z EPZEP (X,Y Z)EP ( ˆ X, ˆ Y Z)[c X(X, ˆX) + c Y (Y, ˆY)] = inf P PX,Y, ˆ X, ˆ Y,Z EPZEP (X,Y Z)[c X(X,F(Ga(Za),Gy(Zy))) + c Y (Y,D(Gy(Zy)))] = inf P PX,Y,Z EPZEP (X,Y Z)[c X(X,F(Ga(Za),Gy(Zy))) + c Y (Y,D(Gy(Zy)))] = inf P PX,Y,Z EX,Y,Z P [c X(X,F(Ga(Za),Gy(Zy))) + c Y (Y,D(Gy(Zy)))]. Note that in Eq. equation 8, PX,Y,Z = P((X,Y) PX,Y,Z PZ) and with a proposed Q(Z X), we can rewrite Eq. equation 8 as inf P PX,Y,Z EPX,YEPZ[c X(X,F(Ga(Za),Gy(Zy))) + c Y (Y,D(Gy(Zy)))] = inf QZ=PZ EPX,YEQ(Z X) c X(X,F(Ga(Za),Gy(Zy))) + c Y (Y,D(Gy(Zy))) . (9) A.2 FROM UNIMODAL TO MULTIMODAL The proof is similar to Proposition 2, and we present a sketch to it. We can first show P(ˆX1 M, ˆY Z = z) are Dirac for all z Z. Then we use the fact that c((X1 M,Y),(ˆX1 M, ˆY)) = M i=1 c Xi(Xi, ˆXi) + c Y (Y, ˆY). Finally, we follow the tower rule of expectation and the conditional independence property similar to the proof in Proposition 2 and this concludes the proof. B FULL BASELINE MODELS & RESULTS For a detailed description of the baselines, we point the reader to MFN (Zadeh et al., 2018a), MARN (Zadeh et al., 2018b), TFN (Zadeh et al., 2017), BC-LSTM (Poria et al., 2017), MVLSTM (Rajagopalan et al., 2016), EF-LSTM (Hochreiter & Schmidhuber, 1997; Graves et al., 2013; Schuster & Paliwal, 1997), DF (Nojavanasghari et al., 2016), MV-HCRF (Song et al., 2012; 2013), EF-HCRF (Quattoni et al., 2007; Morency et al., 2007), THMM (Morency et al., 2011), SVM-MD (Zadeh et al., 2016) and RF (Breiman, 2001). We use the following extra notations for full descriptions of the baseline models described in Section 3.2, paragraph 3: Variants of EF-LSTM: EF-LSTM (Early Fusion LSTM) uses a single LSTM (Hochreiter & Schmidhuber, 1997) on concatenated multimodal inputs. We also implement the EF-SLSTM (stacked) (Graves et al., 2013), EF-BLSTM (bidirectional) (Schuster & Paliwal, 1997) and EFSBLSTM (stacked bidirectional) versions. Variants of EF-HCRF: EF-HCRF: (Hidden Conditional Random Field) (Quattoni et al., 2007) uses a HCRF to learn a set of latent variables conditioned on the concatenated input at each time step. Published as a conference paper at ICLR 2019 EF-LDHCRF (Latent Discriminative HCRFs) (Morency et al., 2007) are a class of models that learn hidden states in a CRF using a latent code between observed concatenated input and hidden output. EF-HSSHCRF: (Hierarchical Sequence Summarization HCRF) (Song et al., 2013) is a layered model that uses HCRFs with latent variables to learn hidden spatio-temporal dynamics. Variants of MV-HCRF: MV-HCRF: Multi-view HCRF (Song et al., 2012) is an extension of the HCRF for Multi-view data, explicitly capturing view-shared and view specific sub-structures. MVLDHCRF: (Morency et al., 2007) is a variation of the MV-HCRF model that uses LDHCRF instead of HCRF. MV-HSSHCRF: (Song et al., 2013) further extends EF-HSSHCRF by performing Multiview hierarchical sequence summary representation. In the following, we provide the full results for all baselines models described in Section 3.2, paragraph 3. Table 3 contains results for multimodal speaker traits recognition on the POM dataset. Table 4 contains results for the multimodal sentiment analysis on the CMU-MOSI, ICT-MMMO, You Tube, and MOUD datasets. Table 5 contains results for multimodal emotion recognition on the IEMOCAP dataset. MFM consistently achieves state-of-the-art or competitive results for all six multimodal datasets. We believe that by our MFM design, the multimodal discriminative factor Fy has successfully learned more meaningful representations by distilling discriminative features. This highlights the benefit of learning factorized multimodal representations towards discriminative tasks. C MULTIMODAL FEATURES For each of the multimodal time series datasets as mentioned in Section 3.2, paragraph 3, we extracted the following multimodal features: Language: We use pre-trained word embeddings (glove.840B.300d) (Pennington et al., 2014) to convert the video transcripts into a sequence of 300 dimensional word vectors. Visual: We use Facet (i Motions, 2017) to extract a set of features including per-frame basic and advanced emotions and facial action units as indicators of facial muscle movement (Ekman et al., 1980; Ekman, 1992). Acoustic: We use COVAREP (Degottex et al., 2014) to extract low level acoustic features including 12 Mel-frequency cepstral coefficients (MFCCs), pitch tracking and voiced/unvoiced segmenting features, glottal source parameters, peak slope parameters and maxima dispersion quotients. To reach the same time alignment between different modalities we choose the granularity of the input to be at the level of words. The words are aligned with audio using P2FA (Yuan & Liberman, 2008) to get their exact utterance times. We use expected feature values across the entire word for visual and acoustic features since they are extracted at a higher frequencies. We make a note that the features for some of these datasets are constantly being updated. The authors of Zadeh et al. (2018a) notified us of a discrepancy in the sampling rate for acoustic feature extraction in the ICT-MMMO, You Tube and MOUD datasets which led to inaccurate word-level alignment between the three modalities. They publicly released the updated multimodal features. We performed all experiments on the latest versions of these datasets which can be accessed from https: //github.com/A2Zadeh/CMU-Multimodal SDK. All baseline models were retrained with extensive hyperparameter search for fair comparison. D INFORMATION AND GRADIENT-BASED INTERPRETATION Information-Based Interpretation: We choose the normalized Hilbert-Schmidt Independence Criterion (Gretton et al., 2005; Wu et al., 2018) as the approximation (see Sugiyama & Yamada (2012); Wu et al. (2018)) of our MI measure: MI(F , ˆXi) = HSICnorm(F , ˆXi) = tr(KF HK ˆ Xi H) HKF H F HK ˆ Xi H F , (10) where represents y or ai, n is the number of {F , ˆXi} pairs, H = I 1 n11 , KF Rn n is the Gram matrix of F with KF ij = k1(F i,F j), K ˆ Xi Rn n is the Gram matrix of ˆXi with K ˆ Xijk = k2(ˆXij, ˆXik). k1( , ) and k2( , ) are predefined kernel functions. The most common choice for the kernel is the RBF kernel. However, if we consider time series data with various time steps, we need to either perform data augmentation or choose another kernel choice. For example, we can adopt the Global Alignment Kernel (Cuturi et al., 2007) which considers the Published as a conference paper at ICLR 2019 Table 3: Results for personality trait recognition on the POM dataset. The best results are highlighted in bold and SOT A shows the change in performance over previous state of the art. Improvements are highlighted in green. MFM achieves state-of-the-art or competitive performance on all datasets and metrics. Dataset POM Speaker Personality Traits Task Con Pas Voi Dom Cre Viv Exp Ent Res Tru Rel Out Tho Ner Per Hum Metric r Majority -0.041 -0.029 -0.104 -0.031 -0.122 -0.044 -0.065 -0.105 0.006 -0.077 -0.024 -0.085 -0.130 0.097 -0.127 -0.069 SVM 0.063 0.086 -0.004 0.141 0.113 0.076 0.134 0.141 0.166 0.168 0.104 0.066 0.134 0.068 0.064 0.147 DF 0.240 0.273 0.017 0.139 0.112 0.173 0.118 0.217 0.148 0.143 0.019 0.093 0.041 0.136 0.168 0.259 EF-LSTM 0.200 0.302 0.031 0.079 0.170 0.244 0.265 0.240 0.142 0.062 0.083 0.152 0.260 0.105 0.217 0.227 EF-SLSTM 0.221 0.327 0.042 0.151 0.177 0.239 0.268 0.248 0.204 0.069 0.092 0.215 0.252 0.159 0.218 0.196 EF-BLSTM 0.162 0.289 -0.034 0.135 0.191 0.279 0.274 0.231 0.184 0.154 0.093 0.147 0.245 0.166 0.243 0.272 EF-SBLSTM 0.174 0.310 0.021 0.088 0.170 0.224 0.261 0.241 0.155 0.163 0.097 0.120 0.215 0.121 0.216 0.171 MV-LSTM 0.358 0.416 0.131 0.146 0.280 0.347 0.323 0.326 0.295 0.237 0.119 0.238 0.284 0.258 0.239 0.317 BC-LSTM 0.359 0.425 0.081 0.234 0.358 0.417 0.450 0.361 0.293 0.109 0.075 0.078 0.363 0.184 0.344 0.319 TFN 0.089 0.201 0.030 0.020 0.124 0.204 0.171 0.223 -0.051 -0.064 0.114 0.060 0.048 -0.002 0.106 0.213 MARN 0.340 0.410 0.166 0.235 0.340 0.374 0.406 0.378 0.282 0.147 0.215 0.204 0.348 0.235 0.303 0.287 MFN 0.395 0.428 0.193 0.313 0.367 0.431 0.452 0.395 0.333 0.296 0.255 0.259 0.381 0.318 0.377 0.386 MFM 0.431 0.450 0.197 0.411 0.380 0.448 0.467 0.452 0.368 0.212 0.309 0.333 0.404 0.333 0.334 0.408 SOT A 0.036 0.022 0.004 0.097 0.013 0.017 0.015 0.057 0.035 0.054 0.074 0.023 0.015 0.022 Table 4: Sentiment prediction results on CMU-MOSI, ICT-MMMO, You Tube and MOUD. The best results are highlighted in bold and SOT A shows the change in performance over previous state of the art (SOTA). Improvements are highlighted in green. MFM achieves state-of-the-art or competitive performance on all datasets and metrics. Dataset CMU-MOSI ICT-MMMO You Tube MOUD Task Sentiment Sentiment Sentiment Sentiment Metric Acc_7 Acc_2 F1 MAE r Acc_2 F1 Acc_3 F1 Acc_2 F1 Majority 17.5 50.2 50.1 1.864 0.057 40.0 22.9 42.4 25.2 60.4 45.5 RF 21.3 56.4 56.3 - - 70.0 69.8 33.3 32.3 64.2 63.3 SVM-MD 26.5 71.6 72.3 1.100 0.559 68.8 68.7 42.4 37.9 59.4 45.5 THMM 17.8 53.8 53.0 - 50.7 45.4 42.4 27.9 61.3 57.0 SAL-CNN - 73.0 - - - - - - - - - C-MKL 30.2 72.3 72.0 - - - - - - - - EF-HCRF 24.6 65.3 65.4 - - 50.0 50.3 44.1 43.8 54.7 54.7 EF-LDHCRF 24.6 64.0 64.0 - - 73.8 73.1 45.8 45.0 52.8 49.3 MV-HCRF 22.6 44.8 27.7 - - 36.3 19.3 27.1 19.7 60.4 45.5 MV-LDHCRF 24.6 64.0 64.0 - - 68.8 67.1 44.1 44.0 53.8 46.9 CMV-HCRF 22.3 44.8 27.7 - - 36.3 19.3 30.5 14.3 60.4 45.5 CMV-LDHCRF 24.6 63.6 63.6 - - 51.3 51.4 42.4 42.0 53.8 47.8 EF-HSSHCRF 24.6 63.3 63.4 - - 50.0 51.3 37.3 35.6 52.8 49.3 MV-HSSHCRF 24.6 65.6 65.7 - - 62.5 63.1 44.1 44.0 47.2 46.4 DF 26.8 72.3 72.1 1.143 0.518 65.0 58.7 45.8 32.0 67.0 67.1 EF-LSTM 32.4 74.3 74.3 1.023 0.622 66.3 65.0 44.1 43.6 67.0 64.3 EF-SLSTM 29.3 72.7 72.8 1.081 0.600 72.5 70.9 40.7 41.2 56.6 51.4 EF-BLSTM 28.9 72.0 72.0 1.080 0.577 63.8 49.6 42.4 38.1 58.5 58.9 EF-SBLSTM 26.8 73.3 73.2 1.037 0.619 62.5 49.0 37.3 33.2 63.2 63.3 MV-LSTM 33.2 73.9 74.0 1.019 0.601 72.5 72.3 45.8 43.3 57.6 48.2 BC-LSTM 28.7 73.9 73.9 1.079 0.581 70.0 70.1 45.0 45.1 72.6 72.9 TFN 28.7 74.6 74.5 1.040 0.587 72.5 72.6 45.0 41.0 63.2 61.7 MARN 34.7 77.1 77.0 0.968 0.625 71.3 70.2 48.3 44.9 81.1 81.2 MFN 34.1 77.4 77.3 0.965 0.632 73.8 73.1 51.7 51.6 81.1 80.4 MFM 36.2 78.1 78.1 0.951 0.662 81.3 79.2 53.3 52.4 82.1 81.7 SOT A 1.5 0.7 0.8 0.014 0.030 7.5 6.1 1.6 0.8 1.0 0.5 Published as a conference paper at ICLR 2019 Table 5: Emotion recognition results on IEMOCAP test set. The best results are highlighted in bold and SOT A shows the change in performance over previous SOTA. Improvements are highlighted in green. MFM achieves state-of-the-art or competitive performance on all datasets and metrics. Dataset IEMOCAP Emotions Task Happy Sad Angry Frustrated Excited Neutral Metric Acc_2 F1 Acc_2 F1 Acc_2 F1 Acc_2 F1 Acc_2 F1 Acc_2 F1 Majority 85.6 79.0 79.4 70.3 75.8 65.4 79.5 70.4 89.6 84.7 59.1 44.0 SVM 86.1 81.5 81.1 78.8 82.5 82.4 77.3 71.1 86.4 86.0 65.2 64.9 RF 85.5 80.7 80.1 76.5 81.9 82.0 78.6 75.3 88.9 85.1 63.2 57.3 THMM 85.6 79.2 79.5 79.8 79.3 73.0 71.6 69.6 86.0 84.6 58.6 46.4 EF-HCRF 85.7 79.2 79.4 70.3 75.8 65.4 79.5 70.4 89.6 84.7 59.1 44.0 EF-LDHCRF 85.8 79.5 79.4 70.3 75.8 65.4 79.5 70.4 89.6 84.7 59.1 44.0 MV-HCRF 15.0 4.9 79.4 70.3 24.2 9.4 79.5 70.4 89.6 84.7 59.1 44.0 MV-LDHCRF 85.7 79.2 79.4 70.3 75.8 65.4 79.5 70.4 89.6 84.7 59.1 44.0 CMV-HCRF 14.4 3.6 79.4 70.3 24.2 9.4 79.5 70.4 89.6 84.7 59.1 44.0 CMV-LDHCRF 85.8 79.5 79.4 70.3 75.8 65.4 79.5 70.4 89.6 84.7 59.1 44.0 EF-HSSHCRF 85.8 79.5 79.4 70.3 75.8 65.4 79.5 70.4 89.6 84.7 59.1 44.0 MV-HSSHCRF 85.8 79.5 79.4 70.3 75.8 65.4 79.5 70.4 89.6 84.7 59.1 44.0 DF 86.0 81.0 81.8 81.2 75.8 65.4 78.4 76.8 89.6 84.7 59.1 44.0 EF-LSTM 85.2 83.3 82.1 81.1 84.5 84.3 79.5 70.4 89.6 84.7 68.2 67.1 EF-SLSTM 85.6 79.0 80.7 80.2 82.8 82.2 77.5 69.7 89.3 86.2 68.8 68.5 EF-BLSTM 85.0 83.7 81.8 81.6 84.2 83.3 79.5 70.4 89.6 84.7 67.1 66.6 EF-SBLSTM 86.0 84.2 80.2 80.5 85.2 84.5 79.5 70.4 89.6 84.7 67.8 67.1 MV-LSTM 85.9 81.3 80.4 74.0 85.1 84.3 79.5 73.8 88.9 85.8 67.0 66.7 BC-LSTM 84.9 81.7 83.2 81.7 83.5 84.2 80.0 76.1 86.9 85.4 67.5 64.1 TFN 84.8 83.6 83.4 82.8 83.4 84.2 74.1 74.3 75.6 78.0 67.5 65.4 MARN 86.7 83.6 82.0 81.2 84.6 84.2 79.5 76.6 89.6 87.1 66.8 65.9 MFN 90.1 85.3 85.8 79.2 87.0 86.0 80.3 76.9 89.8 86.3 71.8 61.7 MFM 90.2 85.8 88.4 86.1 87.5 86.7 80.4 74.5 90.0 87.1 72.1 68.1 SOT A 0.1 0.5 2.6 3.3 0.5 0.7 0.1 0.2 0.3 alignment between two varying-length time series when computing the kernel score between them. To simplify our analysis, we choose to augment data before we calculate the kernel score with the RBF kernel. More specifically, we perform averaging over time series data: T t=1 Xt with X = [X1,X2, ,XT ]. (11) The bandwidth of the RBF kernel is set as 1.0 throughout the experiments. Table 6: Information-Based interpretation results showing ratios ri = MI(Fy, ˆXi) MI(Fai, ˆXi), i {(ℓ)anguage,(v)isual,(a)coustic} for the POM dataset for personality traits prediction. Ratio rℓ(language) rv (visual) ra (acoustic) POM 1.090 0.996 0.898 Here, we provide an additional interpretation result for the POM dataset in Table 6. We observe that the language modality is also the most informative while the visual and acoustic modalities are almost equally informative. This result is in agreement with behavioral studies which have observed that non-verbal behaviors are particularly informative of personality traits (Guimond & Massrieh, 2012; Levine et al., 2009; Mohammadi et al., 2010). For example, the same sentence this movie was great can convey significantly different messages on speaker confidence depending on whether it was said in a loud and exciting voice, with eye contact, or powerful gesticulation. Gradient-Based Interpretation: MFM reconstructs xi as follows: ˆxi = Fi(fai, fy), fai = Gai(zai), fy = Gy(zy), zai Q(Zai Xi = xi), zy Q(Zy X1 M = x1 M). (12) Equation equation 12 also explains how we obtain fy P(Fy X1 M = x1 M). The gradient flow through time is defined as: fy(ˆxi) =[ fy ˆx1 i 2 F , fy ˆx2 i 2 F , , fy ˆx T i 2 F ]. (13) Published as a conference paper at ICLR 2019 E ENCODER AND DECODER DESIGN FOR MULTIMODAL SYNTHETIC IMAGE DATASET For experiments on the multimodal synthetic image dataset, we use convolutional+fully-connected layers for the encoder and deconvolutional+fully-connected layers for the decoder (Zeiler et al., 2010). Different convolutional layers are each applied on the input SVHN and MNIST images to learn modality-specific generative factors. Next, we concatenate the features from two more convolutional layers on SVHN and MNIST to learn the multimodal-discriminative factor. The multimodal discriminative factor is passed through fully-connected layers to predict the label. For generation, we concatenate the multimodal discriminative factors and the modality-specific generative factor together and use a deconvolutional layer to generate digits. F ENCODER AND DECODER DESIGN FOR MULTIMODAL TIME SERIES DATASETS Figure 5 illustrates how MFM operates on multimodal time series data. The encoder Q(Zy X1 M) can be parametrized by any model that performs multimodal fusion (Nojavanasghari et al., 2016; Zadeh et al., 2018a). We choose the Memory Fusion Network (MFN) (Zadeh et al., 2018a) as our encoder Q(Zy X1 M). We use encoder LSTM networks and decoder LSTM networks (Cho et al., 2014) to parametrize functions Q(Za1 M X1 M) and F1 M respectively, and FCNNs to parametrize functions Gy, Ga{1 M} and D. Figure 5: Recurrent neural architecture for MFM. The encoder Q(Zy X1 M) can be parametrized by any model that performs multimodal fusion (Nojavanasghari et al., 2016; Zadeh et al., 2018a). We use encoder LSTM networks and decoder LSTM networks (Cho et al., 2014) to parametrize functions Q(Za1 M X1 M) and F1 M respectively, and FCNNs to parametrize functions Gy, Ga{1 M} and D. G SURROGATE INFERENCE GRAPHICAL MODEL We illustrate the surrogate inference for addressing the missing modalities issue in Figure 6. The surrogate inference model infers the latent codes given the present modalities. These inferred latent codes can then be used for reconstructing the missing modalities or label prediction in the presence of missing modalities. Published as a conference paper at ICLR 2019 Figure 6: The surrogate inference graphical model to deal with missing modalities in MFM. Red lines denote original inference in MFM and green lines denote surrogate inference to infer latent codes given present modalities. H COMPARISON WITH HSU & GLASS (2018) A similar approach for factorizing the latent factors was recently proposed by Hsu & Glass (2018) in work that was performed independently and in parallel. In comparison with MFM, there are several major differences that can be categorized into the (1) prior matching discrepancy, (2) inference network, (3) discriminative objective, (4) multimodal fusion, (5) scale of experiments. 1. MFM uses MMD(QZ,PZ) (see Equation 4) as the prior matching discrepancy while Hsu & Glass (2018) use KL(QZ X,PZ)) (see Section 2.2.1 in Hsu & Glass (2018)). 2. MFM considers multimodal and unimodal inference in a single network (see Figure 1(b)), while Hsu & Glass (2018) considers separate networks (see Figure 1 in Hsu & Glass (2018)). They further propose to match the coherence between these two networks using an additional loss term (see Equation 7 in Hsu & Glass (2018)). 3. MFM learns to predict the labels using a generative framework (see Figure 1(a)), while Hsu & Glass (2018) use an additional hinge loss to separate the latent factors from different labels (see Equation 9 in Hsu & Glass (2018)). 4. MFM is a flexible framework that can be combined with any multimodal fusion encoder (see Section 2.4), while Hsu & Glass (2018) considers a fixed multimodal encoder (similar to early fusion) (see Section 4.1 in Hsu & Glass (2018)). 5. We evaluate the performance of MFM over a much larger scale of datasets. We perform experiments on six multimodal time-series datasets that take on the form of videos with the language, visual, and acoustic modalities. These datasets span three core research areas of multimodal personality traits recognition, multimodal sentiment analysis, and multimodal emotion recognition. On the other hand, Hsu & Glass (2018) evaluates their model on a spoken digit dataset which randomly combines a digit image with a spoken digit (see Section 4 in Hsu & Glass (2018)). MFM further considers experiments to evaluate reconstruction and prediction in the presence of missing modalities (see Section 3.2) which Hsu & Glass (2018) do not. Lastly, we compares to over 20 baseline models in our experiments (see Section 3.2) and explore the choice of various multimodal encoders in MFM. Hsu & Glass (2018) only compares to the JMVAE baseline model (Suzuki et al., 2016a) which resembles the MD model in our ablation study (see Section 4.4 in Hsu & Glass (2018)). In Table 7, we provide a comparison on the CMU-MOSI, ICT-MMMO, You Tube and MOUD datasets to test the disentanglement and prediction performance for the model described in Hsu & Glass (2018). These experimental results show that across these datasets and metrics, MFM performs better than the model proposed in Hsu & Glass (2018). We would like to highlight that at the time of submission, the code for (Hsu & Glass, 2018) had not been made public and we reimplemented their model to experiment on our datasets. I COMPARISON WITH β-VAE Although β-VAE (Higgins et al., 2017) was designed to handle unimodal data, we provide an extension to multimodal data. To achieve this, we set the choice of prior matching discrepancy as the KL-divergence KL(QZ X,PZ) and set β large (i.e. β {10,50,100,200}) to encourage disentanglement of latent variables. We train a β-VAE to model multimodal data using the same Published as a conference paper at ICLR 2019 Table 7: Comparison with Hsu & Glass (2018) for sentiment analysis on CMU-MOSI, ICT-MMMO, You Tube, and MOUD. MFM outperforms the baselines across these datasets and metrics. Dataset CMU-MOSI ICT-MMMO You Tube MOUD Task Sentiment Sentiment Sentiment Sentiment Metric Acc_7 Acc_2 F1 MAE r Acc_2 F1 Acc_3 F1 Acc_2 F1 without factorization (EF) 32.4 74.3 74.3 1.023 0.622 72.5 70.9 44.1 43.6 67.0 64.3 without factorization (MFN) 34.1 77.4 77.3 0.965 0.632 73.8 73.1 51.7 51.6 81.1 80.4 Hsu & Glass (2018) 33.8 75.2 75.2 1.049 0.584 77.5 75.0 51.7 48.6 66.0 62.9 MFM 36.2 78.1 78.1 0.951 0.662 81.3 79.2 53.3 52.4 82.1 81.7 Table 8: Comparison with β-VAE for multimodal sentiment analysis on CMU-MOSI, ICT-MMMO, You Tube, and MOUD. MFM outperforms β-VAE across these datasets and metrics. Dataset CMU-MOSI ICT-MMMO You Tube MOUD Task Sentiment Sentiment Sentiment Sentiment Metric Acc_7 Acc_2 F1 MAE r Acc_2 F1 Acc_3 F1 Acc_2 F1 β-VAE 29.7 71.3 71.3 1.094 0.552 65.0 59.8 46.7 31.3 60.4 54.2 MFM 36.2 78.1 78.1 0.951 0.662 81.3 79.2 53.3 52.4 82.1 81.7 factorization as proposed in our model (i.e. modality-specific generative factors Za{1 M} and a multimodal discriminative factor Zy). To provide a fair comparison to our discriminative model, we fine tune by training a classifier on top of the multimodal discriminative factor Zy to the label Y. We provide experimental results in Table 8 on the CMU-MOSI, ICT-MMMO, You Tube and MOUD datasets. MFM outperforms β-VAE across these datasets and metrics.