# generalizable_multilinear_attention_network__46cc07e9.pdf Generalizable Multi-Linear Attention Network Tao Jin Zhejiang University jint_zju@zju.edu.cn Zhou Zhao Zhejiang University zhaozhou@zju.edu.cn The majority of existing multimodal sequential learning methods focus on how to obtain powerful individual representations and neglect to effectively capture the multimodal joint representation. Bilinear attention network (BAN) is a commonly used integration method, which leverages tensor operations to associate the features of different modalities. However, BAN has a poor compatibility for more modalities, since the computational complexity of the attention map increases exponentially with the number of modalities. Based on this concern, we propose a new method called generalizable multi-linear attention network (MAN), which can associate more modalities in acceptable complexity with hierarchical approximation decomposition. Specifically, considering the fact that softmax attention kernels cannot be decomposed as linear operation directly, we adopt the addition random features mechanism to approximate the non-linear softmax functions with enough theoretical analysis. Furthermore, we also introduce the local sequential constraints, which can be combined with ARF conveniently, as positional information. We conduct extensive experiments on several datasets of corresponding tasks, the experimental results show that MAN could achieve competitive results compared with baseline methods, showcasing the effectiveness of our contributions. 1 Introduction Multimodal learning draws increasing attention in recent years, which aims to process and understand the information from multiple modalities (i.e. vision, language, audio) with machine learning skills. Many endeavors have been devoted to the multimodal interaction and robust individual representation learning [29, 6, 39, 40]. Existing multimodal interaction methods could be categorized into Transformer-based [29, 6] and Non-Transformer-based methods [39, 40]. With deep stacked attention blocks [30] and suitable number of training samples, Transformer-based methods could achieve relatively good performances with relatively high computational complexity. After obtaining the individual representations of all the modalities, joint representation learning [5, 15, 14, 19, 36] is performed by leveraging simple matrix and vector operations. Bilinear functions have been widely utilized in this field. Specifically, bilinear pooling [15] employs low-rank decomposition to approximate high-order tensor operations for input vectors. Bilinear attention network (BAN) [14] uses a bilinear attention distribution, on top of low-rank bilinear pooling. In general, BAN exploits bilinear interactions between two groups of input channels and bilinear pooling extracts the joint representations for each pair of channels. Besides, bilinear pooling could be extended to multi-linear pooling with acceptable computational complexity. However, BAN has the limitation to generalize to more modalities due to the complexity. For example, when we want to integrate the features with the same space Rd T of m modalities, we would obtain a high-order tensor A RT m that captures multimodal interactions. It is observed that A scales exponentially with the number of modalities. Motivated by the observations, in this paper, we propose a novel method called generalizable multilinear attention network (MAN), which could be leveraged to integrate as many modalities as possible Corresponding Author 35th Conference on Neural Information Processing Systems (Neur IPS 2021). with linear complexity. Concretely, we employ multi-linear pooling to extract the joint representations for multiple feature channels of different modalities. With the previous assumption, we could obtain T m joint representations for m modalities in total. Following the main idea of BAN, an interaction tensor A RT m represents the attention weights of the joint representations (in practise, we devise the combinational addition attention map). To reduce the computational complexity and space requirements, we devise a hierarchical approximation decomposition (HAD) mechanism for the high-order tensor A, leading to the efficient approximation decomposition. Specifically, considering the fact that softmax attention kernels cannot be decomposed as linear operation directly, we devise the addition random features (ARF) mechanism to approximate the non-linear softmax function. Furthermore, we facilitate the sparse attention distributions by leveraging local sequential constraints (LSC), which can be combined with ARF conveniently. We conduct extensive experiments on three tasks, the experimental results show that MAN could achieve competitive results compared with the state-of-the-art methods. To sum up, the contributions of our work are four-fold: We propose the multi-linear attention network (MAN), which is the extension of bilinear attention network (BAN), to learn and utilize multi-linear attention distributions on top of the multi-linear pooling operations. To reduce the computational complexity and space requirements, we devise a hierarchical approximation decomposition mechanism for the combinational addition interaction tensor, leading to the efficient approximation decomposition. Most importantly, considering that softmax attention kernels cannot be decomposed as linear operation directly, we adopt the addition random features (ARF) approach to approximate the non-linear softmax functions with enough theoretical analysis. To process multimodal sequential features, we introduce the local sequential constraints (LSC), which can be combined with ARF conveniently. To be compatible with Transformer, the multi-head and residual properties are introduced into MAN. We conduct extensive experiments on three different tasks, the experimental results show that MAN could achieve competitive results compared with the baseline methods. In this section, we first introduce the low-rank multi-linear pooling, which is the extension of bilinear pooling [15]. Then we illustrate the details of multi-linear attention network. 2.1 Multi-Linear Pooling We first review the low-rank bilinear pooling. [15] proposes to reduce the rank of bilinear weight matrix Wi Rd1 d2 with CP decomposition [7]. Wi could be rewrited as the multiplication of two smaller matrices Ui VT i , where Ui Rd1 d and Vi Rd2 d, d denotes the value of rank with the constraint d min(d1, d2). We calculate the scalar output fi as follows: fi = x T Wiy x T Ui VT i y = 1T UT i x VT i y (1) where 1 Rd consists of ones and denotes element-wise multiplication, note that we omit the bias term for convenience. When the output is a vector f Rc, the 1 is replaced by a pooling matrix: f = PT UT x VT y (2) where P Rd c, U Rd1 d, and V Rd2 d. By leveraging the pooling matrix P, the number of parameters is reduced significantly. Based on the bilinear pooling, we introduce the calculation process of multi-linear pooling. Suppose the number of modalities is m and the feature (channel) dimensions of them are {dj|j [1, m]}, Eq. 2 is extended to: j=1 UT j vj) = PT (UT 1 v1 UT 2 v2 ... UT mvm) (3) Step 1. Multi-linear Attention Map Step 2. Multi-linear Attention Networks Figure 1: The core module of multi-linear attention network (only one head), where we provide detailed input and output of each step, making it easy to understand. Note that, for convinience, we give an example of three modalities. where vj Rdj and Uj Rdj d denote the feature vector of j-th modality and corresponding projection matrix. Theoretically, P and {Uj|j [1, m]} are the components of d-rank approximation of the high-order tensor. 2.2 Multi-Linear Attention Network Since the features of most modalities are multi-channel (for example, video features have temporal dimension), we propose MAN to learn and utilize the multi-linear attention distributions on top of the multi-linear pooling operations. Suppose the feature matrices of m modalities are {Vj Rdj Tj|j [1, m]}, we could obtain Qm j=1 Tj groups of joint representations. To combine these information, we introduce the multi-linear attention map A R Qm j=1 Tj as follows: f k = ((A 1 (VT 1 U 1)k) 2 (VT 2 U 2)k) 3 ... m (VT m U m)k (4) where U j Rdj K, (VT j U j)k RTj, the operator j denotes the j-th mode product between a tensor and a vector, and f k denotes the k-th element of intermediate representation. Note that Eq. 4 is a multi-linear model for m groups of input channels, where A is a multi-linear weight matrix. To make it easier to understand, Eq. 4 could be rewritten as follows: tm=1 At1,t2,...,tm( j=1 VT j,tj U j,k) (5) where Vj,tj Rdj denotes the tj-th channel of input Vj and U j,k Rdj denotes the k-th column of Uj, At1,t2,...,tm denotes an element in tensor A with indices [t1, t2, ..., tm]. The intermediate result f RK consists of K elements [f 1, ..., f k, ..., f K]. Then the multi-linear joint representation is f = PT f where f Rc and P RK c. For convenience, we define the multi-linear attention network as a function of m multi-channel inputs parameterized by a multi-linear attention map as follows: f = MAN({Vj|j [1, m]}; A) (6) Combinational Additive Attention Map. We argue that multi-linear attention map should have several properties: (1) The attention logits are calculated by leveraging the interaction information between different modalities. Each modality is in a symmetric (equal) position. (2) Following the main idea of BAN, the attention map can be calculated with Hadamard product and tensor operations similarly. However, since the absolute values of most multiplicative factors are between 0 and 1, directly employing softmax function may result in even attention distributions when the attention map is large. (3) As for the processing of multimodal sequential features, the time interval (relative position) of different modalities should be considered. In general, locally close moments are more relevant. Following the idea of Transformer, most positional encoding methods can be integrated with the inner product of two vectors. Therefore, we utilize simply combinational addition operation (CAO) to replace the Hadamard product. The attention map A is defined as follows: A := softmax(A), At1,t2,...,tm = ( q=1, =j (VT q,tq U q )(U T j Vj,tj)) (7) where U j Rdj K , A R Qm j=1 Tj and the softmax function is applied element-wisely along all the dimensions. In Eq. 7, the number of additive items (VT q,tq U q )(U T j Vj,tj) is positively correlated with the number of modalities, resulting in the bigger difference between elements and sparser attention distributions. Besides, the combinational addition operation has satisfactory compatibility with local sequential constraints introduced in the subsequent section. Local Sequential Constraints (LSC) for Multimodal Temporal Features. When the number of modalities increases, the attention tensor A becomes larger and the probability distributions are much denser. As mentioned in the previous section, the features of temporal proximity are more relevant. So we devise an attenuation function related to the time steps of associated modalities. Suppose the number of time steps is T, we assign each time step a T s -dimensional vector ( T s denotes the number of chunks and s adjusts the scale), which consists of e and e. For example, the temporal vectors for the 1-st and 2-nd temporal chunks are [e, e, e, ..., e] and [e, e, e, ..., e], the inner product result is ( T s 1) e2. Similarly, the inner product result between any t1 and t2 time steps are ( T s |t2 t1|) e2, which fits the property of attenuation function. In practice, we simply rewrite the Eq. 7, concatenating the temporal vectors to the features of different modalities: At1,t2,...,tm = ( h VT q,tq U q , Etq ih VT j,tj U j , Etj i T ) (8) To balance the weights of modality association and local constraints, we adjust the value of e. Hierarchical Approximation Decomposition. Considering the fact that the size of combinational additive interaction tensor A scales exponentially to the number of modalities. We devise a hierarchical decomposition mechanism to cope with this problem. Without restrictions, A can be decomposed with CP algorithm: j=1 Bj, At1,t2,...,tm = 1T ( j=1 Bj,tj) (9) where 1 RH, Bj RH Tj and H denotes the rank value, is the defined outer product operation for multiple matrices, which is a bit different from traditional vectorial operation. The right equation is the detailed calculation for each element, where Bj,tj RH denotes the tj-th channel of Bj. We directly substitude the Eq. 8 into Eq. 5 as follows2: j=1 Bj,tj))( j=1 VT j,tj U j) = 1T ( j=1 Bj VT j U j) (10) where we obtain a simple conclusion. Bj RH Tj and U j Rdj K can be treated as two linear projection weights for different dimensions, respectively. Addition Random Features. Although Eq. 9 is quite simple and pretty, the high-order tensor A cannot be decomposed directly due to the non-linearity caused by softmax function. Thus, it is necessary to find a substitute approximate solution. Inspired by the Performer [3], we propose the addition random features mechanism, which can be treated as an extension of vanilla random features mechanism for more than two modalities. The detailed process is shown as follows1: 2The detailed proof and complete algorithm process are shown in the appendix. SM({vj|j [1, m]}) = Ew N(0,IK ) j=1 exp w vj vj 2 where vj denotes the feature vector of j-th modality, SM({vj|j [1, m]}) is equal to exp(Pm j=1 Pm k=1, =j v T k vj), E and N (0, IK ) denote expectation and sampling distribution. The detailed derivations can be found in the supplementary materials. Eq. 10 is employed to calculate At1,t2,...,tm directly. Bj in Eq. 8 can be obtained by project the features of j-th modality directly with ARF mechanism. In this way, the number of random features is equal to the rank value H of A. Note that softmax function consists of exponential operation and sum normalization, we need the value of normalized denominator. In practice, we directly employ {Bj|j = [1, m]} to calculate it1: j=1 Bj 1) (12) where P denotes the sum of all elements in the tensor and 1 RTj. We could keep P A and utilize the integration result f to divide it. Note that, during the calculation of ARF for softmax function, a reduction factor 1 H is introduced into both numerator and denominator, thus it directly cancel out. We could treat the expectation operation in Eq. 10 as the summation operation. The core module of MAN is shown in Fig. 1. Multi-head Multi-linear Attention. Following the main idea of Transformer [30] that projecting the features into multiple representation spaces could capture more information, we also adopt the multi-head strategy in the multi-linear attention. In practice, we add multiple linear mappings to the input features before fusing them (i.e. multiple U j and U j for {VT j U j, VT j U j |j [1, m]} ). The results of all the heads are concatenated. The calculation follows the process described above. Stacked Multi-linear Attention. The residual connection is proved to be effective in most structures [8]. Therefore, such strategy can also be introduced into multi-linear attention network. Specifically, Eq. 6 is modified as: f l n+1 = MAN(f l n, {Vj|j [1, m], j = l}; A) 1T + f l n (13) where 1 RTl, f l n Rdl Tl denotes the n-th output of l-th modality and f l 0 = Vl (if d1 g =K=c, g denotes the number of heads). Complexity. Without approximation decomposition, the computational complexity increases exponentially with respect to the number of modalities. Concretely, we need Qm j=1 Tj low-rank multimodal pooling operations, leading to the unfriendly complexity O(Qm j=1 Tj). While the complexity scales linearly O(Pm j=1 Tj) with the number of modalities by leveraging approximation decomposition. 3 Experiments 3.1 Datasets We evaluate MAN on three challenging tasks, multimodal sentiment analysis, multimodal speaker traits recognition, and multimodal video retrieval. In this section, we provide a brief introduction of the datasets (CMU-MOSI [41] for multimodal sentiment analysis, POM [25] for multimodal speaker traits recognition, MSR-VTT [33] and LSMDC [28] for multimodal video retrieval). CMU-MOSI: The CMU-MOSI dataset is a collection of 93 opinion videos from You Tube movie reviews. Each video consists of multiple opinion segments (2199 segments in total) and each segment is annotated with the score in the range [ 3, 3], where 3 and 3 indicate highly negative and positive. There are 1284 segments in the training set, 229 in the validation set, and 686 in the test set. POM: POM is a multimodal speaker traits recognition dataset made up of 903 movie review videos. Each video is annotated for various personality and speaker traits, specifically: Confident (con), Model \Metric BA F1 MAE Corr MA MV-LSTM [27] 73.9 74.0 1.019 0.601 33.2 TFN [38] 73.9 73.4 1.040 0.633 32.1 MARN [40] 77.1 77.0 0.968 0.625 34.7 MFN [39] 77.4 77.3 0.965 0.632 34.1 RAVEN [31] 78.0 0.915 0.691 LMF [19] 76.4 75.7 0.912 0.668 32.8 MAN (Non-Transformer) 79.4 79.3 0.894 0.697 37.8 Mul T [29] 83.0 82.8 0.870 0.698 40.0 MAN (Transformer) 82.7 83.0 0.866 0.707 40.5 Table 1: MAN achieves superior performances over baseline models for CMU-MOSI dataset. We report BA (binary accuracy), F1, Corr (Pearson Correlation Coefficient), and MA (Multi-class accuracy, all higher is better), MAE (Mean-absolute Error, lower is better). Model \ Trait Con Pas Voi Dom Cre Viv Exp Ent MA7 MA7 MA7 MA7 MA7 MA7 MA7 MA7 MV-LSTM [27] 25.6 28.6 28.1 34.5 25.6 32.5 32.5 29.6 TFN [38] 24.1 31.0 31.5 34.5 24.6 25.6 27.6 29.1 MARN [40] 29.1 33.0 - - 31.5 - - - MFN [39] 34.5 35.5 37.4 41.9 34.5 36.9 36.0 37.9 LMF [29] 35.9 35.9 34.8 39.6 34.5 35.9 37.8 36.5 MAN (Non-Transformer) 37.5 38.3 37.9 43.6 37.0 38.7 40.5 38.9 Mul T [29] 34.5 34.5 36.5 38.9 37.4 36.9 37.9 39.4 MAN (Tranformer) 37.9 38.9 37.9 44.8 37.9 39.9 40.9 40.4 Model \ Trait Res Tru Rel Out Tho Ner Per Hum MA5 MA5 MA5 MA5 MA5 MA5 MA7 MA5 MV-LSTM [27] 33.0 52.2 50.7 38.4 37.9 42.4 26.1 38.9 TFN [38] 30.5 38.9 35.5 37.4 33.0 42.4 27.6 33.0 MARN [40] 36.9 - 52.2 - - 47.3 31.0 44.8 MFN [39] 38.4 57.1 53.2 46.8 47.3 47.8 34.0 47.3 LMF [29] 35.5 54.2 53.2 44.8 42.7 43.5 34.9 45.8 MAN (Non-Transformer) 39.6 59.5 55.2 47.4 46.5 47.0 37.2 49.8 Mul T [29] 41.4 60.6 54.2 43.3 49.3 46.3 33.5 43.3 MAN (Transformer) 41.9 61.4 55.2 49.1 44.8 48.3 38.3 51.2 Table 2: MAN achieves superior performances over baseline models in POM dataset (multimodal personality traits recognition). MA(5,7) denotes multi-class accuracy for (5,7) classes. Passionate (Pas), Voice Pleasant (Voi), Dominant (Dom), Credible (Cre), Vivid (Viv), Expertise (Exp), Entertaining (Ent), Reserved (Res), Trusting (Tru), Relaxed (Rel), Outgoing (Out), Thorough (Tho), Nervous (Ner), Persuasive (Per) and Humorous (Hum). The training, validation, and test set distributions are approximately 600, 100, and 203, respectively. MSR-VTT: MSR-VTT is composed of 10K You Tube videos, collected using 257 queries from a commercial video search engine. Each video is 10 to 30s long, and is paired with 20 natural sentences describing it, obtained from Amazon Mechanical Turk Workers. We report results on the train/test splits introduced in [34] that uses 9000 videos for training and 1000 for testing. LSMDC: It contains 118081 short video clips (about 45s) extracted from 202 movies. Each clip is annotated with a caption, extracted from either the movie script or the audio description. The test set is composed of 1000 videos, from movies not presented in the training set. Model \Metric BA F1 MAE Corr MA w/o CAO 78.4 78.3 0.924 0.668 35.2 w/o LSC 78.4 78.3 0.928 0.673 36.5 w/o HAD 79.4 79.3 0.901 0.693 38.1 MAN (Non-Transformer) 79.4 79.3 0.894 0.697 37.8 w/o CAO 81.5 81.6 0.880 0.692 38.9 w/o LSC 81.4 81.3 0.883 0.698 39.8 w/o HAD 82.6 82.8 0.872 0.707 40.5 MAN (Transformer) 82.7 83.0 0.866 0.707 40.5 Table 3: Ablation study on CMU-MOSI dataset. Metric \Model w/o CAO w/o HAD w/o LSC MAN FLOPs 8.31 106 10.23 106 3.07 105 3.87 105 Table 4: Computational complexity of different variants on CMU-MOSI, note that we only consider the FLOPs of one-layer integration module. 3.2 Experiments for Multimodal Sentiment Analysis and Speaker Traits Recognition Data Preprocessing: Each dataset (CMU-MOSI, POM) consists of three modalities, including textual, visual, and audio modalities. For textual features, we employ the pre-trained 300-dimensional Glove embeddings [24]. For visual features, we utilize Facet [11] to indicate 35 facial action units, which records facial muscle movement for representing the basic and advanced emotions. For audio features, we use COVAREP [4] acoustic analysis framework. To align the different modalities along the temporal dimension, we perform word alignment with P2FA [37], which aligns the modalities at the word granularity. Experimental Details: Transformer-based and Non-Transformer-based methods both have their practical meanings of existence. For example, without enough labeled data, non-transformer-based methods may be an optimal solution due to less computational complexity. Henceforce, we combine MAN with both Transformer-based and Non-Transformer-based backbones and make corresponding comparisons with existing methods. We refer to [29] and [19] for our implementation of backbones, respectively. Note that we do not remove the original structures of [29] and [19], but add MAN as an extra module. We fuse their results for the tasks. The hyperparameters of MAN include Adam learning rate 0.001, the structure of integration network (1 layer of integration block, with hidden sizes of 40, number of heads 10, number of random features 24). Hidden size denotes the common size of d1, d2,...,dm. We divide all the time steps into 4 chunks and apply local sequential constraints. Furthermore, for the specific hyperparameters of backbones, we use similar values as original papers. Results: Table 1 presents the overall comparison of MAN and existing methods on CMU-MOSI dataset. As for the Non-Transformer-based methods, we could observe that MV-LSTM, MFN, MARN, RAVEN perform worse than MAN as they pay more attention to multimodal interaction and representation learning, and ignore the importance of subsequent multimodal integration. Besides, MAN achieves the best performances on all the metrics among the existing multimodal integration methods TFN, LMF and gains a large margin. Since TFN, LMF neglect the fine-grained temporal interaction which includes rich structured information for multimodal modeling. Particularly, the performance of MAN decreases the MAE from 0.912 to 0.894 compared to the best counterparts. As for the Transformer-based methods, MAN also performs better than Mul T on all the metrics, which demonstrates that, to some extent, subsequent multimodal joint representation learning could be integrated with the early multimodal interaction, even though only multimodal interaction has been able to achieve competitive performances. In general, the best performances of MAN attribute to the advanced hierarchical approximation decomposition which imposes fine-grained temporal correlation with acceptable computational complexity, as well as the addition random features mechanism that provides support for the approximation decomposition. Table 2 shows the experimental results of different methods on speaker traits recognition dataset POM, where we report the multi-class accuracy of all the traits. The similar observation could be found from the table, MAN achieves competitive performances compared with both Transformer-based and Non-Transformer-based methods on most of the traits. The only difference is that the metrics of Transformer-based and Non-Transformer-based methods are close. Therefore, even the Transformer has great potential in multimodal learning, its ability may be limited by the number of training samples. We argue that Non-Transformer-based methods still have their meanings in the real-world applications. Figure 2: The evaluation of different numbers of random features, where we report the values of Corr and MA on CMU-MOSI. Ablation Study: We set some control experiments on CMU-MOSI to verify the effectiveness of MAN and the results are shown in Table 3 and Table 4, where w/o CAO denotes the model without combinational addition operation in Eq. 7, we simply employ the continued Hadamard product and do not decompose the attention map, in this way, the local constraints also cannot be utilized. w/o HAD denotes the model without hierarchical approximation decomposition, but with other contributions. w/o LSC denotes the model with all the contributions except for local sequential constraints. We could observe that w/o HAD performs similarly to MAN with much more computational complexity, since w/o HAD needs to generate a high-order attention map with exponential complexity. Besides, MAN and w/o HAD perform much better than w/o CAO , since w/o CAO generates denser attention distributions and cannot capture really important information. Additionally, by comparing the metrics of w/o LSC and MAN, we argue that the local constraints have impacts on better capturing the correlation between multimodal sequential features. Since we utilize addition random features mechanism to approximate the softmax function, the number of random features should be considered. We examine the performances of MAN (Non-Transformer) on CMU-MOSI with different values of H. As shown in Fig. 2, when the value of H is big enough, MAN achieves competitive performances of Corr and MA close to w/o HAD . 3.3 Experiments for Multimodal Video Retrieval Data Preprocessing: The videos contain abundant multimodal information. Thus, we use multiple pre-trained models for extracting features. Concretely, we utilize following seven experts: Motion embeddings are extracted from S3D [32] trained on the kinetics dataset. Scene embeddings are extracted with Dense Net-161 [10] trained on the Places365 dataset [42]. OCR embeddings are extracted in three stages. First, the pixel link text detection model is used to detect the overlaid text. Then, the detected boxes are passed through the text recognition model. Finally, each character sequence is encoded using a word2vec embedding. Audio embeddings are obtained with a VGGish model, trained on the You Tube-8m dataset. Speech features are extracted using the Google Cloud speech API, to extract word tokens from the audio stream, which are then encoded via pre-trained word2vec embeddings [22]. Face features are extracted by Res Net-50 [8] trained for face classification on the VGGFace2 dataset. Appearance features are extracted from the final global average pooling layer of SENet-154 [9] trained on Image Net. Experimental Details: We also combine MAN with Transformer-based and Non-Transformer-based backbones. For MAN (Non-Transformer), each word of the text query is encoded with pre-trained word2vec embeddings and then passed through the GPT model. For MAN (Transformer), we adopt the basic structure of [6] and train the word vectors from scratch. We refer to [18] and [6] for the implementation of backbones and keep the original structures, only adding MAN as an extra module. We further fuse their results. The hyperparameters of MAN include Adam learning rate 5 10 5, which we decay by a multiplicative factor 0.95 every 1000 optimization steps, the structure of integration network (1 layer of integration block, with hidden size 512, number of heads 8, number of random features 512). Hidden size denotes the common size of d1, d2,...,dm. We divide the time steps into 10 chunks and employ local sequential constraints. We also use the same hyperparameters of specific backbones as the original papers. Text Video Video Text Model \Metric R@1 R@5 R@10 Md R R@1 R@5 R@10 Md R JSFusion [34] 10.2 31.2 43.2 13 - - - - HT [21] 14.9 40.2 52.8 9 - - - - CE [18] 20.9 48.8 62.4 6 20.6 50.3 64.0 5.3 MAN (Non-Transformer) 21.4 51.0 63.5 5 21.6 51.3 65.4 5 MMT [6] 24.6 54.0 67.1 4 24.4 56.0 67.8 4 MAN (Transformer) 24.1 55.7 68.1 4 24.4 57.1 68.5 4 Table 5: Retrieval performances on the MSR-VTT dataset. Text Video Video Text Method \Metric R@1 R@5 R@10 Md R R@1 R@5 R@10 Md R CT-SAN [35] 5.1 16.3 25.2 46 - - - - JSFusion [34] 9.1 21.2 34.1 36 - - - - CCA [16] (rep. by [20]) 7.5 21.7 31.0 33 - - - - MEE [20] 10.1 25.6 34.6 27 - - - - CE [18] 11.2 26.9 34.8 25.3 - - - - MAN (Non-Transformer) 11.8 27.9 36.0 24 - - - - MMT [6] 13.2 29.2 38.8 21 12.1 29.3 37.9 22.5 MAN (Transformer) 13.6 30.0 39.4 20 12.5 29.8 38.7 22 Table 6: Retrieval performances on the LSMDC dataset. Results: We report the evaluation results of MAN and the competing text-video retrieval methods on MSR-VTT (Table 5) and LSMDC (Table 6). In order to be fair, we still divide all the methods into two categories, Non-Transformer-based and Transformer-based methods. As for the former, MAN outperforms existing state-of-the-art methods in all the metrics on both MSR-VTT and LSMDC. Benefiting from the fine-grained temporal correlation of different modalities, the final representation contains not only the global multimodal information, but also local temporal interaction information. When we replace the backbone with Transformer, the metrics of MMT and MAN (Transformer) are very close, in other words, the effect of MAN is diminished. It may be because that the sample number of MSR-VTT and LSMDC is sufficient to stimulate the potential of Transformer. 4 Related Work Bilinear Model. [5] proposes multimodal compact bilinear pooling network (MCB), which calculates the outer product between the visual and textual embeddings. The count-sketch projection is utilized to project the product on a lower-dimensional space. [15] proposes multimodal low-rank bilinear pooling network (MLB), where the full bilinear interaction between image and question spaces are parametrized by a tensor. To limit the number of free parameters, this tensor is constrained to be of low rank r. [1] proposes multimodal tucker network (MUTAN), which reduces the size of mono-modal embeddings while modeling their interaction as accurately as possible with a full bilinear integration scheme. [14] proposes bilinear attention network (BAN) to utilize the given vision-language information seamlessly by finding the bilinear attention distributions. Concretely, BAN considers bilinear interactions among two groups of input channels, while low-rank bilinear pooling extracts the joint representations for each pair of channels. Based on BAN, [2] and [17] propose relational reasoning by constructing a fine-grained graph between different input channels. [12] also tries to decompose the temporal projection tensor, but without theoretical support. Non-Linear Approximation. Many endeavors have been devoted to the approximation of non-linear function. [13] employs a generalizable statistic expectation to approximate the non-linear kernels, specifically, [13] first defines a basic transformation (approximation of simply dot-product) and then utilizes Maclaurin expansion to expand out the non-linear functions to employ the conclusion of basic transformation. Such approximation accumulates two approximate errors. [26] proposes Tensor Sketching to approximate polynomial kernel functions based on Fourier transform. [5] develops pairlevel integration of two vectors by leveraging the conclusions of [13] and [26], such method achieves superior results on visual classification. Recently, [3] and [23] introduce the kernel decomposition into Transformer, they mainly approximate the exponential kernel with statistic expectation. The existing kernel decomposition methods are all proposed to process two vectors, considering the multimodal tasks, we first try to generalize the decomposition methods to more objects. 5 Conclusion We propose a generalizable method called multi-linear attention network (MAN) to integrate as many modalities as possible. Compared with the existing multimodal joint representation learning methods, MAN conducts the hierarchical approximation decomposition (HAD) to reduce the exponential computational complexity. Considering that softmax function cannot be decomposed directly, we propose the addition random features mechanism (ARF) to reconstruct the non-linear operation with orthogonal projections. Furthermore, we employ local sequential constraints (LSC) to suppress the association scores of features that are virtually irrelevant. In the future, we would focus on how to combine powerful interaction methods (i.e. Transformer) with joint representation learning methods (i.e. MAN) more seamlessly. Acknowledgments This work was supported in part by the National Key R&D Program of China under Grant No.2020YFC0832505, National Natural Science Foundation of China under Grant No.61836002, No.62072397 and Zhejiang Natural Science Foundation under Grant LR19F020006. [1] Hedi Ben-Younes, Rémi Cadene, Matthieu Cord, and Nicolas Thome. Mutan: Multimodal tucker fusion for visual question answering. In Proceedings of the IEEE international conference on computer vision, pages 2612 2620, 2017. [2] Remi Cadene, Hedi Ben-Younes, Matthieu Cord, and Nicolas Thome. Murel: Multimodal relational reasoning for visual question answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1989 1998, 2019. [3] Krzysztof Choromanski, Valerii Likhosherstov, David Dohan, Xingyou Song, Andreea Gane, Tamas Sarlos, Peter Hawkins, Jared Davis, Afroz Mohiuddin, Lukasz Kaiser, et al. Rethinking attention with performers. ar Xiv preprint ar Xiv:2009.14794, 2020. [4] Gilles Degottex, John Kane, Thomas Drugman, Tuomo Raitio, and Stefan Scherer. Covarep a collaborative voice analysis repository for speech technologies. In ICASSP, 2014. [5] Akira Fukui, Dong Huk Park, Daylen Yang, Anna Rohrbach, Trevor Darrell, and Marcus Rohrbach. Multimodal compact bilinear pooling for visual question answering and visual grounding. ar Xiv preprint ar Xiv:1606.01847, 2016. [6] Valentin Gabeur, Chen Sun, Karteek Alahari, and Cordelia Schmid. Multi-modal transformer for video retrieval. In European Conference on Computer Vision (ECCV), volume 5. Springer, 2020. [7] Richard A Harshman et al. Foundations of the parafac procedure: Models and conditions for an" explanatory" multimodal factor analysis. 1970. [8] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [9] Jie Hu, Li Shen, and Gang Sun. Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7132 7141, 2018. [10] Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q Weinberger. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4700 4708, 2017. [11] i Motions. Facial expression analysis. 2017. [12] Tao Jin, Siyu Huang, Yingming Li, and Zhongfei Zhang. Dual low-rank multimodal fusion. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, pages 377 387, 2020. [13] Purushottam Kar and Harish Karnick. Random feature maps for dot product kernels. In Artificial intelligence and statistics, pages 583 591. PMLR, 2012. [14] Jin-Hwa Kim, Jaehyun Jun, and Byoung-Tak Zhang. Bilinear attention networks. ar Xiv preprint ar Xiv:1805.07932, 2018. [15] Jin-Hwa Kim, Kyoung-Woon On, Woosang Lim, Jeonghee Kim, Jung-Woo Ha, and Byoung Tak Zhang. Hadamard product for low-rank bilinear pooling. ar Xiv preprint ar Xiv:1610.04325, 2016. [16] Benjamin Klein, Guy Lev, Gil Sadeh, and Lior Wolf. Associating neural word embeddings with deep image representations using fisher vectors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4437 4446, 2015. [17] Linjie Li, Zhe Gan, Yu Cheng, and Jingjing Liu. Relation-aware graph attention network for visual question answering. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 10313 10322, 2019. [18] Yang Liu, Samuel Albanie, Arsha Nagrani, and Andrew Zisserman. Use what you have: Video retrieval using representations from collaborative experts. ar Xiv preprint ar Xiv:1907.13487, 2019. [19] Zhun Liu, Ying Shen, Varun Bharadhwaj Lakshminarasimhan, Paul Pu Liang, Amir Zadeh, and Louis-Philippe Morency. Efficient low-rank multimodal fusion with modality-specific factors. In ACL, 2018. [20] Antoine Miech, Ivan Laptev, and Josef Sivic. Learning a text-video embedding from incomplete and heterogeneous data. ar Xiv preprint ar Xiv:1804.02516, 2018. [21] Antoine Miech, Dimitri Zhukov, Jean-Baptiste Alayrac, Makarand Tapaswi, Ivan Laptev, and Josef Sivic. Howto100m: Learning a text-video embedding by watching hundred million narrated video clips. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2630 2640, 2019. [22] Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. Efficient estimation of word representations in vector space. ar Xiv preprint ar Xiv:1301.3781, 2013. [23] Hao Peng, Nikolaos Pappas, Dani Yogatama, Roy Schwartz, Noah A Smith, and Lingpeng Kong. Random feature attention. ar Xiv preprint ar Xiv:2103.02143, 2021. [24] Jeffrey Pennington, Richard Socher, and Christopher D Manning. Glove: Global vectors for word representation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pages 1532 1543, 2014. [25] Verónica Pérez-Rosas, Rada Mihalcea, and Louis-Philippe Morency. Utterance-level multimodal sentiment analysis. In ACL, 2013. [26] Ninh Pham and Rasmus Pagh. Fast and scalable polynomial kernels via explicit feature maps. In Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 239 247, 2013. [27] Shyam Sundar Rajagopalan, Louis-Philippe Morency, Tadas Baltrusaitis, and Roland Goecke. Extending long short-term memory for multi-view structured learning. In ECCV, 2016. [28] Anna Rohrbach, Marcus Rohrbach, Niket Tandon, and Bernt Schiele. A dataset for movie description. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3202 3212, 2015. [29] Yao-Hung Hubert Tsai, Shaojie Bai, Paul Pu Liang, J Zico Kolter, Louis-Philippe Morency, and Ruslan Salakhutdinov. Multimodal transformer for unaligned multimodal language sequences. In Proceedings of the conference. Association for Computational Linguistics. Meeting, volume 2019, page 6558. NIH Public Access, 2019. [30] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. ar Xiv preprint ar Xiv:1706.03762, 2017. [31] Yansen Wang, Ying Shen, Zhun Liu, Paul Pu Liang, Amir Zadeh, and Louis-Philippe Morency. Words can shift: Dynamically adjusting word representations using nonverbal behaviors. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 7216 7223, 2019. [32] Saining Xie, Chen Sun, Jonathan Huang, Zhuowen Tu, and Kevin Murphy. Rethinking spatiotemporal feature learning: Speed-accuracy trade-offs in video classification. In Proceedings of the European Conference on Computer Vision (ECCV), pages 305 321, 2018. [33] Jun Xu, Tao Mei, Ting Yao, and Yong Rui. Msr-vtt: A large video description dataset for bridging video and language. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5288 5296, 2016. [34] Youngjae Yu, Jongseok Kim, and Gunhee Kim. A joint sequence fusion model for video question answering and retrieval. In Proceedings of the European Conference on Computer Vision (ECCV), pages 471 487, 2018. [35] Youngjae Yu, Hyungjin Ko, Jongwook Choi, and Gunhee Kim. End-to-end concept word detection for video captioning, retrieval, and question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3165 3173, 2017. [36] Zhou Yu, Jun Yu, Jianping Fan, and Dacheng Tao. Multi-modal factorized bilinear pooling with co-attention learning for visual question answering. In ICCV, 2017. [37] Jiahong Yuan and Mark Liberman. Speaker identification on the scotus corpus. Journal of the Acoustical Society of America, 2008. [38] Amir Zadeh, Minghai Chen, Soujanya Poria, Erik Cambria, and Louis-Philippe Morency. Tensor fusion network for multimodal sentiment analysis. In EMNLP, 2017. [39] Amir Zadeh, Paul Pu Liang, Navonil Mazumder, Soujanya Poria, Erik Cambria, and Louis Philippe Morency. Memory fusion network for multi-view sequential learning. In AAAI, 2018. [40] Amir Zadeh, Paul Pu Liang, Soujanya Poria, Prateek Vij, Erik Cambria, and Louis-Philippe Morency. Multi-attention recurrent network for human communication comprehension. In AAAI, 2018. [41] Amir Zadeh, Rowan Zellers, Eli Pincus, and Louis-Philippe Morency. Mosi: multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos. ar Xiv preprint ar Xiv:1606.06259, 2016. [42] Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. Places: A 10 million image database for scene recognition. IEEE transactions on pattern analysis and machine intelligence, 40(6):1452 1464, 2017.