# merlot_multimodal_neural_script_knowledge_models__4a6a46ce.pdf MERLOT: Multimodal Neural Script Knowledge Models Rowan Zellers Ximing Lu Jack Hessel Youngjae Yu Jae Sung Park Jize Cao Ali Farhadi Yejin Choi Paul G. Allen School of Computer Science & Engineering, University of Washington Allen Institute for Artificial Intelligence https://rowanzellers.com/merlot As humans, we understand events in the visual world contextually, performing multimodal reasoning across time to make inferences about the past, present, and future. We introduce MERLOT, a model that learns multimodal script knowledge by watching millions of You Tube videos with transcribed speech in an entirely label-free, self-supervised manner. By pretraining with a mix of both framelevel (spatial) and video-level (temporal) objectives, our model not only learns to match images to temporally corresponding words, but also to contextualize what is happening globally over time. As a result, MERLOT exhibits strong out-of-the-box representations of temporal commonsense, and achieves state-ofthe-art performance on 12 different video QA datasets when finetuned. It also transfers well to the world of static images, allowing models to reason about the dynamic context behind visual scenes. On Visual Commonsense Reasoning, MERLOT answers questions correctly with 80.6% accuracy, outperforming stateof-the-art models of similar size by over 3%, even those that make heavy use of auxiliary supervised data (like object bounding boxes). Ablation analyses demonstrate the complementary importance of: 1) training on videos versus static images; 2) scaling the magnitude and diversity of the pretraining video corpus; and 3) using diverse objectives that encourage full-stack multimodal reasoning, from the recognition to cognition level. Visual Commonsense Reasoning Why is the man Commonsense Single-Image QA What s she holding onto before he leaves? Which of the chef s hands has a watch? Figure 1: Multimodal Event Representation Learning Over Time. We learn representations of multimodal script knowledge from 6 million You Tube videos. These representations can then be applied to a variety of downstream tasks that require commonsense or temporal visual reasoning. 1 Introduction The human capacity for commonsense reasoning is shaped by how we experience causes and effects over time. Consider the still image of people dining at a restaurant in the bottom right of Figure 1: while a literal, concrete description like people sitting at a table eating" might be technically correct for the static scene, it doesn t capture the richer temporal, commonsense inferences that are nonetheless obvious: before sitting down, the people had to meet up, agree where to go, and enter the : Equal contribution. 35th Conference on Neural Information Processing Systems (Neur IPS 2021). restaurant; at present, the man is pointing because the server just came to the table, and she might want to know whose food is whose; and after, it is likely the server will return to the kitchen to help another table. Teaching machines this type of script knowledge [95] is a significant challenge in no small part because enumerating all facts, inferences, and counterfactuals is prohibitive. As a result, the highest performing models on vision-and-language tasks, including Visual Commonsense Reasoning (VCR) (where Figure 1 s scene originates from), learn about the visual world exclusively through static images paired with literal captions [108, 22, 69, 75, 119, 36]. Though some captions might hint at the past and future, it is not obvious that even training on, e.g., 400M literal image/text pairs [89] will result in models capable of temporal reasoning. In this paper, we introduce MERLOT, short for Multimodal Event Representation Learning Over Time. MERLOT is a model that learns commonsense representations of multimodal events by selfsupervised pretraining over 6M unlabelled You Tube videos. With the goal of learning multimodal reasoning capacity beyond static images/literal captions, we train MERLOT to a) match individual video frames with contextualized representations of the associated transcripts, and to b), contextualize those frame-level representations over time by unmasking" distant word-level corruptions [27] and reordering scrambled video frames. We validate our model on a diverse suite of video tasks, requiring both recognitionand cognition-level reasoning across long and short timescales; when finetuned, MERLOT achieves a new state-of-theart on 12 such tasks. Additionally, we show that our script-knowledge representations transfer to the single image domain. On Visual Commonsense Reasoning (VCR; [123]), our model achieves particularly strong performance, outperforming models that require heavy visual supervision (in the form of object detection bounding boxes, or images paired with pristine captions). Beyond finetuning, we show both quantitatively and qualitatively that MERLOT has a strong outof-the-box understanding of everyday events and situations. Given a scrambled visual story, [50, 2], MERLOT can sort image sequences to match captions which tell a globally coherent narrative. Despite considerable domain shift from videos to static images, MERLOT outperforms strong baselines like CLIP [89] and UNITER [22], which independently match images to text and thus cannot reason over long-term contexts as effectively. This capacity for temporal coherence emerges during pretraining: analysis of MERLOT s attention patterns (Figure 11) show that regions attend to captions that are distant in time (and vice versa), allowing it perform cross-modal coreference to piece together a holistic view of situations. Finally, ablations of MERLOT show that 1) pretraining works better when we train on videos rather than still images, aided crucially by our strategy of corrupting highly visual words in the masked language modeling task, 2) using a diverse set of videos covering many aspects of everyday situations improves downstream performance compared to curated instructional video corpora [107, 80] which both cover a smaller slice of the visual world (confirming hypotheses from past work [47]); and 3) MERLOT s performance does not saturate even after many epochs of training on the pretraining corpus we curated, YT-Temporal-180M, as it continues to improve performance simply with more pretraining. The combination of these results suggests that learning full-stack visual reasoning and multimodal world knowledge from video data is a promising path forward for future research. In summary, our main contributions are: 1. MERLOT a performant end-to-end vision and language model, that learns powerful multimodal world representations from videos and their transcripts using no labeled data. 2. YT-Temporal-180M, a diverse corpus of frames/ASR derived from a filtered set of 6M diverse You Tube videos, which we show greatly aids performance, and 3. A set of experiments/ablations demonstrating the strong performance of MERLOT on a set of 14 tasks, spanning finetuning and zero-shot transfer, and images and videos. At rowanzellers.com/merlot, we have released code, data, and models for public research use. 2 Related Work 2.1 Joint representations of written text and images There is a long history of work on learning joint text-image representations [14]. Recently, several papers have proposed Visual BERT models [108, 22, 8, 69, 75, 119, 36], trained on image captioning datasets such as MSCOCO [71]. In general, features are extracted using Anderson et al. [10] s frozen object detector, which was originally trained on Visual Genome [60]. Some exceptions are Zhang et al. [125], who use an even larger object detector trained on more labeled data; Kim et al. [57], who use an Image Net-pretrained backbone [26], and Shen et al. [100], who study a CLIP backbone [89] pretrained on web image-caption pairs. Overall, these approaches all learn visual representations of static images, and rely on significant human annotation in doing so (e.g. through literal image descriptions). Instead, our approach learns dynamic visual representations purely from videos their frames, and a transcript of what is said thus using no human annotation. 2.2 Learning from videos, with automatic speech recognition (ASR) transcripts Prior works have used web videos with ASR to build weakly-supervised object detectors [87], action detectors/classifiers [120, 6, 62, 84], instruction aligners [77, 5, 19], video captioners [96, 46, 86, 101], and visual reference resolvers [49]. Of late, works have sought to learn multimodal representations transferable to many tasks from uncurated sets of (usually how-to) videos [80, 106, 107, 81, 127, 9, 7, 4]; generally these are applied to video understanding tasks like activity recognition. One challenge is designing an appropriate objective for learning video-level representations. Lei et al. [67] s Clip BERT model learns vision-language representations from image captions, which more literally describe image content versus the longer ASR transcripts we consider. Tang et al. [109] use a pretrained dense image captioner [59] to provide auxiliary labels for web how-to videos. Both approaches use (supervised) Res Nets pretrained on Image Net [43] as their visual backbones. MERLOT is trained using a combination of objectives requiring no manual supervision; it nonetheless outperforms both prior approaches on downstream tasks. 2.3 Temporal ordering and forecasting There has been a large body of work on analyzing what happens next in videos [58]. Some modeling choices include using pixels [34, 113], graphs [11], euclidean distance using sensors [3], or studying cycle consistency across time [32]. In addition to extrapolation, past work has studied deshuffling objectives in videos [82, 115], though this has mostly been limited to the visual modality. In contrast to these papers, our goal is learning multimodal script knowledge representations: using both language and vision as complementary views into the world, instead of just tracking what changes on-screen. 3 MERLOT: Multimodal Event Representation Learning Over Time We now present our unified model for learning script knowledge through web videos; including our pretraining dataset, architecture, and objectives. 3.1 YT-Temporal-180M We collect YT-Temporal-180M, a dataset for learning multimodal script knowledge, derived from 6 million public You Tube videos. Our YT-Temporal-180M intentionally spans many domains, datasets, and topics. We began with 27 million candidate video IDs (which we then filtered), including instructional videos from How To100M [80], lifestyle vlogs of everyday events from the VLOG dataset [35], and You Tube s auto-suggested videos for popular topics like science or home improvement. Our intent (in making the corpus as diverse as possible) was to encourage the model to learn about a broad range of objects, actions, and scenes [47]: we will later show through an ablation that limiting our pretraining to only instructional videos indeed hurts performance downstream. We filtered videos using the You Tube API, which provides access to videos themselves, their ASR track (automatically transcribed speech tokens), and other metadata. We discard videos 1) without Image Encoder Unmask words 1,1 CLS H,W Image Encoder CLS 1,1 H,W CLS 1 L 1 L CLS Joint Vision & Language Transformer Encoder It s thin plastic, so it ll be easy So I ll cut it with It s thin plastic, so it ll be easy We re making a greenhouse. So I ll cut it with a circular For my morning routine, I Language only encoder Temporal ordering I1 w1 w2 I2 Contrastive Frame-caption Figure 2: Left: MERLOT learns to match contextualized captions with their corresponding video frames. Right: the same image encoding is provided, along with (masked) word embeddings, into a joint vision-language Transformer model; it then unmasks ground words (like saw in this example) and puts scrambled video frames into the correct order. an English ASR track; 2) that are over 20 minutes long; 3) that belong to visually ungrounded" categories like video game commentaries; and 4) that have thumbnails unlikely to contain objects, according to a lightweight image classifier. We add punctuation to the ASR by applying a sequenceto-sequence model trained to add punctuation to sentences/paragraphs from news articles. Full details of the scraping and filtering are in Appendix A. Each video V might contain thousands of frames. In this work, we represent a video V as a sequence of consecutive video segments {st}. Each segment st consists of: a. an image frame It, extracted from the middle timestep of the segment, b. the words wt spoken during the segment, with a total length of L tokens. To split the videos into segments, we byte-pair-encode (BPE; [97, 88]) each video transcript and align tokens with You Tube s word-level timestamps. This enables us to split the videos into segments of L=32 BPE tokens each (Appendix A.4); our final dataset has 180 million segments of this form. 3.2 MERLOT Architecture A diagram of MERLOT is given in Figure 2. MERLOT takes a sequence of video frames {st} as input. We encode each frame It using an image encoder, embed the words wt using a learned embedding, and jointly encode both using a Transformer [112]. After pretraining, the architecture can be applied to a variety of vision-and-language tasks with minimal modification. For video QA, for example, we pass several video frames to the image encoder, the question to the text encoder, and extract a single vector representation from the CLS token position. For each task, we learn a lightweight classification head mapping from this hidden state to the task s label space; specific modeling/optimization details are given in Appendix E.2. Image encoder. We train our image encoder end-to-end, alongside the rest of the model, from random initialization (thus without learning from supervised data). While most performant visionand-language models pre-extract features from a (supervised) object detector [108, 69, 75, 22, 68], for the sake of pre-training efficiency we use a grid-based hybrid Res Net/Vision Transformer.1 Specifically: our encoder uses a Res Net-50 backbone, followed by a 12-layer, 768-dimensional Vision Transformer [43, 112, 31]. We made additional modifications that improve efficiency, including: 1) we trained on smaller, widescreen images of size 192x352 (because most You Tube videos are 1Standard object detectors have expensive operations for proposing regions, and extracting features from those regions (Ro I-pooling); our grid approach avoids these. Recent work has proposed using grid features broadly [53], yet on tasks like VCR these approaches have so far underperformed the more expensive object detector backbones [123]; our results suggest that grid features can perform well broadly. widescreen) using a patch size of 16x16 pixels; 2) we mirror [31] s alterations of removing the C5 block in Res Net-50; and 3) we save compute further by average-pooling the final-layer region cells using a kernel size of 2 2. With these modifications, our image encoder requires 40 giga FLOPs for a forward pass, which is 2% of the 2 tera FLOPs required for the Faster-RCNN. In summary: given an image of size W H, the image encoder will output a W/32 H/32 feature map, along with two CLS hidden states: one for pooling a global representation of the image, and another for pretraining (Task 1.). Joint Vision-Language Encoder. The joint encoder is a 12-layer, 768-dimensional Transformer [112], mirroring the Ro BERTa base architecture [72]; we initialize it with pretrained Ro BERTa weights. To compute joint representations, we first embed the tokens {wt} via lookup, and then add position embeddings to both language and vision components (i.e., {It}). The position embeddings differ between different segments, so as to distinguish between images and captions at different timesteps. Finally, we pass the independent visual and textual feature maps to our joint encoder. The tokens wt in each segment begin with a CLS token; recall that the feature maps for each frame It start with one as well. At those positions, we will later pool final-layer hidden-state representations, for use in pretraining along with downstream tasks. 3.3 Pretraining Tasks and Objectives We use the following three objectives to pretrain MERLOT, that cover full-stack visual reasoning from recognition subtasks (like object detection) that operate at the frame level, to more cognitive tasks that operate at the video level. 1. Contrastive frame-transcript matching [126, 89]. We want to ensure that the underlying image encoder produces helpful image representations. Thus, we use the video transcript to compute a language-only representation of each video segment; and use a contrastive loss to maximize its similarity to corresponding representations from the image encoder.2 Unlike what is the case for many image captions, the words wt in each segment are often not sufficient to describe the gist of It, or even what the key objects might be for that, video-level contextualization is often required. We thus pass the entire transcript into the language-only encoder, which then extracts hidden states for each segment at the segment-level CLS tokens. Given matching representations for each frame It and caption wt as positive examples, the negative examples come from all other frame-caption pairs in the batch whether or not they come from the same video. We project both of these representations into a size-768 hidden state which is then unit-L2-normalized, and compute an all-pairs dot-product between all image and text representations. We divide these logits by a temperature of τ = 0.05, and then apply a pairwise cross entropy loss to encourage matching captions and frames. 2. (Attention) Masked Language Modeling When providing words into the joint vision-andlanguage encoder, we randomly replace 20% with a MASK token, a random word, or the same word; MERLOT must then reconstruct the correct word with a cross-entropy loss, following [27]. This approach is commonly used by visual BERT models in the image captioning domain, where captions are concise, and thus the identity of masked concrete words is difficult for models to recover given language context alone. However, we observed qualitatively that videos break these assumptions: people tend to ramble, and often mention key objects multiple times. Thus, applying vanilla BERT-style masking often causes ungrounded fillers like umm or yeah to get masked, while the (repeated) names of important objects are often partially masked, penalizing the learning of multimodal representations. We introduce a simple solution to this problem, that we call attention masking: we use attention weights from a language-only transformer (introduced in the previous objective) as a heuristic for which words are grounded. 50% of the time, we mask out a random token; the other 50% of the time, we mask out one of the top 20% most-attended-to-tokens. We then apply Span BERT masking [54], randomly corrupting the following or preceding tokens with an average length of 0.5 tokens in each direction; this makes it harder for models to over-rely on BPE artifacts. We show in ablations that this improves performance. 2To save memory, our language-only encoder for this subtask shares parameters with the joint vision-andlanguage encoder. Q A QA R Q AR Vi LBERT [75] 73.3 74.6 54.8 Unicoder-VL [68] 73.4 74.4 54.9 VLBERT [69] 73.8 74.4 55.2 UNITER [22] 75.0 77.2 58.2 VILLA [36] 76.4 79.1 60.6 ERNIE-Vi L [119] 77.0 80.3 62.1 MERLOT (base-sized) 80.6 80.4 65.1 Table 1: Results on VCR [123]. We compare against SOTA models of the same base size as ours (12-layer vision-and-language Transformers). MERLOT performs best on all metrics. Spearman Pairwise acc Distance ( ) ( ) ( ) CLIP [89] .609 78.7 .638 UNITER [22] .545 75.2 .745 MERLOT .733 84.5 .498 Table 2: Results unscrambling SIND visual stories[50, 2]. Captions are provided in the correct order; models must arrange the images temporally. MERLOT performs best on all metrics by reasoning over the entire story, instead of independently matching images with captions. 3. Temporal Reordering. We have the model order the image frames in a video, forcing it to explicitly learn temporal reasoning and giving it an interface to measure such temporal reasoning. Here, 40% of the time, we randomly pick an integer i between 2 and N (the number of segments provided to the joint encoder). Then we randomly scramble i video frames chosen at random, by replacing the segment-level position embeddings (e.g. [image_t]) for that frame with a random and unique position embedding, e.g. [image_unk_0]). These random position embeddings are learned, and separate from the unshuffled position embeddings. This allows the model to order each shuffled frame conditioned on frames provided in the correct order (if any). To compute the reordering loss, we extract hidden states from each frame at the CLS token position. For each pair of frames, we concatenate their hidden states hti and htj and pass the result through a two-layer MLP, predicting if ti < tj or ti > tj. We optimize this using a cross-entropy loss. 3.4 Pretraining MERLOT We pretrain our model for 40 epochs over our video dataset. We preprocess the dataset into examples with sequences of N=16 video segments each, each containing up to L=32 BPE tokens.3 The language-only encoder computes contrastive representations given this entire sequence, its total length is thus 512 tokens. To save memory, we provide the joint vision-language encoder 4 groups of N = 4 segments each. At an image training resolution of 192 352, the joint model s sequence length is 396 tokens. To combine the losses, we multiply the contrastive loss by a coefficient of 0.25, which we found scaled its gradient magnitudes to roughly the same magnitude as the Mask LM loss. We train the model using a v3-1024 TPU pod, at a batch size of 1024 sequences (or 16k segments) in total. This pretraining process on this hardware takes 30 hours. We provide additional information about hyperparameters and experimental setup in Appendix E.1. 4 Experiments: Transferring MERLOT to Downstream Tasks In this section, we explore MERLOT on 14 different tasks, covering vision-language reasoning on static images as well as videos; we present analysis and ablations to dig deeper into our performance. 4.1 Image tasks VCR. We consider VCR [123], a task and dataset where models must answer commonsense visual questions about images. These questions, about e.g. what might happen next or what are people s intentions, force MERLOT to transfer video-level understanding to the world of single images. VCR provides additional referring expression information to models in the form of bounding boxes around named entities. For example, if Person1 is referenced in the question, the location of Person1 is also given in the image. We provide this information to models by drawing (in pixel space) a 3To train the model on as much data as possible, we merged together the segments of short videos, and split up longer videos, such that all preprocessed examples in our dataset have exactly N=16 video segments. Tasks Split Vid. Length Act BERT [127] Clip BERT8x2 [67] SOTA MERLOT MSRVTT-QA Test Short - 37.4 41.5 [118] 43.1 MSR-VTT-MC Test Short 88.2 - 88.2 [127] 90.9 TGIF-Action Test Short - 82.8 82.8 [67] 94.0 TGIF-Transition Test Short - 87.8 87.8 [67] 96.2 TGIF-Frame QA Test Short - 60.3 60.3 [67] 69.5 LSMDC-Fi B QA Test Short 48.6 - 48.6 [127] 52.9 LSMDC-MC Test Short - - 73.5 [121] 81.7 Activity Net QA Test Long - - 38.9 [118] 41.4 Drama-QA Val Long - - 81.0 [56] 81.4 TVQA Test Long - - 76.2 [56] 78.7 TVQA+ Test Long - - 76.2 [56] 80.9 VLEP Test Long - - 67.5 [66] 68.4 Table 3: Comparison with state-of-the-art methods on video reasoning tasks. MERLOT outperforms state of the art methods in 12 downstream tasks that involve short and long videos. colored highlight around the referenced entity (Appendix E.3.1), this differs from prior works (that integrate these entities into detection architectures). Our results on the three VCR settings, in comparison to other models at the same ( base ) scale, are given in Table 1. Our model outperforms these other models, that all learn from exclusively static images (paired with captions and supervised object detections). Unsupervised ordering of Visual Stories. To probe our model s ability to do out-of-the-box commonsense reasoning over events in images, we next consider the Visual Storytelling dataset [50, 74]. Each story in this dataset contains five images and captions in a certain order; the order tells a joint narrative between the captions and the images. Past work has considered unshuffling imagecaption pairs [2], but we take a slightly different approach in this work to avoid language-only biases, which can rely on discursive clues to order text [27, 102]. In our formulation, models are given the captions in sorted order, and must match frames to the captions. Our formulation disarms language-only baselines, while still allowing us to quantify MERLOT s capacity for commonsense temporal reasoning. We compare MERLOT with two strong out-of-the-box baselines for text-image matching: CLIP [89], which encodes each caption and image separately and computes similarity through a dot product, and UNITER [22] which jointly represents each image/caption pair, and is trained in part using a textimage matching objective. We use our temporal reordering loss to find the most probable ordering of the video frames (Appendix E.1.1); for CLIP and UNITER we compute a maximum-weight bipartite matching [63] over the pairwise image-text similarity scores. Results over 5K stories are given in Table 2. MERLOT s performance in comparison to the algorithms trained from image-literal caption pairs suggests that, with no fine-tuning, our model has strong capability to reason about past and future events expressed in collections of temporal visual stories. 4.2 Video Reasoning We report results on 12 video reasoning tasks: TVQA [64], TVQA(+) [65], VLEP [66], MSRVTT-QA [117], MSRVTT-Multichoice [121], LSMDC-Multichoice, LSMDC fill-in-the-blank QA [110, 92], Activity Net QA [122, 45], TGIFQA [52], and Drama QA [23]. We apply MERLOT to these tasks in the same way. We sample a sequence of 5 to 7 still frames from each video clip, initialize new parameters only to map the model s pooled CLS hidden state into the output labels, and finetune MERLOT with a softmax cross entropy loss; see Appendix E.2 for details. As shown in Table 3, for all these datasets MERLOT sets a new state-of-the-art. Given the diversity of tasks and the strengths of the comparison models, these results provide strong evidence that MERLOT learned strong multimodal and temporal representations. 4.3 Ablations We present ablations over VCR and TVQA+ to study the effect of several modeling decisions. Training setup VCR TVQA+ One segment (N=1) 73.8 75.2 One segment, attention masking 73.5 74.5 Four segments 74.1 73.3 Four segments, attention masking 75.2 75.8 (a) Context helps together with attention masking. Pretraining on more segments at once improves performance, but more context can encourage language-only representation learning. Attention masking counteracts this, giving an additional 1 point boost. Training setup VCR TVQA+ No contrastive V-L loss 57.5 67.6 No temporal ordering loss 75.5 75.6 All losses 75.2 75.8 (b) Contrastive V+L loss is crucial. Removing it makes performance drop significantly; the temporal ordering loss is not as important for downstream finetuning. No boxes 74.8 Drawn-on boxes 79.4 (c) Drawing on bounding boxes helps, suggesting that our model uses it to decode the referring expression information (e.g. person1). Dataset VCR Conceptual COCO 58.9 How To100M 66.3 YT-Temporal-180M 75.2 How To100M-sized YT-Temporal-180M 72.8 YTT180M, raw ASR 72.8 (d) Diverse (video) data is important. Applying our architecture to caption data leads to poor results. Our model performs better on How To100M, yet still below our (more diverse) YT-Temporal-180M, even when controlled for size. Using raw ASR (vs. denoised ASR) reduces performance. # epochs VCR 5 epochs 75.2 10 epochs 75.9 20 epochs 77.0 30 epochs 78.5 40 epochs 79.4 (e) Training for longer helps, with performance increasing monotonically over training iterations. Table 4: Ablation study on the validation set of VCR question answering (Q A) and TVQA+, in accuraty (%). We put a next to the configurations we chose for MERLOT. Context size. Table 4a shows the effect of varying the number of segments N given to the joint vision-and-language encoder during pretraining. In the first two rows, we provide only a single video segment (N=1) to the model.4 In this limited regime, we find that our attention masking approach (preferential masking of tokens that were highly attended-to by the contrastive language-only encoder) does not outperform a strong baseline of masking spans randomly [54]. Yet, when we expand the sequence length to N=4 segments/128 tokens, our masking becomes more effective, improving by 1 point over the baseline. This supports our hypothesis (Section 3.3.2.) that text-only shortcuts become increasingly viable with length, and that our attention-masking approach counteracts them.5 Losses. In Table 4b, we ablate the losses. We find that the contrastive frame-transcript matching loss is crucial to performance, suggesting that an explicit objective is critical for the (randomly initialized) image backbone to learn visual representations. The temporal ordering loss appears less critical for downstream tasks; it helps for TVQA but performance drops slightly for VCR. Thus, we find that it helps primarily as an interface by which we can query the model about temporal events (i.e. for the story ordering experiments); the model might be learning this information from other objectives. Drawing bounding boxes. Table 4c shows the effects of providing grounding information to VCR models by drawing boxes. Performance drops 5% when they are removed, suggesting that they help. Dataset source. In Table 4d, we investigate pretraining MERLOT on two datasets beyond YTTemporal-180M. First, we train on 3 million static image-caption pairs from Conceptual Captions [99] combined with MSCOCO [71]; for fair comparison, we train for the same number of steps as 5 epochs on our dataset. The resulting model achieves 58.9% accuracy on VCR. We suspect this might be due to 1) a smaller context window (Table 4a), and 2) overfitting (5 epochs on YT-Temporal180M corresponds to 300 epochs on the caption data). Because our vision pipeline is trained from scratch, the scale of the curated/supervised image pairing corpora is a concern. We next investigate the impact of video selection, comparing YT-Temporal-180M with How To100M [80]. To control for number of videos, we train for an equivalent amount of steps: 5 epochs on our dataset, 30 epochs on How To100M, and likewise 30 epochs on a How To100M-sized YTTemporal-180M . Using diverse YT-Temporal-180M data vs. only instructional videos improves VCR performance by 6.5 points. This suggests that the how-to domain is limited in terms of visual 4We keep the effective batch size the same, so that we use 4 the number of sequences at 1 4th the length. 5Additional qualitative analyses of the attention patterns produced by the language-only encoder are in Appendix C.1; we find that highly attended-to tokens are typically more visual , and, thus, masking them may make the Masked LM objective require more cross-modal reasoning. Figure 3: Zero-shot story ordering (same setup as Table 2). MERLOT performs temporal commonsense reasoning accross frames. In the first row, it uses the old man mentioned to identify the kids as parent-aged; in the second, it identifies riding a merry-go-round as an activity that takes a while. phenomena covered, and that other domains (like web dramas and VLOGs) provide helpful signal for tasks like VCR [47]. Using all the data gives an additional 2.4-point performance boost. Last, we investigate our choice to preprocess the You Tube ASR text with a language model (adding punctuation, etc); using raw ASR instead of this preprocessing reduces performance by 2.4 points. Pretraining longer. Last, in Table 4e, we investigate the effect of pretraining MERLOT for longer. The performance increases monotonically and doesn t begin to plateau, which suggests that had we pretrained MERLOT for even longer, its performance could improve even further. 4.4 Qualitative examples In Figure 3, we show two qualitative examples of MERLOT s zero-shot story ordering capability. More examples (and a comparison with the best-scoring baseline, CLIP [89]) are in Appendix C.2. The examples here show that MERLOT has a strong understanding of events, transcending individual frames. In the first row, it orders the story correctly, performing vision-and-language coreference across several frames (e.g. frames and captions 2 and 3 use he to refer to the old man only mentioned in the first caption). Without resolving this coreference (establishing the subject as an elderly family member), it seems unlikely that anyone would describe the adults in frame (3) as kids. Investigating the attention patterns of MERLOT (Appendix C.3) backs up this claim; they show that MERLOT frequently addresses video tasks by merging attention across (distant) video segments. MERLOT gets the second row wrong , but for an interesting reason. It reverses the order of frames (3) and (4), which groups the merry-go-round pictures together even though caption (3) mentions a barn. This seems to capture the temporal commonsense intuition that people might ride a merry-go-round for a while, i.e., it is not an atomic event [25]. 5 Conclusion, Limitations, and Broader Impacts We introduced Multimodal Event Representation Learning Over Time (MERLOT). We trained the model through a combination of self-supervised objectives on 6M You Tube videos, in service of learning powerful multimodal representations that go beyond single frames. The model achieves strong performance on tasks requiring event-level reasoning over videos and static images. We hope that MERLOT can inspire future work for learning vision+language representations in a more human-like fashion compared to learning from literal captions and their corresponding images. There are several potential limitations of MERLOT that would make for promising avenues of future work, including: 1) exploring finer-grained temporal reasoning pretraining objectives vs. frame ordering e.g., a temporal frame localization within transcripts; and 2) learning multilingually from non-English videos and communities on You Tube. Like other pretraining work, MERLOT risks some potential negative impacts. We discuss these in more detail below, in addition to the steps we took to reduce these harms. 5.1 Data collection and privacy. As with other corpora gathered from the web used for pretraining data, YT-Temporal-180M contains publicly available content posted by users. We thus shaped our data gathering and release strategy to minimize inherent privacy and consent harms (Appendix A.5). Perhaps most importantly, we plan to only share video IDs for download, following a release strategy from prior work [1, 80] and giving users the right to opt out of not just You Tube, but our dataset as well. 5.2 Social biases. The curation choices we made in this work could cause the model to exhibit undesirable social biases for this reason, along with others, we do not advocate for deployed use-cases. For example, 30% of the data selected for by our filtering pipeline was local broadcast news (uploaded to You Tube). Including these news videos seems to perform better than filtering them out and only using how-to videos (Table 4b), however, there are risks when training on them. Local broadcast news (at least in the US) dedicates significant time to covering crime, sometimes in a racist and sensationalized manner [38, 29, 44]. Indeed, running a topic model over our data identifies several crime categories (Appendix B). Past work has shown correlation between watching local news and having more explicit racialized beliefs about crime [28]; it seems likely therefore that training models on this data could teach them learn the same racist patterns. Additionally, there are inherent social biases on You Tube and treating these videos as equivalent to the world [111] can embed hegemonic perspectives [42, 114, 13]. Most popular You Tubers are men [30] and video practices emerging on You Tube are often gendered [83]. You Tube also has problems with hate, including radical alt-right and alt-lite content [90]. These problems as with other problems in representation and power are themselves amplified by the You Tube algorithm [15] that recommends content to users. Though we downloaded videos independently of You Tube s recommender system, by filtering based on what content has views, we are implicitly filtering based on this algorithm. The dynamics of You Tube (i.e., which videos get popular/monetized) influence the style and content of videos that get made and uploaded to the platform; this in turn shapes and is shaped by culture more broadly [104]. 5.3 Dual use. The video QA tasks that we studied carry risk of dual use, through possible downstream applications like surveillance [91, 128]. It seems unlikely that purely technological fixes and defenses which themselves can be problematic [40] could resolve these dynamics. Studying how well video-level pretraining enables surveillance applications might be an important avenue for future work, if only to inform stakeholders and policymakers about these risks. 5.4 Energy consumption. The pretraining that we used in this work was expensive upfront [105]. Our results suggest that scaling up the amount of data and compute that we used might yield additional performance gains but at increased environmental cost. To pretrain more efficiently, we used a much more lightweight architecture (in terms of FLOPs) than is standard for today s vision and language models. We hope that our public release of the model (for research use) can further amortize this cost. 5.5 Synthesizing these risks. With these issues in mind, we release MERLOT and YT-Temporal-180M for researchers. We view our work, and our research artifacts, to be part of a larger conversation on the limits of pretrained foundation models [17]. These models have broad impact to real-world areas like healthcare, law, and education. At the same time, these models have significant risks, including the harms that we outlined. We believe that further academic research into this video-and-language pretraining paradigm is important especially to probe its limits and possible harms. We hope that our paper, code, and data release can contribute to this direction. Acknowledgements and Funding Transparency Statement We thank the anonymous reviewers for their helpful feedback that improved this work, along with Oren Etzioni and Gabriel Ilharco. Thanks also to Zak Stone and the Google Cloud TPU team for providing access to the TPU machines used for conducting experiments, and for help with the computing infrastructure. Last, but not least, thanks to all the You Tubers who share interesting videos with the world. This work was funded by DARPA MCS program through NIWC Pacific (N66001-19-2-4031), and the Allen Institute for AI. [1] Sami Abu-El-Haija, Nisarg Kothari, Joonseok Lee, Paul Natsev, George Toderici, Balakrishnan Varadarajan, and Sudheendra Vijayanarasimhan. Youtube-8m: A large-scale video classification benchmark. ar Xiv preprint ar Xiv:1609.08675, 2016. [2] Harsh Agrawal, Arjun Chandrasekaran, Dhruv Batra, Devi Parikh, and Mohit Bansal. Sort Story: Sorting Jumbled Images and Captions into Stories. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 925 931, 2016. [3] Pulkit Agrawal, Ashvin Nair, Pieter Abbeel, Jitendra Malik, and Sergey Levine. Learning to poke by poking: experiential learning of intuitive physics. In Proceedings of the 30th International Conference on Neural Information Processing Systems, pages 5092 5100, 2016. [4] Hassan Akbari, Linagzhe Yuan, Rui Qian, Wei-Hong Chuang, Shih-Fu Chang, Yin Cui, and Boqing Gong. VATT: transformers for multimodal self-supervised learning from raw video, audio and text. ar Xiv preprint ar Xiv:2104.11178, 2021. [5] Jean-Baptiste Alayrac, Piotr Bojanowski, Nishant Agrawal, Josef Sivic, Ivan Laptev, and Simon Lacoste-Julien. Unsupervised learning from narrated instruction videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4575 4583, 2016. [6] Jean-Baptiste Alayrac, Ivan Laptev, Josef Sivic, and Simon Lacoste-Julien. Joint discovery of object states and manipulation actions. In ICCV, 2017. [7] Jean-Baptiste Alayrac, Adrià Recasens, Rosalia Schneider, Relja Arandjelovi c, Jason Ramapuram, Jeffrey De Fauw, Lucas Smaira, Sander Dieleman, and Andrew Zisserman. Selfsupervised multimodal versatile networks. ar Xiv preprint ar Xiv:2006.16228, 2020. [8] Chris Alberti, Jeffrey Ling, Michael Collins, and David Reitter. Fusion of detected objects in text for visual question answering. ar Xiv preprint ar Xiv:1908.05054, 2019. [9] Elad Amrani, Rami Ben-Ari, Daniel Rotman, and Alex Bronstein. Noise estimation using density estimation for self-supervised multimodal learning. ar Xiv preprint ar Xiv:2003.03186, 2020. [10] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. Bottom-up and top-down attention for image captioning and visual question answering. In CVPR, 2018. [11] Peter Battaglia, Razvan Pascanu, Matthew Lai, Danilo Jimenez Rezende, and Koray kavukcuoglu. Interaction networks for learning about objects, relations and physics. In Proceedings of the 30th International Conference on Neural Information Processing Systems, pages 4509 4517, 2016. [12] Emily Bender and Batya Friedman. Data statements for nlp: Toward mitigating system bias and enabling better science. Transactions of the Association for Computational Linguistics, 2019. [13] Emily M Bender, Timnit Gebru, Angelina Mc Millan-Major, and Shmargaret Shmitchell. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pages 610 623, 2021. [14] Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8): 1798 1828, 2013. [15] Sophie Bishop. Anxiety, panic and self-optimization: Inequalities and the youtube algorithm. Convergence, 24(1):69 84, 2018. [16] David M Blei, Andrew Y Ng, and Michael I Jordan. Latent dirichlet allocation. Journal of machine Learning research, 3(Jan):993 1022, 2003. [17] Rishi Bommasani, Drew A Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, et al. On the opportunities and risks of foundation models. ar Xiv e-prints, pages ar Xiv 2108, 2021. [18] Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, et al. Extracting training data from large language models. ar Xiv preprint ar Xiv:2012.07805, 2020. [19] Chien-Yi Chang, De-An Huang, Yanan Sui, Li Fei-Fei, and Juan Carlos Niebles. D3tw: Discriminative differentiable dynamic time warping for weakly supervised action alignment and segmentation. In CVPR, 2019. [20] Snigdha Chaturvedi, Haoruo Peng, and Dan Roth. Story Comprehension for Predicting What Happens Next. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 1603 1614, 2017. [21] Peihao Chen, Deng Huang, Dongliang He, Xiang Long, Runhao Zeng, Shilei Wen, Mingkui Tan, and Chuang Gan. Rspnet: Relative speed perception for unsupervised video representation learning. In AAAI, 2021. [22] Yen-Chun Chen, Linjie Li, Licheng Yu, Ahmed El Kholy, Faisal Ahmed, Zhe Gan, Yu Cheng, and Jingjing Liu. UNITER: Learning universal image-text representations. ar Xiv preprint ar Xiv:1909.11740, 2019. [23] Seongho Choi, Kyoung-Woon On, Yu-Jung Heo, Ahjeong Seo, Youwon Jang, Seungchan Lee, Minsu Lee, and Byoung-Tak Zhang. Drama QA: character-centered video story understanding with hierarchical qa. ar Xiv preprint ar Xiv:2005.03356, 2020. [24] Kevin Clark, Urvashi Khandelwal, Omer Levy, and Christopher D Manning. What does bert look at? an analysis of bert s attention. In Proceedings of the 2019 ACL Workshop Blackbox NLP: Analyzing and Interpreting Neural Networks for NLP, pages 276 286, 2019. [25] William Croft. Verbs: Aspect and causal structure. OUP Oxford, 2012. [26] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on, pages 248 255. Ieee, 2009. [27] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. ar Xiv preprint ar Xiv:1810.04805, 2018. [28] Travis L Dixon. Crime news and racialized beliefs: Understanding the relationship between local news viewing and perceptions of african americans and crime. Journal of Communication, 58(1):106 125, 2008. [29] Travis L Dixon and Daniel Linz. Overrepresentation and underrepresentation of african americans and latinos as lawbreakers on television news. Journal of communication, 50(2): 131 154, 2000. [30] Nicola Döring and M Rohangis Mohseni. Male dominance and sexism on youtube: results of three content analyses. Feminist Media Studies, 19(4):512 524, 2019. [31] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. [32] Dave Epstein, Jiajun Wu, Cordelia Schmid, and Chen Sun. Learning temporal dynamics from cycles in narrated video. ar Xiv preprint ar Xiv:2101.02337, 2021. [33] Francis Ferraro, Nasrin Mostafazadeh, Ishan Misra, Aishwarya Agrawal, Jacob Devlin, Ross Girshick, Xiaodong He, Pushmeet Kohli, Dhruv Batra, C Lawrence Zitnick, et al. Visual storytelling. ar Xiv preprint ar Xiv:1604.03968, 2016. [34] Chelsea Finn, Ian Goodfellow, and Sergey Levine. Unsupervised learning for physical interaction through video prediction. In Proceedings of the 30th International Conference on Neural Information Processing Systems, pages 64 72, 2016. [35] David F Fouhey, Wei-cheng Kuo, Alexei A Efros, and Jitendra Malik. From lifestyle vlogs to everyday interactions. In CVPR, 2018. [36] Zhe Gan, Yen-Chun Chen, Linjie Li, Chen Zhu, Yu Cheng, and Jingjing Liu. Largescale adversarial training for vision-and-language representation learning. ar Xiv preprint ar Xiv:2006.06195, 2020. [37] Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna Wallach, Hal Daumeé III, and Kate Crawford. Datasheets for datasets. ar Xiv preprint ar Xiv:1803.09010, 2018. [38] Franklin D Gilliam Jr, Shanto Iyengar, Adam Simon, and Oliver Wright. Crime in black and white: The violent, scary world of local news. Harvard International Journal of press/politics, 1(3):6 23, 1996. [39] Jonathan Gordon and Benjamin Van Durme. Reporting bias and knowledge acquisition. In Proceedings of the 2013 workshop on Automated knowledge base construction, pages 25 30. ACM, 2013. [40] Ben Green. Good isn t good enough. In Proceedings of the AI for Social Good workshop at Neur IPS, 2019. [41] Herbert P Grice. Logic and conversation. In Speech acts, pages 41 58. Brill, 1975. [42] Donna Haraway. Situated knowledges: The science question in feminism and the privilege of partial perspective. Feminist studies, 14(3):575 599, 1988. [43] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [44] Don Heider. White news: Why local news programs don t cover people of color. Routledge, 2014. [45] Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem, and Juan Carlos Niebles. Activitynet: A large-scale video benchmark for human activity understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 961 970, 2015. [46] Jack Hessel, Bo Pang, Zhenhai Zhu, and Radu Soricut. A case study on combining ASR and visual features for generating instructional video captions. In Co NLL, November 2019. [47] Jack Hessel, Zhenhai Zhu, Bo Pang, and Radu Soricut. Beyond instructional videos: Probing for more diverse visual-textual grounding on youtube. In EMNLP, 2020. [48] Ari Holtzman, Jan Buys, Maxwell Forbes, and Yejin Choi. The curious case of neural text degeneration. ar Xiv preprint ar Xiv:1904.09751, 2019. [49] De-An Huang, Joseph J. Lim, Li Fei-Fei, and Juan Carlos Niebles. Unsupervised visuallinguistic reference resolution in instructional videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017. [50] Ting-Hao Kenneth Huang, Francis Ferraro, Nasrin Mostafazadeh, Ishan Misra, Aishwarya Agrawal, Jacob Devlin, Ross Girshick, Xiaodong He, Pushmeet Kohli, Dhruv Batra, C. Lawrence Zitnick, Devi Parikh, Lucy Vanderwende, Michel Galley, and Margaret Mitchell. Visual storytelling. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 1233 1239, San Diego, California, June 2016. Association for Computational Linguistics. doi: 10.18653/v1/N16-1147. URL https://www.aclweb.org/anthology/N16-1147. [51] Sarthak Jain and Byron C Wallace. Attention is not explanation. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 3543 3556, 2019. [52] Yunseok Jang, Yale Song, Youngjae Yu, Youngjin Kim, and Gunhee Kim. Tgif-qa: Toward spatio-temporal reasoning in visual question answering. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017). Honolulu, Hawaii, pages 2680 8, 2017. [53] Huaizu Jiang, Ishan Misra, Marcus Rohrbach, Erik Learned-Miller, and Xinlei Chen. In defense of grid features for visual question answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10267 10276, 2020. [54] Mandar Joshi, Danqi Chen, Yinhan Liu, Daniel S Weld, Luke Zettlemoyer, and Omer Levy. Spanbert: Improving pre-training by representing and predicting spans. Transactions of the Association for Computational Linguistics, 8:64 77, 2020. [55] Ruogu Kang, Laura Dabbish, Nathaniel Fruchter, and Sara Kiesler. my data just goes everywhere: user mental models of the internet and implications for privacy and security. In Eleventh Symposium On Usable Privacy and Security ({SOUPS} 2015), pages 39 52, 2015. [56] Seonhoon Kim, Seohyeong Jeong, Eun-Byul Kim, Inho Kang, and Nojun Kwak. Selfsupervised pre-training and contrastive representation learning for multiple-choice video qa. Ar Xiv, abs/2009.08043, 2020. [57] Wonjae Kim, Bokyung Son, and Ildoo Kim. Vilt: Vision-and-language transformer without convolution or region supervision. ar Xiv preprint ar Xiv:2102.03334, 2021. [58] Kris M Kitani, Brian D Ziebart, James Andrew Bagnell, and Martial Hebert. Activity forecasting. In European Conference on Computer Vision, pages 201 214. Springer, 2012. [59] Ranjay Krishna, Kenji Hata, Frederic Ren, Li Fei-Fei, and Juan Carlos Niebles. Dense Captioning Events in Videos. In International Conference on Computer Vision (ICCV), 2017. [60] Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen, Yannis Kalantidis, Li-Jia Li, David A Shamma, et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. International Journal of Computer Vision, 123(1):32 73, 2017. [61] H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre. HMDB: a large video database for human motion recognition. In Proceedings of the International Conference on Computer Vision (ICCV), 2011. [62] Hilde Kuehne, Ahsan Iqbal, Alexander Richard, and Juergen Gall. Mining youtube-a dataset for learning fine-grained action concepts from webly supervised video data. ar Xiv preprint ar Xiv:1906.01012, 2019. [63] Harold W Kuhn. The hungarian method for the assignment problem. Naval research logistics quarterly, 2(1-2):83 97, 1955. [64] Jie Lei, Licheng Yu, Mohit Bansal, and Tamara L Berg. Tvqa: Localized, compositional video question answering. In EMNLP, 2018. [65] Jie Lei, Licheng Yu, Tamara L Berg, and Mohit Bansal. Tvqa+: Spatio-temporal grounding for video question answering. In Tech Report, ar Xiv, 2019. [66] Jie Lei, Licheng Yu, Tamara L Berg, and Mohit Bansal. What is more likely to happen next? video-and-language future event prediction. ar Xiv preprint ar Xiv:2010.07999, 2020. [67] Jie Lei, Linjie Li, Luowei Zhou, Zhe Gan, Tamara L Berg, Mohit Bansal, and Jingjing Liu. Less is more: Clipbert for video-and-language learning via sparse sampling. ar Xiv preprint ar Xiv:2102.06183, 2021. [68] Gen Li, Nan Duan, Yuejian Fang, Ming Gong, Daxin Jiang, and Ming Zhou. Unicoder-vl: A universal encoder for vision and language by cross-modal pre-training. In AAAI, pages 11336 11344, 2020. [69] Liunian Harold Li, Mark Yatskar, Da Yin, Cho-Jui Hsieh, and Kai-Wei Chang. Visualbert: A simple and performant baseline for vision and language. ar Xiv preprint ar Xiv:1908.03557, 2019. [70] Yingwei Li, Yi Li, and Nuno Vasconcelos. Resound: Towards action recognition without representation bias. In Proceedings of the European Conference on Computer Vision (ECCV), pages 513 528, 2018. [71] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft COCO: Common objects in context. In European conference on computer vision, pages 740 755. Springer, 2014. [72] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining approach. ar Xiv preprint ar Xiv:1907.11692, 2019. [73] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. [74] Cewu Lu, Ranjay Krishna, Michael Bernstein, and Li Fei-Fei. Visual relationship detection with language priors. In European Conference on Computer Vision, 2016. [75] Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. Vi LBERT: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. In Advances in Neural Information Processing Systems, pages 13 23, 2019. [76] Tegan Maharaj, Nicolas Ballas, Anna Rohrbach, Aaron C Courville, and Christopher Joseph Pal. A dataset and exploration of models for understanding video data through fillin-the-blank question-answering. In Computer Vision and Pattern Recognition (CVPR), 2017. URL http://openaccess.thecvf.com/content_cvpr_2017/papers/Maharaj_ A_Dataset_and_CVPR_2017_paper.pdf. [77] Jonathan Malmaud, Jonathan Huang, Vivek Rathod, Nick Johnston, Andrew Rabinovich, and Kevin Murphy. What s cookin ? interpreting cooking videos using text, speech and vision. In NAACL, 2015. [78] Alice E Marwick and danah boyd. Networked privacy: How teenagers negotiate context in social media. New media & society, 16(7):1051 1067, 2014. [79] Andrew Kachites Mc Callum. Mallet: A machine learning for language toolkit. 2002. URL http://mallet.cs.umass.edu. [80] Antoine Miech, Dimitri Zhukov, Jean-Baptiste Alayrac, Makarand Tapaswi, Ivan Laptev, and Josef Sivic. How To100M: Learning a Text-Video Embedding by Watching Hundred Million Narrated Video Clips. In ICCV, 2019. [81] Antoine Miech, Jean-Baptiste Alayrac, Lucas Smaira, Ivan Laptev, Josef Sivic, and Andrew Zisserman. End-to-end learning of visual representations from uncurated instructional videos. In CVPR, 2020. [82] Ishan Misra, C Lawrence Zitnick, and Martial Hebert. Shuffle and learn: unsupervised learning using temporal order verification. In European Conference on Computer Vision, pages 527 544. Springer, 2016. [83] Heather Molyneaux, Susan O Donnell, Kerri Gibson, Janice Singer, et al. Exploring the gender divide on youtube: An analysis of the creation and reception of vlogs. American Communication Journal, 10(2):1 14, 2008. [84] Yasufumi Moriya, Ramon Sanabria, Florian Metze, and Gareth JF Jones. Grounding object detections with transcriptions. ar Xiv preprint ar Xiv:1906.06147, 2019. [85] Meinard Müller. Dynamic time warping. Information retrieval for music and motion, pages 69 84, 2007. [86] Shruti Palaskar, Jindrich Libovick y, Spandana Gella, and Florian Metze. Multimodal abstractive summarization for how2 videos. ar Xiv preprint ar Xiv:1906.07901, 2019. [87] Alessandro Prest, Christian Leistner, Javier Civera, Cordelia Schmid, and Vittorio Ferrari. Learning object class detectors from weakly annotated video. In 2012 IEEE Conference on Computer Vision and Pattern Recognition, pages 3282 3289. IEEE, 2012. [88] Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. Language models are unsupervised multitask learners. Technical report, Open AI, 2019. [89] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. ar Xiv preprint ar Xiv:2103.00020, 2021. [90] Manoel Horta Ribeiro, Raphael Ottoni, Robert West, Virgílio AF Almeida, and Wagner Meira Jr. Auditing radicalization pathways on youtube. In Proceedings of the 2020 conference on fairness, accountability, and transparency, pages 131 141, 2020. [91] Neil M Richards. The dangers of surveillance. Harv. L. Rev., 126:1934, 2012. [92] Anna Rohrbach, Atousa Torabi, Marcus Rohrbach, Niket Tandon, Chris Pal, Hugo Larochelle, Aaron Courville, and Bernt Schiele. Movie description. International Journal of Computer Vision, 2017. URL http://link.springer.com/article/10.1007/ s11263-016-0987-1?wt_mc=Internal.Event.1.SEM.Article Author Online First. [93] Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4510 4520, 2018. [94] Oleksandr Savsunenko. How tensorflow s tf.image.resize stole 60 days of my life. Technical report, Hacker Noon. [95] Roger C. Schank and Robert P. Abelson. Scripts, plans, and knowledge. In Proceedings of the 4th International Joint Conference on Artificial Intelligence - Volume 1, IJCAI 75, pages 151 157, San Francisco, CA, USA, 1975. Morgan Kaufmann Publishers Inc. URL http://dl.acm.org/citation.cfm?id=1624626.1624649. [96] Ozan Sener, Amir R Zamir, Silvio Savarese, and Ashutosh Saxena. Unsupervised semantic parsing of video collections. In ICCV, 2015. [97] Rico Sennrich, Barry Haddow, and Alexandra Birch. Neural machine translation of rare words with subword units. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1715 1725, 2016. [98] Sofia Serrano and Noah A Smith. Is attention interpretable? In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 2931 2951, 2019. [99] Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut. Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2556 2565, 2018. [100] Sheng Shen, Liunian Harold Li, Hao Tan, Mohit Bansal, Anna Rohrbach, Kai-Wei Chang, Zhewei Yao, and Kurt Keutzer. How much can clip benefit vision-and-language tasks? ar Xiv preprint ar Xiv:2107.06383, 2021. [101] Botian Shi, Lei Ji, Yaobo Liang, Nan Duan, Peng Chen, Zhendong Niu, and Ming Zhou. Dense procedure captioning in narrated instructional videos. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6382 6391, 2019. [102] Wei Shi and Vera Demberg. Next sentence prediction helps implicit discourse relation classification within and across domains. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 5794 5800, 2019. [103] Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. A dataset of 101 human action classes from videos in the wild. Center for Research in Computer Vision, 2(11), 2012. [104] Michael Strangelove. Watching You Tube. University of Toronto press, 2020. [105] Emma Strubell, Ananya Ganesh, and Andrew Mc Callum. Energy and policy considerations for deep learning in nlp. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 3645 3650, 2019. [106] Chen Sun, Fabien Baradel, Kevin Murphy, and Cordelia Schmid. Contrastive bidirectional transformer for temporal representation learning. ar Xiv preprint ar Xiv:1906.05743, 2019. [107] Chen Sun, Austin Myers, Carl Vondrick, Kevin Murphy, and Cordelia Schmid. Video BERT: A joint model for video and language representation learning. In ICCV, 2019. [108] Hao Tan and Mohit Bansal. LXMERT: Learning cross-modality encoder representations from transformers. In EMNLP, 2019. [109] Zineng Tang, Jie Lei, and Mohit Bansal. Decembert: Learning from noisy instructional videos via dense captions and entropy minimization. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2415 2426, 2021. [110] Atousa Torabi, Niket Tandon, and Leon Sigal. Learning language-visual embedding for movie understanding with natural-language. ar Xiv preprint, 2016. URL http://arxiv.org/pdf/ 1609.08124v1.pdf. [111] Antonio Torralba and Alexei A Efros. Unbiased look at dataset bias. In Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, pages 1521 1528. IEEE, 2011. [112] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in neural information processing systems, pages 5998 6008, 2017. [113] Jacob Walker, Carl Doersch, Abhinav Gupta, and Martial Hebert. An uncertain future: Forecasting from static images using variational autoencoders. In European Conference on Computer Vision, pages 835 851. Springer, 2016. [114] Zeerak Waseem, Smarika Lulz, Joachim Bingel, and Isabelle Augenstein. Disembodied machine learning: On the illusion of objectivity in nlp. ar Xiv preprint ar Xiv:2101.11974, 2021. [115] Donglai Wei, Joseph J Lim, Andrew Zisserman, and William T Freeman. Learning and using the arrow of time. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8052 8060, 2018. [116] Sarah Wiegreffe and Yuval Pinter. Attention is not not explanation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 11 20, 2019. [117] Dejing Xu, Zhou Zhao, Jun Xiao, Fei Wu, Hanwang Zhang, Xiangnan He, and Yueting Zhuang. Video question answering via gradually refined attention over appearance and motion. In Proceedings of the 25th ACM international conference on Multimedia, pages 1645 1653, 2017. [118] Antoine Yang, Antoine Miech, Josef Sivic, Ivan Laptev, and Cordelia Schmid. Just ask: Learning to answer questions from millions of narrated videos. ar Xiv preprint ar Xiv:2012.00451, 2020. [119] Fei Yu, Jiji Tang, Weichong Yin, Yu Sun, Hao Tian, Hua Wu, and Haifeng Wang. Ernie-vil: Knowledge enhanced vision-language representations through scene graph. ar Xiv preprint ar Xiv:2006.16934, 2020. [120] Shoou-I Yu, Lu Jiang, and Alexander Hauptmann. Instructional videos for unsupervised harvesting and learning of action examples. In ACM MM, 2014. [121] Youngjae Yu, Jongseok Kim, and Gunhee Kim. A joint sequence fusion model for video question answering and retrieval. In Proceedings of the European Conference on Computer Vision (ECCV), pages 471 487, 2018. [122] Zhou Yu, Dejing Xu, Jun Yu, Ting Yu, Zhou Zhao, Yueting Zhuang, and Dacheng Tao. Activity Net-QA: a dataset for understanding complex web videos via question answering. In AAAI, pages 9127 9134, 2019. [123] Rowan Zellers, Yonatan Bisk, Ali Farhadi, and Yejin Choi. From recognition to cognition: Visual commonsense reasoning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6720 6731, 2019. [124] Rowan Zellers, Ari Holtzman, Hannah Rashkin, Yonatan Bisk, Ali Farhadi, Franziska Roesner, and Yejin Choi. Defending against neural fake news. In Advances in Neural Information Processing Systems 32, 2019. [125] Pengchuan Zhang, Xiujun Li, Xiaowei Hu, Jianwei Yang, Lei Zhang, Lijuan Wang, Yejin Choi, and Jianfeng Gao. Vinvl: Revisiting visual representations in vision-language models. ar Xiv preprint ar Xiv:2101.00529, 2021. [126] Yuhao Zhang, Hang Jiang, Yasuhide Miura, Christopher D Manning, and Curtis P Langlotz. Contrastive learning of medical visual representations from paired images and text. ar Xiv preprint ar Xiv:2010.00747, 2020. [127] Linchao Zhu and Yi Yang. Act BERT: Learning global-local video-text representations. In CVPR, 2020. [128] Shoshana Zuboff. Big other: surveillance capitalism and the prospects of an information civilization. Journal of Information Technology, 30(1):75 89, 2015.