# learning_stateaware_visual_representations_from_audible_interactions__9d0b20a6.pdf Learning State-Aware Visual Representations from Audible Interactions Himangi Mittal1, Pedro Morgado1, Unnat Jain2, Abhinav Gupta1 1Carnegie Mellon University 2Meta AI Research We propose a self-supervised algorithm to learn representations from egocentric video data. Recently, significant efforts have been made to capture humans interacting with their own environments as they go about their daily activities. In result, several large egocentric datasets of interaction-rich multi-modal data have emerged. However, learning representations from videos can be challenging. First, given the uncurated nature of long-form continuous videos, learning effective representations require focusing on moments in time when interactions take place. Second, visual representations of daily activities should be sensitive to changes in the state of the environment. However, current successful multimodal learning frameworks encourage representation invariance over time. To address these challenges, we leverage audio signals to identify moments of likely interactions which are conducive to better learning. We also propose a novel selfsupervised objective that learns from audible state changes caused by interactions. We validate these contributions extensively on two large-scale egocentric datasets, EPIC-Kitchens-100 and the recently released Ego4D, and show improvements on several downstream tasks, including action recognition, long-term action anticipation, and object state change classification. Code and pretrained model are available here: https://github.com/Himangi M/Rep LAI 1 Introduction Recent successes in self-supervised learning (SSL) [48, 10, 31, 28] has brought into question the need for human annotations in order to learn strong visual representations. However, current approaches are bottlenecked by the lack of rich data they learn from static images which lack temporal information and restrict the ability to learn object deformations and state changes. It is clear that we need videos to learn rich representations in self-supervised manner. Learning representations from videos is however quite challenging. The first challenge is choosing the right SSL loss. Approaches such as [67, 54] have attempted to learn representations that are invariant to object deformations/viewpoints. However, many downstream tasks require representations that are sensitive to these deformations. Another alternative has been to use the multi-modal data [3, 43, 57] and learn representations via audio. But again most of these approaches seek to align audio and visual features in a common space, leading to invariant representations as well. The second challenge is dealing with the fact that current video-based SSL approaches exploit the curated nature of video datasets, such as Kinetics [9]. These approaches are designed to leverage carefully selected clips, displaying a single action or object interaction. This is in contrast to the predominantly untrimmed real-world data characteristic of large egocentric datasets of daily activities. Here, unlike action centric datasets, the most interesting or interaction-rich clips have NOT been carefully selected by human annotators. Thus, learning from untrimmed video poses a major challenge, as a significant portion of the data does not focus on the concepts we want to learn. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Before moment with NO interaction Before moment of interaction After moment with NO interaction After moment of interaction No State Change State Changes from a bok choy to chopped bok choy Moment of Interaction (Mo I) Random Moment Video Spectrogram Figure 1: Moments of audible interactions in a long, untrimmed video. Random moments in time are likely to contain NO interactions, e.g., timestamp shown by the gray box. Since no interactions occur, no changes in the before and after states are observed. By sampling from moments of interaction (Mo I), as shown in the red box, we can learn representations that are sensitive to the change in the visual state caused by interactions. In this work, we ask the question, Can we learn meaningful representations from interaction-rich, multi-modal streams of egocentric data? Learning from continuous streams of data requires focusing on the right moments when the actual interactions are likely to occur. Consider, for example, the acts of opening a fridge or placing a pan on the stove. Actions like these create clear and consistent sound signatures due to the physical interaction between objects. These moments can be easily detected from audio alone and can be used to target training on interesting portions of the untrimmed videos. We show that even a simple spectrogram-based handcrafted detector is sufficient to identify interesting moments in time, and that representation learning benefits substantially from using them to sample training clips. But what should the loss be? Prior work on audio-visual correspondence (AVC) [15, 4, 43] uses the natural co-occurrence of sounds and the visual manifestations of their sources as the source of supervision. However, since the AVC objective still favors invariance, the learned representations are not informative of the changes that happen over time (e.g., representations that can distinguish between closed and opened fridge, or vegetables before and after chopping them). To better capture state changes, we introduce a novel audio-visual self-supervised objective, in which audio representations at key moments in time are required to be informative of the change in the corresponding visual representations over time. The intuition behind this objective is that transitions between object states are often marked by characteristic sounds. Thus, models optimized under this objective would associate the distinct sounds not only with the objects themselves (as accomplished with AVC), but also with the transition between two different states of the object. To this end, we introduce Rep LAI Representation Learning from Audible Interactions, a selfsupervised algorithm for representation learning from videos of audible interactions. Rep LAI uses the audio signals in two unique ways: (1) to identify moments in time that are conducive to better self-supervised learning and (2) to learn representations that focus on the visual state changes caused by audible interactions. We validate these contributions extensively on two egocentric datasets, EPIC-Kitchens-100 [14] and the recently released Ego4D [27], where we demonstrate the benefits of Rep LAI for several downstream tasks, including action recognition, long term action anticipation, and object state change classification. 2 Related Work Self-supervised learning. Self-supervised learning methods operate on an unlabeled dataset by explicitly defining pretext tasks such as solving jigsaw puzzle [47], patch location prediction [16], inpainting [50], and image rotation [25] prediction. Following these, the next wave of self-supervised methods has been based on contrastive learning that learns representations with the help of data augmentation and instance discrimination [10, 28, 48, 31, 8]. These methods have shown rapid progress in self-supervised learning for images. While these approaches explore the spatial information of images, Rep LAI leverages the temporal information of videos. Video representation learning. Relevant to our proposed approach is self-supervised representation learning for videos where the spatiotemporal pretext tasks are designed such as temporal order prediction [40, 70, 35, 69], predicting motion and appearance statistics [65], pace prediction [66], temporal cycle consistency [18, 68], and video colorization [64]. Contrastive learning has also been widely adopted in the domain of video [55, 29, 32, 57, 71, 30, 22] with impressive results on action recognition tasks. These methods however learn representations that are invariant to spatio-temporal augmentations, such as temporal jittering, and thus are incapable of representing object state changes. Closer to the objective of Rep LAI, we include relevant literature on audio-visual representation learning from videos, where the audio stream is additionally utilized. Audio-visual representation learning. Learning without additional supervision has also been explored in the context of the audio modality with the help of audio-visual correspondence (AVC) [4, 5]. As stated simply, AVC is the binary classification task of predicting if a video clip and a short audio clip correspond with each other or not (details in Sec. 3.4). Similar tasks like temporal synchronization [36, 49] between audio and video, audio classification [6, 3, 11], spatial alignment prediction between audio and 360-degree videos [41], optimal combination of self-supervised tasks [52] have been shown beneficial for learning effective multi-modal video representations. Other works explore contrastive learning for both audio and video modality [43, 51, 42] as a cross-modal instance discrimination task. Fine-grained video understanding. Real-world videos are often untrimmed in nature and have multiple actions in a single video. Along this line, fine-grained analysis has been studied for videos in the form of a query-response temporal attention mechanism [72], bi-directional RNN s[58], and semi-supervised learning problem [17]. While these works only utilize the visual modality, another line of work has also explored multi-modal fine-grained video understanding as a transformer-based model [34], by exploiting the correspondence between modalities [44], or by exploring how to best combine multiple modalities - audio, visual, and language [2]. In our work, we try to conduct fine-grained video understanding in a self-supervised manner. Egocentric datasets. Egocentric datasets offer new opportunities to learn from a first-person point of view, where the world is seen through the eyes of an agent. Many egocentric datasets have been developed such as Epic-kitchens [13, 14] which consist of daily activities performed in a kitchen environment, Activities of Daily Living [53], UT Ego [37, 60], the Disney Dataset [20], and the recently released large-scale Ego4D dataset [27] which consists of day-to-day life activities in multiple scenarios such as household, outdoor spaces, workplace, etc. Multiple challenges and downstream tasks have explored for egocentric datasets like action recognition [34, 33, 38], action localization [56], action anticipation [26, 59, 39, 1, 23], human-object interactions [45, 12, 7], parsing social interactions [46], and domain adaptation [44]. In our work, we evaluate the efficiency of the representations learned by our self-supervised approach on the EPIC-Kitchens-100 and Ego4D datasets over multiple downstream tasks. In this section, we detail our approach to learn audio-visual representations from and for interactionrich egocentric data in a self-supervised manner, i.e., without relying on human annotated labels. Sec. 3.1 provides an overview of Rep LAI and motivates the two key contributions of this work identifying moments of interaction (Mo I) and learning from audible visual state changes . Sec. 3.2 details the proposed approach for Mo I detection and section Sec. 3.3 explains the proposed selfsupervised objective for learning state-aware representations. Sec. 3.4 explains the objective of audio-visual correspondence learning used to train Rep LAI. Sec. 3.5 brings both objectives together and includes necessary details for reproducibility. 3.1 Overview Given a dataset D = {(vi, ai)N i=1} containing N long (untrimmed) audio-visual streams, our goal is to learn visual and audio encoders, denoted f V and f A, that can effectively represent egocentric data. An overview of the proposed approach is depicted in Fig. 2. For each sample (v, a) D, we search for moments of interaction (Mo I) using the audio stream, and extract short audio and visual clips around these Mo I. These trimmed clips are then encoded into a vectorized representation using f V and f A. The whole system is trained to optimize two self-supervised losses an audio-visual Untrimmed Audio-Visual Pairs Trim Untrimmed audio stream 𝑎 Untrimmed visual stream 𝑣 ℒ"!# 𝑡 Mo I Detection Audio-Visual Encoders Clip Extraction Self-Supervised , Visual state Figure 2: Overview of Rep LAI. Rep LAI seeks to learn audio and visual encoders (f A and f V ) by (1) detecting and focusing training data on moments of interaction (Mo I) present in untrimmed videos and (2) solving a combination of two tasks audio-visual correspondence (AVC) and audio identifiable state changes (ASt C). correspondence loss LAVC, and a novel self-supervised loss that learns from audible state changes LASt C. Why detect moments of interaction (Mo I)? Untrimmed video of daily activities often contains long periods without interactions, which aren t useful for training. Instead, we search for moments in time that are more likely to contain interactions which we refer to as moments of interaction (Mo I). Why learn from audible state changes? Visual representations of daily activities should be informative of the state of the environment and/or objects being interacted with. Moreover, changes in the environment are usually caused by physical interactions, which produce distinct sound signatures. We hypothesize that state-aware representations can be obtained by learning to associate audio with the change of visual representation during a moment of interaction. 3.2 Audio-driven detection of moments of interaction Audio signals are particularly informative of moments of interaction. To complete day-to-day activities, we physically interact with objects in our environments. These interactions typically produce distinct audio patterns - short bursts of energy that span all frequencies. This is illustrated in Fig. 1, where we visualize the untrimmed visual and audio data of a person performing a series of actions in the kitchen. The audio data is represented as a log mel spectrogram, where the x-axis represents time and y-axis the audio frequency in log-scale. As can be seen, moments of interaction appear in the spectrogram as vertical edges, which can be easily detected. Once detected, short clips around the moments of interaction are collected into a dataset DMo I, and used for training. The remaining question is how to locate the timestamp of such vertical edges? Intuitively, we do this by finding robust local maxima in the total energy (summed over all frequencies) of the spectrogram. Concretely, let M(t, ω) be the value of the log mel spectrogram of an audio clip at time t and frequency ω. To remove the influence of background noise and overall audio intensity/volume, we compute the z-score normalization of the spectrogram for each frequency independently M(t, ω) = s(t,ω) µω σω+ϵ , where ϵ is small constant for numerical stability. Here, µω and σω are the mean and standard deviation of M(t, ω) over time, respectively.1 Next, we define moments of interaction as the set of timestamps which are local maxima of P ω s(t, ω) (or peaks for short). Moreover, to avoid weak local maxima that may be caused by the noisy nature of audio signals, we ignore peaks with small prominence (lower than 1)2. For further robustness, when multiple close peaks are found (less than 50ms apart), only the highest prominence peak is kept. 3.3 Learning from audible state changes Physical interactions often cause both state changes in the environment and distinct audio signals. To leverage this natural co-occurrence, we propose a self-supervised task that seeks to associate the audio with changes in the visual state during a moment of interaction. 1Specifically, µω = Et[M(t, ω)], σ2 ω = Et[(M(t, ω) µω)2], and ϵ = 1e 5. 2The prominence of a peak is defined as the difference between the peak value and the minimum value in a small window around it. Learning from Audible State Changes Trim Audio 𝑎 Audio-Visual Encoders Clip Extraction ASt C Projection Heads Loss Function (a) The proposed ASt C formulation (Sec. 3.3) Audio-Visual Correspondence Trim Audio 𝑎 Audio-Visual Encoders Clip Extraction Loss Function AVC Projection (b) Audio-visual correspondence (Sec. 3.4) Figure 3: Rep LAI architecture for ASt C and AVC tasks. In both cases, short clips are extracted around moments of interaction (Mo I). (a). In ASt C, the representation of a visual state change vt is matched to the corresponding audio at. (b). AVC seeks to associate audio at with the corresponding visual clips vt. The proposed task is optimized by minimizing a loss with two negative log-likelihood terms to: (1) increase the probability of associating the audio with the visual state change in the forward (i.e. correct) direction, (2) decrease the probability of associating the audio with the visual state change in the backward (i.e. incorrect) direction. Consider, for example, the interaction of closing a fridge door . To optimize for this task, the audio of closing the door should be (1) similar to the visual transition opened door closed door and (2) dissimilar to the (backwards) transition closed open. This encourages learning of representations that are informative of object states, making them useful for a variety of egocentric tasks. Specifically, the audible state change (ASt C) loss is defined as LASt C = Evt,at DMo I log pfrwd(vt, at) log 1 pbkwd(vt, at) . (1) The probabilities (pfrwd, pbkwd) are computed from cross-modal similarities pfrwd(vt, at) = σ sim vfrwd t , at /τ , (2) pbkwd(vt, at) = σ sim vbkwd t , at /τ , (3) where τ = 0.2 is a temperature hyper-parameter, and σ denotes the sigmoid function. For better readability, we absorb the notations for the audio projection MLP head h ASt C A and the state change projection MLP head h ASt C V within sim( , ), but their usage is clearly illustrated in Fig. 3a. Audio representations (at) are obtained by encoding the trimmed audio clips at via the audio encoder f A (shared across all objectives). As explained above, at is further projected via h ASt C A to a space where similarity to visual state changes is enforced. State change representations ( vfrwd t , vbkwd t ) are computed by considering two non-overlapping visual clips for each moment of interaction t, at timestamps t δ and t + δ. The two clips, vt δ and vt+δ, are encoded via the visual encoder f V (shared across all tasks) and a projection MLP head h ASt C V (specific to the ASt C task). Specifically, we represent forward and backward state changes as vfrwd t = h ASt C V f V (vt+δ) h ASt C V f V (vt δ), (4) vbkwd t = h ASt C V f V (vt δ) h ASt C V f V (vt+δ). (5) In summary, optimizing the loss of Eq. 1 not only requires the audio representation at to be aligned with representation of the visual change vfrwd t that took place, but also to be different from the hypothetical backward state change vbkwd t . 3.4 Learning from audio-visual correspondences [15, 4, 43] Audio-visual correspondence (AVC) is a well-studied self-supervised methodology for learning unimodal audio and visual encoders. The key idea is to bring visual and audio clips into a common feature space, where the representations of audio-visual pairs are aligned. Note that AVC differs from the proposed ASt C task, as AVC seeks to associate the audio at with the corresponding visual clips vt, as opposed to the change in visual state vt. As a result, visual representations learned through AVC are biased towards static concepts, while those learned through ASt C are more sensitive to dynamic concepts. Since both types of representations can be useful for egocentric tasks, we further train the visual and audio encoders, f V and f A, for the AVC task. Specifically, consider a dataset of audio-visual pairs (vi, ai) with representations vi = f V (vi) and ai = f A(ai). In particular, we let (vi, ai) be short clips extracted from sample i around one of the detected moments of interest. Then, following [43, 61], audio-visual correspondence is established by minimizing a cross-modal Info NCE loss of the form LAVC = Evi,ai D log esim(vi,ai)/τ P j esim(vi,aj)/τ log esim(vi,ai)/τ P j esim(vj,ai)/τ where τ = 0.07 is a temperature hyper-parameter and sim( , ) denotes the cosine similarity. Both terms in Eq. 6 help bring vi and ai (i.e. the positives) together. The key difference is whether the negative set is composed of audio representations aj or visual representations vj where j = i For readability of Eq. 6, we once again absorb the notation for the audio and visual projection MLP heads (h AVC A and h AVC V ) within sim( , ), and illustrate their usage in Fig. 3b. Fig. 3b also shows that we apply the AVC loss twice to associate both the visual clips (extracted slightly before and after the moment of interaction t) to the corresponding audio. 3.5 Training The audio-visual representation models f A and f V are trained to minimize both AVC and ASt C losses L = αLAVC + (1 α)LASt C (7) where α is a weighting hyper-parameter between the two terms. While we experimented with different values of α, we found that equal weighting produced best results. Implementation details. We follow prior work on audio visual correspondence [43], and use an R(2+1)D video encoder [62] with depth 18 and a 10-layer 2D CNN as the audio encoder. Two video clips are extracted around moments of interaction at a frame rate of 16 FPS each with a duration of 0.5s, and separated by a gap of 0.2s. Video clips are augmented by random resizing, cropping, and horizontal flipping resulting in clips of 8 frames at a resolution of 112 112,. As for the audio, we extract clips of 2s at 44.1k Hz and downsample them to 16k Hz. If the audio is stereo, we average the two waveforms to downgrade to mono, and then convert the mono signal to a log mel spectrogram with 80 frequency bands and 128 temporal frames. Models are trained with stochastic gradient descent for 100 epochs with a batch size of 128 trained over 4 GTX 1080 Ti GPUs, a learning rate of 0.005 and a momentum of 0.9. For Ego4D, we use a batch size of 512 trained over 8 RTX 2080 Ti GPUs with a learning rate of 0.05. The two loss terms in Eq. 7 are equally weighted with α = 0.5. 4 Experiments In this section, we demonstrate the benefits of identifying moments of interaction and learning state-aware representations through an audible state-change objective. We also show that, while large scale audio-visual correspondence (AVC) is beneficial, it is not sufficient to learn state-aware representations required for egocentric tasks. The setup used for our experiments is described in Sec. 4.1. Results and discussion of main takeaways are presented in Sec. 4.2. 4.1 Experimental Setup Datasets. We evaluate on two egocentric datasets: EPIC-Kitchens-100 [14] and Ego4D [27]. EPICKitchens-100 contains 100 hours of activities in the kitchen. Ego4D contains 3670 hours of egocentric video covering daily activities in the home, workplace, social settings, etc. For experiments on Ego4D, we use all videos from the Forecasting and Hand-Object interaction subsets. Baselines and ablations. We consider various baselines as well as ablated versions of Rep LAI. Random represents an untrained (randomly initialized) model. AVID [43] and XDC [3] are two state-of-the-art models pre-trained on 2M audio-visual pairs from Audio Set [24] that only leverage audio-visual correspondence. For the full method Rep LAI, we initialize the model weights from AVID before training on moments of interaction to minimize both AVC and state change loss, ASt C. Method LAVC LASt C Mo I Sampling AVC Pretraining [43] Top1 Acc Top5 Acc Verb Noun Verb Noun (1) Random 20.38 4.96 64.75 19.83 (2) XDC [3] 24.46 6.75 68.04 22.71 (3) AVID [43] 26.62 9.00 69.79 25.50 (4) Rep LAI w/o AVC 29.92 10.46 70.58 29.00 (5) Rep LAI w/o ASt C 29.29 9.67 73.33 29.54 (6) Rep LAI w/o Mo I 28.71 8.33 73.17 27.29 (7) Rep LAI (scratch) 25.75 8.12 71.25 27.29 (8) Rep LAI 31.71 11.25 73.54 30.54 Table 1: Action recognition on EPIC-Kitchens-100. Top1 and top5 accuracy (%) is reported. : Higher is better. We also evaluate our method trained without AVID initialization (Rep LAI from scratch), trained with only AVC (Rep LAI w/o ASt C ), only state change losses (Rep LAI w/o AVC ), and trained on random moments in time (Rep LAI w/o Mo I). Finally, we compare our approach with the fully supervised methods presented in Ego4D [27]. Downstream tasks. After self-supervised pre-training, the models are evaluated on a range of egocentric downstream tasks. This is done, as is standard, by appending a task specific decoder to the backbone model, and training the decoder on a small annotated dataset. The tasks are: Video action recognition (AR) on EPIC-Kitchens-100 and Ego4D. Given a short video clip, the task is to classify the verb and noun of the action taking place. This is done using two separate linear classifiers trained for this task. We report the top-1 and top-5 accuracies, following [14] (Tab. 1) and [27] (Tab. 2). We also evaluate on the unseen participants, head classes, and tail classes of EPIC-Kitchens-100 in Tab. 3. Through this task, we assess the efficacy of the spatial-temporal representations learned by the model in differentiating among different verbs and nouns. Long-term action anticipation (LTA) on Ego4D. Given a video, the task is to predict the camera wearer s future sequence of actions. For this task, the model is first presented with 4 consecutive clips of 2s, which are encoded using our visual backbone f V . Following [27], the representations are concatenated and fed to 20 separate linear classification heads to predict the following 20 actions. Performance is measured using the edit distance metric ED@(Z=20) [27].3 With the help of this task, we can evaluate if the representations learned by the model can be employed for long-horizon planning where the actions can change and may be of arbitrary duration. Results are reported in Tab. 2. State change classification (St CC) on Ego4D. Given a video clip, the task is to classify if an object undergoes a state change or not. The video clip is encoded by f V and a state change classification head is used which performs global average pooling on the entire feature tensor and is followed by a classification layer. Performance is measured through the State Change Classification Accuracy (%), and reported in Tab. 2. This task is ideal for assessing the ability of the model in understanding the temporal change happening in the state of an object. 4.2 Discussion of results As can be seen in Tab. 1 and Tab. 2, Rep LAI outperforms all other methods across all downstream tasks. Overall, this can be attributed to its ability to focus on interactions, both by detecting when they occur and by learning representations that are sensitive to interactions. A closer analysis of these results reveals several insights that we discuss next. Rep LAI enhances large-scale AVC driven approaches. Prior work on self-supervised audio-visual learning has shown strong audio-visual representations for action recognition [43, 42]. One question that we seek to answer is, how useful these are representations to egocentric tasks and what are their limitations? To answer this question, we compare our model trained from scratch, Rep LAI (Scratch), 3Edit distance measures the minimum number of operations required to convert the predicted sequence of actions to ground truth. To account for multi-modality of future actions, it also allows the model to make Z = 20 predictions, and only accounts for the best prediction. St CC AR LTA Method LAVC LASt C Mo I AVC Pretraining [43] Acc Top1 Acc ED@(Z=20) Verb Noun Verb Noun (S1) I3D-Res Net-50 [9, 27] NA NA NA NA 68.70 - - - - (S2) Slow Fast [21, 27] NA NA NA NA - - - 0.747 0.808 (S3) MVi T [19, 27] NA NA NA NA - - - 0.707 0.901 (1) Random 51.80 17.4 7.7 0.831 0.936 (2) XDC [3] 58.90 17.90 8.70 0.823 0.928 (3) AVID [43] 61.11 18.3 10.7 0.811 0.919 (4) Rep LAI w/o AVC 64.00 20.3 12.4 0.781 0.854 (5) Rep LAI w/o ASt C 63.60 21.1 13.5 0.774 0.853 (6) Rep LAI w/o Mo I 62.90 19.8 11.2 0.792 0.868 (7) Rep LAI (scratch) 66.20 22.2 14.1 0.760 0.840 (8) Rep LAI 66.30 22.5 14.7 0.755 0.834 Table 2: Performance on several downstream tasks on Ego4D. St CC: State Change Classification (%). AR: Action Recognition (%). LTA: Long-term action anticipation. : Higher is better. : Lower is better. Unseen Participants Tail Classes Head Classes Top1 Acc Top5 Acc Top1 Acc Top5 Acc Top1 Acc Top5 Acc Methods Verb Noun Verb Noun Verb Noun Verb Noun Verb Noun Verb Noun XDC [3] 24.29 6.96 67.79 23.00 15.89 4.17 44.92 9.77 24.78 6.95 72.28 24.74 AVID [43] 26.17 8.67 68.75 24.12 16.80 4.43 47.14 12.89 27.95 9.82 73.20 28.43 Rep LAI w/o AVC 28.67 9.38 72.04 27.88 18.49 5.21 47.79 12.63 30.59 10.21 73.33 30.90 Rep LAI w/o Mo I 27.71 7.92 72.08 26.88 16.80 4.04 49.74 12.76 29.36 10.65 76.33 30.41 Rep LAI 31.58 10.17 73.46 29.96 20.05 6.12 52.08 16.54 33.41 11.58 77.77 34.33 Table 3: Video action recognition (AR) accuracy (%) on EPIC-Kitchens-100 for unseen participants, head classes, and tail classes. Top1 and top5 accuracy (%) is reported. : Higher is better. with our model using the weights from AVID [43] as initialization for both the visual and audio encoders. We also compare our method to standalone AVID and XDC i.e. without further selfsupervised training. Comparing rows (2), (3) and (8) in Tab. 1 and Tab. 2, it is clear that Rep LAI enhances large-scale AVC pre-training by significant margins, leading to absolute improvements of 5% in top-1 verb accuracy on EPIC-Kitchens-100, 4.2% on Ego4D, 5.2% increase in state-change classification accuracy, and 5.6% reduction on the edit distance for long-term anticipation compared to AVID. Comparing rows (7) and (8), we also see that large-scale AVID pre-training enhances the representations learned by Rep LAI on EPIC-Kitchens-100 significantly but only marginally on Ego-4D. This is likely due to the significantly large diversity of scenes in Ego4D. Thus, while relying on large-scale audio-visual pre-training (as with AVID) can help avoid overfitting on smaller egocentric datasets, this is less critical when training on larger and more diverse data. Detecting moments of interaction (Mo I) helps representation learning. We hypothesize that to learn good representations for egocentric data of daily activities, self-supervised learning should focus on moments in time when interactions occur. To assess whether our audio-driven Mo I detection algorithm helps representation learning, we compare Rep LAI with an ablated version, Rep LAI w/o Mo I, where the model is trained on audio-visual clips extracted at random from the untrimmed videos. As can be seen by comparing rows (6) and (8) in Tab. 1 and Tab. 2, sampling clips around Mo I leads to significantly better representations for all egocentric downstream tasks that we study. Moreover, even though Rep LAI w/o Mo I trains with ASt C, it is unable to fully leverage the state change objective function without the information of moments of interactions which leads to a worse performance. This suggests that, an explicit state change objective function and sampling video clips around moments of interactions (which are likely to be aligned with the actual state changes) together provide an information-rich feedback to our model in better understanding how the state changes by an interaction and how the actions transition over time. These results also clearly show that the proposed Mo I detection procedure is able to find moments in time that are especially useful for learning representations of daily activities. We emphasize the simplicity and effectiveness of our audio-driven detector, which shows how informative audio can be when searching for moments of interaction. In the future, we believe that learning-based approaches could further enhance Mo I detection, and further improve the learned audio-visual representations. We also show several qualitative examples of detected Mo I in the supplement. Rep LAI (ours) t-SNE Before Grate After Grate Audio Spectrogram Action Timeline Moment of Interaction: Cutting carrot Moment of Interaction: Grating carrot Figure 4: t-SNE visualization of the feature representations learned by Rep LAI and AVID for a video consisting of fine-grained actions over time. For a simpler visualization, we consider all the videos belonging to a single participant. A larger spread in the t-SNE of Rep LAI indicates more distinct state-aware representations. AVC and ASt C are complementary. To assess the impact of both terms in Eq. 7, we evaluate Rep LAI trained without LAVC and without LASt C. Comparing rows (4), (5) to row (2) and row (3) in Tab. 1 and Tab. 2 shows that each term enhances the representations obtained through large-scale audio-visual pre-training (AVID). Furthermore, comparing the ablated models in rows (4) and (5) to the full model in row (8) shows that these two terms are complementary to each other. This is because the AVC and ASt C tasks encourage learning of representations with different characteristics. AVC focuses on learning visual representations that are informative of what kind of sounding objects are present in the video, while ASt C forces the model to differentiate between visual representations that occur before and after state change interactions. Rep LAI encourages state-aware representation learning. To study the representations learned by our approach for different states, we generate a t-SNE plot [63] for Rep LAI and AVID as shown in Fig. 4. For generating a simpler visualization, a small dataset is prepared consisting of all the videos corresponding to a single participant, P01, in EPIC-Kitchens-100 and split into clips of 0.5s. We can observe that there is a larger spread in the t-SNE plot for Rep LAI than AVID. A larger spread indicates that the representations of the various states are significantly different from each other and form more distant clusters as shown by Rep LAI. Whereas, if the state representations are similar to each other, they are clustered together and show lesser spread as shown by AVID. Mo I are the key moments of interactions with an object in an environment where the state is changing. AVID has no such information about the key moments and also does not have an explicit state change objective function. Therefore, it is unable to discriminate between the before and after state of an action and has less effective state-aware information in its representations. Rep LAI representation are more generalizable and robust to long-tail. To assess Rep LAI in a scenario with domain shift, we evaluate on unseen participants that were fully excluded from the pre-training of Rep LAI. Tab. 3 shows that Rep LAI significantly outperforms baselines and ablations, indicating that representation learning by our model provides much better generalization. Moreover, the verb and noun classes in EPIC-Kitchens-100 exhibit a long-tailed distribution. When further compared on head and tail classes separately in Tab. 3, we can observe that Rep LAI outperforms all other methods highlighting its higher robustness on a long-tailed distribution. Self-supervised vs supervised representation learning Tab. 2 also compares Rep LAI to fully supervised methods introduced in Ego4D [27] (rows S1, S2 and S3). We can observe that Rep LAI can also perform competitively to the fully supervised approaches when we have access to larger and more diverse data. With further focus on SSL for untrimmed datasets, SSL methods will be able to match supervised approaches, and our work takes a step towards it. 5 Conclusion In this work, we propose an audio-driven self-supervised method for learning representations of egocentric video of daily activities. We show that in order to learn strong representations for this domain, two important challenges need to be addressed. First, learning should focus on moments of interaction (Mo I). Since these moments only occur sporadically in untrimmed video data, we show that Mo I detection is an important component of representation learning in untrimmed datasets. Second, learning should focus on the consequences of interactions, i.e., changes in the state of an environment caused by agents interacting with the world. In particular, by seeking to identify visible state changes from the audio alone, we can learn representations that are potentially more aware of the state of the environment and hence, particularly useful for egocentric downstream tasks. Acknowledgements We would like to thank DARPA MCS, ONR Young Investigator and DARPA SAIL-ON for the funding. Broader impact Deep learning models are capable of learning (and sometimes even amplifying) biases existing in datasets. While several steps have been taken in datasets like Ego4D to increase geographical diversity, we would like to encourage careful consideration of ethical implications when deploying these model. While public datasets are essential to make progress on how to represent visual egocentric data, premature deployment of our models is likely have negative societal impact, as we did not check for the presence or absence of such biases. [1] Y. Abu Farha, A. Richard, and J. Gall. When will you do what?-anticipating temporal occurrences of activities. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5343 5352, 2018. 3 [2] J.-B. Alayrac, A. Recasens, R. Schneider, R. Arandjelovic, J. Ramapuram, J. De Fauw, L. Smaira, S. Dieleman, and A. Zisserman. Self-supervised multimodal versatile networks. Neur IPS, 2(6): 7, 2020. 3 [3] H. Alwassel, D. Mahajan, B. Korbar, L. Torresani, B. Ghanem, and D. Tran. Self-supervised learning by cross-modal audio-video clustering. Advances in Neural Information Processing Systems, 33:9758 9770, 2020. 1, 3, 6, 7, 8 [4] R. Arandjelovic and A. Zisserman. Look, listen and learn. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 609 617, 2017. 2, 3, 5 [5] R. Arandjelovic and A. Zisserman. Objects that sound. In Proceedings of the European conference on computer vision (ECCV), pages 435 451, 2018. 3 [6] Y. Asano, M. Patrick, C. Rupprecht, and A. Vedaldi. Labelling unlabelled videos from scratch with multi-modal self-supervision. Advances in Neural Information Processing Systems, 33: 4660 4671, 2020. 3 [7] M. Cai, K. Kitani, and Y. Sato. Understanding hand-object manipulation by modeling the contextual relationship between actions, grasp types and object attributes. ar Xiv preprint ar Xiv:1807.08254, 2018. 3 [8] M. Caron, I. Misra, J. Mairal, P. Goyal, P. Bojanowski, and A. Joulin. Unsupervised learning of visual features by contrasting cluster assignments. In Thirty-fourth Conference on Neural Information Processing Systems (Neur IPS), 2020. 2 [9] J. Carreira and A. Zisserman. Quo vadis, action recognition? a new model and the kinetics dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6299 6308, 2017. 1, 8 [10] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, pages 1597 1607. PMLR, 2020. 1, 2 [11] S.-W. Chung, J. S. Chung, and H.-G. Kang. Perfect match: Improved cross-modal embeddings for audio-visual synchronisation. In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 3965 3969. IEEE, 2019. 3 [12] D. Damen, T. Leelasawassuk, O. Haines, A. Calway, and W. W. Mayol-Cuevas. You-do, i-learn: Discovering task relevant objects and their modes of interaction from multi-user egocentric video. In BMVC, volume 2, page 3, 2014. 3 [13] D. Damen, H. Doughty, G. M. Farinella, S. Fidler, A. Furnari, E. Kazakos, D. Moltisanti, J. Munro, T. Perrett, W. Price, and M. Wray. Scaling egocentric vision: The epic-kitchens dataset. In European Conference on Computer Vision (ECCV), 2018. 3 [14] D. Damen, H. Doughty, G. M. Farinella, A. Furnari, E. Kazakos, J. Ma, D. Moltisanti, J. Munro, T. Perrett, W. Price, et al. Rescaling egocentric vision. ar Xiv preprint ar Xiv:2006.13256, 2020. 2, 3, 6, 7 [15] V. R. de Sa. Learning classification with unlabeled data. In Advances in Neural Information Processing Systems, pages 112 119. Citeseer, 1994. 2, 5 [16] C. Doersch, A. Gupta, and A. A. Efros. Unsupervised visual representation learning by context prediction. In Proceedings of the IEEE international conference on computer vision, pages 1422 1430, 2015. 2 [17] H. Doughty and C. G. Snoek. How do you do it? fine-grained action understanding with pseudo-adverbs. ar Xiv preprint ar Xiv:2203.12344, 2022. 3 [18] D. Dwibedi, Y. Aytar, J. Tompson, P. Sermanet, and A. Zisserman. Temporal cycle-consistency learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1801 1810, 2019. 3 [19] H. Fan, B. Xiong, K. Mangalam, Y. Li, Z. Yan, J. Malik, and C. Feichtenhofer. Multiscale vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6824 6835, 2021. 8 [20] A. Fathi, J. K. Hodgins, and J. M. Rehg. Social interactions: A first-person perspective. In 2012 IEEE Conference on Computer Vision and Pattern Recognition, pages 1226 1233. IEEE, 2012. 3 [21] C. Feichtenhofer, H. Fan, J. Malik, and K. He. Slowfast networks for video recognition. In Proceedings of the IEEE/CVF international conference on computer vision, pages 6202 6211, 2019. 8 [22] C. Feichtenhofer, H. Fan, B. Xiong, R. Girshick, and K. He. A large-scale study on unsupervised spatiotemporal representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3299 3309, 2021. 3 [23] A. Furnari and G. M. Farinella. Rolling-unrolling lstms for action anticipation from first-person video. IEEE transactions on pattern analysis and machine intelligence, 43(11):4021 4036, 2020. 3 [24] J. F. Gemmeke, D. P. Ellis, D. Freedman, A. Jansen, W. Lawrence, R. C. Moore, M. Plakal, and M. Ritter. Audio set: An ontology and human-labeled dataset for audio events. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 776 780. IEEE, 2017. 6 [25] S. Gidaris, P. Singh, and N. Komodakis. Unsupervised representation learning by predicting image rotations. In International Conference on Learning Representations, 2018. 2 [26] R. Girdhar and K. Grauman. Anticipative video transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 13505 13515, 2021. 3 [27] K. Grauman, A. Westbury, E. Byrne, Z. Chavis, A. Furnari, R. Girdhar, J. Hamburger, H. Jiang, M. Liu, X. Liu, et al. Ego4d: Around the world in 3,000 hours of egocentric video. In CVPR, 2022. 2, 3, 6, 7, 8, 9 [28] J.-B. Grill, F. Strub, F. Altché, C. Tallec, P. Richemond, E. Buchatskaya, C. Doersch, B. Pires, Z. Guo, M. Azar, et al. Bootstrap your own latent: A new approach to self-supervised learning. In Advances in Neural Information Processing Systems, 2020. 1, 2 [29] T. Han, W. Xie, and A. Zisserman. Self-supervised co-training for video representation learning. Advances in Neural Information Processing Systems, 33:5679 5690, 2020. 3 [30] T. Han, W. Xie, and A. Zisserman. Memory-augmented dense predictive coding for video representation learning. In European conference on computer vision, pages 312 329. Springer, 2020. 3 [31] K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9729 9738, 2020. 1, 2 [32] K. Hu, J. Shao, Y. Liu, B. Raj, M. Savvides, and Z. Shen. Contrast and order representations for video self-supervised learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7939 7949, 2021. 3 [33] E. Kazakos, A. Nagrani, A. Zisserman, and D. Damen. Epic-fusion: Audio-visual temporal binding for egocentric action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5492 5501, 2019. 3 [34] E. Kazakos, J. Huh, A. Nagrani, A. Zisserman, and D. Damen. With a little help from my temporal context: Multimodal egocentric action recognition. ar Xiv preprint ar Xiv:2111.01024, 2021. 3 [35] D. Kim, D. Cho, and I. S. Kweon. Self-supervised video representation learning with space-time cubic puzzles. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 8545 8552, 2019. 3 [36] B. Korbar, D. Tran, and L. Torresani. Cooperative learning of audio and video models from self-supervised synchronization. Advances in Neural Information Processing Systems, 31, 2018. 3 [37] K. Lai, L. Bo, and D. Fox. Unsupervised feature learning for 3d scene labeling. In 2014 IEEE International Conference on Robotics and Automation (ICRA), pages 3050 3057. IEEE, 2014. 3 [38] Y. Li, T. Nagarajan, B. Xiong, and K. Grauman. Ego-exo: Transferring visual representations from third-person to first-person videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6943 6953, 2021. 3 [39] M. Liu, S. Tang, Y. Li, and J. Rehg. Forecasting human object interaction: Joint prediction of motor attention and egocentric activity. 2019. 3 [40] I. Misra, C. L. Zitnick, and M. Hebert. Shuffle and learn: unsupervised learning using temporal order verification. In Proceedings of the European Conference on Computer Vision, pages 527 544. Springer, 2016. 3 [41] P. Morgado, Y. Li, and N. Nvasconcelos. Learning representations from audio-visual spatial alignment. Advances in Neural Information Processing Systems, 33, 2020. 3 [42] P. Morgado, I. Misra, and N. Vasconcelos. Robust audio-visual instance discrimination. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12934 12945, 2021. 3, 7 [43] P. Morgado, N. Vasconcelos, and I. Misra. Audio-visual instance discrimination with crossmodal agreement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12475 12486, 2021. 1, 2, 3, 5, 6, 7, 8 [44] J. Munro and D. Damen. Multi-modal domain adaptation for fine-grained action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 122 132, 2020. 3 [45] T. Nagarajan, C. Feichtenhofer, and K. Grauman. Grounded human-object interaction hotspots from video. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 8688 8697, 2019. 3 [46] E. Ng, D. Xiang, H. Joo, and K. Grauman. You2me: Inferring body pose in egocentric video via first and second person interactions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9890 9900, 2020. 3 [47] M. Noroozi and P. Favaro. Unsupervised learning of visual representations by solving jigsaw puzzles. In Proceedings of the European Conference on Computer Vision, pages 69 84. Springer, 2016. 2 [48] A. v. d. Oord, Y. Li, and O. Vinyals. Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748, 2018. 1, 2 [49] A. Owens and A. A. Efros. Audio-visual scene analysis with self-supervised multisensory features. In Proceedings of the European Conference on Computer Vision (ECCV), pages 631 648, 2018. 3 [50] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2536 2544, 2016. 2 [51] M. Patrick, Y. M. Asano, P. Kuznetsova, R. Fong, J. F. Henriques, G. Zweig, and A. Vedaldi. Multi-modal self-supervision from generalized data transformations. ar Xiv preprint ar Xiv:2003.04298, 2020. 3 [52] A. Piergiovanni, A. Angelova, and M. S. Ryoo. Evolving losses for unsupervised video representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 133 142, 2020. 3 [53] H. Pirsiavash and D. Ramanan. Detecting activities of daily living in first-person camera views. In 2012 IEEE conference on computer vision and pattern recognition, pages 2847 2854. IEEE, 2012. 3 [54] S. Purushwalkam, T. Ye, S. Gupta, and A. Gupta. Aligning videos in space and time. ar Xiv preprint ar Xiv:2007.04515, 2020. 1 [55] R. Qian, T. Meng, B. Gong, M.-H. Yang, H. Wang, S. Belongie, and Y. Cui. Spatiotemporal contrastive video representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6964 6974, 2021. 3 [56] M. Ramazanova, V. Escorcia, F. C. Heilbron, C. Zhao, and B. Ghanem. Owl (observe, watch, listen): Localizing actions in egocentric video via audiovisual temporal context. ar Xiv preprint ar Xiv:2202.04947, 2022. 3 [57] A. Recasens, P. Luc, J.-B. Alayrac, L. Wang, F. Strub, C. Tallec, M. Malinowski, V. P atr aucean, F. Altché, M. Valko, et al. Broaden your views for self-supervised video learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1255 1265, 2021. 1, 3 [58] B. Singh, T. K. Marks, M. Jones, O. Tuzel, and M. Shao. A multi-stream bi-directional recurrent neural network for fine-grained action detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1961 1970, 2016. 3 [59] K. K. Singh, K. Fatahalian, and A. A. Efros. Krishnacam: Using a longitudinal, singleperson, egocentric dataset for scene understanding tasks. In 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), pages 1 9. IEEE, 2016. 3 [60] Y.-C. Su and K. Grauman. Detecting engagement in egocentric video. In European Conference on Computer Vision, pages 454 471. Springer, 2016. 3 [61] Y. Tian, D. Krishnan, and P. Isola. Contrastive multiview coding. In Proceedings of the European Conference on Computer Vision, pages 776 794. Springer, 2020. 6 [62] D. Tran, H. Wang, L. Torresani, J. Ray, Y. Le Cun, and M. Paluri. A closer look at spatiotemporal convolutions for action recognition. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 6450 6459, 2018. 6 [63] L. van der Maaten and G. Hinton. Visualizing data using t-sne. Journal of Machine Learning Research, 2008. 9 [64] C. Vondrick, H. Pirsiavash, and A. Torralba. Anticipating visual representations from unlabeled video. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 98 106, 2016. 3 [65] J. Wang, J. Jiao, L. Bao, S. He, Y. Liu, and W. Liu. Self-supervised spatio-temporal representation learning for videos by predicting motion and appearance statistics. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4006 4015, 2019. 3 [66] J. Wang, J. Jiao, and Y.-H. Liu. Self-supervised video representation learning by pace prediction. In Proceedings of the European Conference on Computer Vision, pages 504 521. Springer, 2020. 3 [67] X. Wang and A. Gupta. Unsupervised learning of visual representations using videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2794 2802, 2015. 1 [68] X. Wang, A. Jabri, and A. A. Efros. Learning correspondence from the cycle-consistency of time. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2566 2576, 2019. 3 [69] D. Wei, J. J. Lim, A. Zisserman, and W. T. Freeman. Learning and using the arrow of time. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018. 3 [70] D. Xu, J. Xiao, Z. Zhao, J. Shao, D. Xie, and Y. Zhuang. Self-supervised spatiotemporal learning via video clip order prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10334 10343, 2019. 3 [71] Z. Zeng, D. Mc Duff, Y. Song, et al. Contrastive learning of global and local video representations. Advances in Neural Information Processing Systems, 34:7025 7040, 2021. 3 [72] C. Zhang, A. Gupta, and A. Zisserman. Temporal query networks for fine-grained video understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4486 4496, 2021. 3 Please do not modify the questions and only use the provided macros for your answers. Note that the Checklist section does not count towards the page limit. In your paper, please delete this instructions block and only keep the Checklist section heading above along with the questions/answers below. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] We discuss the limitations after the conclusions section of the main paper along with the potential avenues for future research inspired by the limitations of our method. (c) Did you discuss any potential negative societal impacts of your work? [Yes] Refer to the section after the conclusions (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] The code and instructions are available publicly. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] Refer to Section 3.5: Implementation Details (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [No] Since the train and test set are too large, it takes a lot of time and resource to run the experiments multiple times. However, errors bars are expected to be small, given their sizes. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] Refer to Section 3.5: Implementation Details 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]