# audiodriven_cospeech_gesture_video_generation__cf092bcf.pdf Audio-Driven Co-Speech Gesture Video Generation Xian Liu1, Qianyi Wu2, Hang Zhou1, Yuanqi Du3, Wayne Wu4, Dahua Lin1,4, Ziwei Liu5 1Multimedia Laboratory, The Chinese University of Hong Kong 2Monash University 3Cornell University 4Shanghai AI Laboratory 5S-Lab, Nanyang Technological University Co-speech gesture is crucial for human-machine interaction and digital entertainment. While previous works mostly map speech audio to human skeletons (e.g., 2D keypoints), directly generating speakers gestures in the image domain remains unsolved. In this work, we formally define and study this challenging problem of audio-driven co-speech gesture video generation, i.e., using a unified framework to generate speaker image sequence driven by speech audio. Our key insight is that the co-speech gestures can be decomposed into common motion patterns and subtle rhythmic dynamics. To this end, we propose a novel framework, Audio-drive N Gesture v Ideo g Eneration (ANGIE), to effectively capture the reusable co-speech gesture patterns as well as fine-grained rhythmic movements. To achieve high-fidelity image sequence generation, we leverage an unsupervised motion representation instead of a structural human body prior (e.g., 2D skeletons). Specifically, 1) we propose a vector quantized motion extractor (VQ-Motion Extractor) to summarize common co-speech gesture patterns from implicit motion representation to codebooks. 2) Moreover, a co-speech gesture GPT with motion refinement (Co-Speech GPT) is devised to complement the subtle prosodic motion details. Extensive experiments demonstrate that our framework renders realistic and vivid co-speech gesture video. Demo video and more resources can be found in: https://alvinliu0.github.io/projects/ANGIE 1 Introduction During daily conversation among humans, speakers naturally emit co-speech gestures to complement the verbal channels and express their thoughts [17, 35, 56]. Such non-verbal behaviors ease speech comprehension [10, 58] and bridge the communicator s gap for better credibility [7, 54]. Therefore, equipping the social robot with conversation skills constitutes a crucial step to human-machine interaction. To achieve it, researchers delve into the task of co-speech gesture generation [21, 39, 62], where audio-coherent human gesture sequences are synthesized in the form of structural human representation (e.g., skeletons). However, such representation contains no appearance information of the target speaker, which is crucial for human perception. As demonstrated in audio-driven talking head synthesis [34, 65], generating real-world subjects in the image domain is highly desirable. To this end, we explore the problem of audio-driven co-speech gesture video generation, i.e., using a unified framework to generate speaker image sequence driven by speech audio (illustrated in Fig. 1). Conventional methods require exhaustive human efforts to pre-define the speech-gesture pairs and connection rules for coherent result [11, 12]. With the development of deep learning, neural networks are leveraged to learn the mapping from encoded audio feature to human skeletons in a data-driven manner [21, 39, 62]. Notably, one category of approaches relies on small-scale Mo Cap datasets in co-speech setting [16, 18, 48], which contributes to specific models with limited capacity and robustness. To capture more general speech-gesture correlations, another category of methods builds large training corpus by exploiting off-the-shelf pose estimators [9, 15] to label enormous online videos as pseudo ground truth [21, 63]. However, the inaccurate pose annotations induce error 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Co-Speech Gesture Image Generator Figure 1: Illustration of Problem Setting. In this paper, we focus on audio-driven co-speech gesture video generation. Given an image with speech audio, we generate aligned speaker image sequence. accumulation in the training phase, which makes the generated results unnatural. Besides, most previous works ignore the problem of co-speech gesture video generation. Only few works [21, 39] animate in the image domain as an independent post-processing step, which borrows from the existing pose-to-image generators [5, 13] to train on the target person s images. How to design a unified framework to generate speaker image sequence driven by speech audio remains unsolved. To effectively learn the mapping from audio to co-speech gesture video, we pinpoint two important observations from current studies: 1) hand-crafted structural human priors like 2D/3D skeletons would eliminate articulated human body region information. Such a zeroth-order motion representation fails to formulate first-order motion like local affine transformation in image animation [44]. Besides, the error in structural prior labeling impairs cross-modal audio-to-gesture learning [33]. 2) Motivated by previous linguistic studies [27, 47], the co-speech gestures could be decomposed into common motion patterns and rhythmic dynamics, where the former ones refer to large-scale motion templates (e.g., periodically put hands up and down), while the latter ones play a refinement role to complement subtle prosodic movements and synchronize with speech audio (e.g., finger flickers). We take inspiration from the above observations and propose a novel framework Audio-drive N Gesture v Ideo g Eneration (ANGIE) to generate co-speech gesture video. The key insight is to summarize common co-speech gesture patterns from motion representation to quantized codebooks and further refine subtle rhythmic details by motion residuals for fine-grained results. In particular, two modules are designed, namely VQ-Motion Extractor and Co-Speech GPT. In VQ-Motion Extractor, we utilize an unsupervised motion representation to depict the articulated human body and first-order gestures [45]. The codebooks are established to quantize the reusable common co-speech gesture patterns from unsupervised motion representation. To guarantee the validity of gesture patterns, we propose a cholesky decomposition based quantization scheme to relax the motion component constraint. The position-irrelevant motion pattern is extracted as final quantization target to represent the relative motion. In this way, the quantized codebooks naturally contain rich common gesture pattern information. With the quantized motion code sequence, in Co-Speech GPT we use a GPT-like [40] structure to predict discrete motion patterns from speech audio. Finally, a motion refinement network is used to complement subtle rhythmic details for fine-grained results. To summarize, our main contributions are three-fold: 1) We explore a challenging problem of audiodriven co-speech gesture video generation. To the best of our knowledge, we are the first to generate co-speech gesture in image domain with a unified framework without any structural human body prior. 2) We propose the VQ-Motion Extractor to quantize the motion representation into common gesture patterns and the Co-Speech GPT to refine subtle rhythmic movement details. The codebooks naturally contain reusable motion pattern information. 3) Extensive experiments demonstrate that the proposed framework ANGIE renders realistic and vivid co-speech gesture video generation results. 2 Related Work Co-Speech Gesture Generation. Synthesizing co-speech gesture has gained research interest in vision [3, 21, 29], graphics [4, 60, 62] and robotics [23, 25, 63] domains. Recent researches resort to deep neural networks to learn the speech-gesture mapping in a data-driven manner, with major focuses on below perspectives: 1) Dataset. One strand of methods use small-scale Mo Cap datasets to learn specific models [16, 18, 30, 42, 48, 51, 55], while another strand of works exploit off-the-shelf estimator to label enormous videos as structural prior [1 3, 21, 33, 39, 61 63]. The dataset scale v.s. pose annotation accuracy often acts as a trade-off in this task: A large amount of speech-gesture pairs facilitate the training of more general models with better capacity and robustness, yet error accumulation in annotations induces unnatural results. 2) Framework architecture. CNN-based [21], Motion Estimator Residual Motion Refinement Image Generator VQ-Motion Extractor VQ-Motion Extractor Cholesky Decomposition Relative Difference Relative Difference arg𝑚𝑖𝑛|| 𝑒 ఓ 𝑑 ఓ|| arg𝑚𝑖𝑛||𝑒 𝑑 || Co-Speech GPT Transformer Layer Quantized Δ𝜇 Recover Shift Recover Factorial Covariance Δ𝜇/ΔL Decoder Motion Representation Co-Speech GPT with Motion Refinement Co-Speech GPT with Motion Refinement Driving Audio Linear Layer Soft Max Soft Max Top1 Selection C Quantized Δ𝐿 𝐿 Recover Covariance Bi-directional Source Image Transformer Layer Figure 2: Overview of the Audio-drive N Gesture v Ideo g Eneration (ANGIE) framework. In VQ-Motion Extractor, the cholesky decomposition with position-irrelevant design transforms the shift-translation µ and covariance C to relative motion pattern representation of µ and L, which are further quantized by codebooks to extract the common motion patterns. Given the driving audio and starting gesture codes, the Co-Speech GPT predicts the future motion fields. A Motion Refinement network further learns motion residuals to complement the subtle rhythmic dynamics. RNN-based [62] and Transformer-based [6] frameworks show promising results. To further improve the diversity of generated gestures and grasp the fine-grained cross-modal associations, components like adversarial loss [21], VAE sampling [30, 59] and hierarchical encoder-decoder design [33] are proposed. 3) Input modality. Some approaches treat single modality of speech audio [2, 19, 21 23, 30, 39] or text transcription [3, 6, 25, 63] as input to drive the co-speech gesture, while some others use both modalities as stimuli for generation [1, 33, 62]. To ease the learning of implicit cross-modal mapping from speech to gesture and create more stable results, recent works involve auxiliary input modality such as speaker style [2], pose mode [59] and motion template [39]. In this work, we take a step further in the above three aspects: 1) For the dataset, we collect a new co-speech gesture dataset in image domain, where an unsupervised motion representation is used to model articulated human body and bypass the inaccuracy from structural prior annotations. 2) For the architecture, a vector quantized (VQ) network with novel discretization scheme is proposed to extract valid relative motion patterns. We further devise a motion refinement network to complement subtle rhythmic dynamics. 3) For the input modality, we explicitly decouple the common motion pattern from co-speech gestures, which serves a similar role as motion template [39] to provide auxiliary input. However, our discrete codebook design is more suitable for finite gesture patterns than continuous representation, which is also proven in recent cross-modal generation tasks [36, 41, 46]. Furthermore, we propose a novel vector quantization network with cholesky decomposition scheme to extract the valid motion patterns. We improve the quantization scheme to encode the relative motion representation that is position (absolute location) irrelevant. A motion refinement network is further devised to complement subtle rhythmic dynamics. Notably, our approach gives an idea on how to deal with the constraints in vector quantization and how to complement sequential results with missing details. Such design could prospectively provide insights for relevant domains like constrained vector quantization problem, cross-modal learning [32] and video generation tasks [46]. Video/Audio-Driven Video Generation. Traditional video-driven approaches for image animation can be categorized into supervised and unsupervised, where the supervised methods typically involve structural human body prior such as landmarks [8, 64] and 3D parametric models [20, 50], while the unsupervised approaches design self-supervised tasks to animate unlabeled images [43 45, 57]. To facilitate broader applications, researchers explore audio-driven video generation, where one of the most relevant tasks is talking face generation [14, 38]. Different from the strong correlations between audio and mouth shape, the mapping from audio to complicated co-speech motion is multi-modal and harder to learn. Most co-speech gesture studies synthesize human skeletons as final results (e.g., 2D keypoints), while only few works [21, 39] generate co-speech images in a post-processing manner. 3 Our Approach We present ANGIE that generates audio-driven co-speech gesture in image domain, where the speakers image sequence is driven by speech audio as shown in Fig. 1. The whole pipeline is illustrated in Fig. 2. To make the content self-contained and narration clearer, we first introduce the preliminaries and problem setting in Sec. 3.1. Then, we present the VQ-Motion Extractor which extracts common co-speech gesture patterns as quantized codebooks in Sec. 3.2. Finally, we elaborate the Co-Speech GPT to complement subtle rhythmic dynamics for fine-grained results in Sec. 3.3. 3.1 Preliminaries and Problem Setting Unsupervised Motion Representation. To achieve high-fidelity image animation, we take inspiration from MRAA [45] that uses an unsupervised motion representation to drive articulated objects. MRAA first estimates a coarse motion representation from the source and driving frames, then predicts dense pixel-wise flow for image generation. Specifically, an encoder-decoder keypoint predictor produces K different heatmaps H1, H2, . . . , HK, where K is region number and Hk denotes the k-th image region. Afterwards, each heatmap is normalized by softmax operation, i.e, P z Z Hk(z) = 1, where z is the image pixel location and Z is the set of all pixels. The key insight behind MRAA is to represent each region s motion by affine transformation with a shift-translation component. The shift-translation component µk R2 and the distribution Ck of the k-th part can be calculated as: z Z Hk(z)z, Ck = X z Z Hk(z)(z µk)(z µk)T, (1) where T is matrix transpose, Ck R2 2 measures the covariance of heatmap value. It naturally depicts the size and shape of an articulated region. To represent the affine transformation Ak R2 2 of the k-th region, we apply singular value decomposition (SVD) to Ck and derive Ak as: Ck = U kΣk(V k)T, Ak = U kΣk 1 2 , (2) where unitary matrices U k, V k and diagonal matrix Σk are the SVD result of covariance matrix Ck. The representation M extracted by motion estimator is the concatenation of [µ; C; A] RK (2+4+4) for K distinct regions, an image generation module with dense pixel-wise flow predictor synthesizes the final generation results. In this work, the motion representation M and image generation module G generally follow MRAA. We suggest the readers referring to [45] for more details. Problem Setting for Co-Speech Gesture Image Generation. We collect training data of massivelyavailable speaking videos with clear co-speech gestures for natural self-reconstruction supervision. Specifically, given an (N + 1)-frame video clip V = {I(0), . . . , I(N)}, the goal of our framework at the training stage is to predict the motion representation c M(1:N) based on the first image frame I(0) and video s accompanying audio sequence a = {a(1), . . . , a(N)}. Further, the image generation module G reconstructs the video frames b I(1:N). At the inference stage, an arbitrary reference image with speech audio clip is provided to generate subsequent image frames. According to the observation in Sec. 1, we decompose the co-speech gestures into common motion patterns and subtle rhythmic dynamics. The overall training setting can be formulated as: b I(1:N) = G(I(0), c M(1:N)(a)), c M(1:N) = c Mpattern (1:N) + c Mrhythmic (1:N) , (3) where Mpattern denotes the gesture pattern (Sec. 3.2) and Mrhythmic is rhythmic movement (Sec. 3.3). 3.2 Vector Quantized Motion Pattern Extractor To decompose the co-speech gestures, we propose to firstly extract the common motion patterns. However, three problems remain: 1) The gesture sequences are different from each other. While some motion sequences share the same action pattern, the dynamic details may vary a lot. How to extract the major motion pattern despite the influence from minor prosodic movements? 2) The covariance matrix C is symmetric positive definite (Eq. 1), which further constrains the range of affine matrix A. How to preserve such characteristic for valid gesture patterns? 3) Since the unsupervised motion representation is extracted in the image pixel space, it is affected by the absolute location of each articulated region. How to represent the position-irrelevant motion pattern information? Vector Quantized Motion Pattern Learning. Our solution to the first problem is to quantize the common motion pattern into a codebook. Since the gesture pattern is finite, it could be summarized to discrete codebook entries. Besides, each codebook entry refers to a certain type of gesture pattern, which matches our goal to extract the common and reusable co-speech gesture patterns. A naive way is to quantize the motion representation M separately as shift-translation µ, covariance matrix C and affine transformation A. Specifically, for a T-frame co-speech gesture sequence I(1:T ), we transform it into [µ; C; A](1:T ) RT K (2+4+4), where [µ; C; A] = M denotes motion representation of K regions as in Sec. 3.1. We first build three codebooks Dµ = {dµ,m}M m=1, DC = {d C,m}M m=1 and DA = {d A,m}M m=1 for each motion component respectively, where M is codebook size. dµ,m, d C,m and d A,m Rℓare the m-th entry of ℓ-channel for each codebook. Then, three separate encoders Eµ, EC and EA are utilized to encode the corresponding context information into latent features of eµ = {eµ,i}T i=1, e C = {e C,i}T i=1 and e A = {e A,i}T i=1 RT ℓ, where T is the temporal dimension and ℓis the channel dimension. Notably, we denote the i-th temporal feature of each motion component as eµ,i, e C,i and e A,i. The feature encoding process can be formulated as: Eµ(µ(1:T )) = eµ, EC(C(1:T )) = e C, EA(A(1:T )) = e A. (4) Following the pipeline of VQ-VAE [53], we individually quantize eµ, e C and e A by substituting each temporal feature eµ,i, e C,i and e A,i to the nearest codebook entry dµ,m, d C,m and d A,m as: eq µ = arg min dµ Dµ ||eµ dµ|| | {z } quantize shift-translation µ , eq C = arg min d C DC ||e C d C|| | {z } quantize covariance matrix C , eq A = arg min d A DA ||e A d A|| | {z } quantize affine transformation A where eq µ = {eq µ,i}T i=1, eq C = {eq C,i}T i=1 and eq A = {eq A,i}T i=1 RT ℓare the quantized code sequence of length T for each motion component. The i-th quantized code of each motion component is denoted as eq µ,i, eq C,i and eq A,i, respectively. Finally, three separate decoders Dµ, DC and DA are leveraged to reconstruct the motion representations of each component as: bµ(1:T ) = Dµ(eq µ), b C(1:T ) = DC(eq C), b A(1:T ) = DA(eq A). (6) Such discrete representation also ease the audio-to-gesture learning (Sec. 3.3): Previous methods predict continuous output as a harder regression problem. While we only need to predict features nearer to the correct codebook entry, which in essence resembles an easier classification problem. Quantization Design for Valid Motion Representation. To extract valid gesture patterns, we have to preserve certain characteristics of motion representation. Especially, the covariance matrix C should be symmetric positive definite (Eq. 1), and the affine transformation A is determined by C through SVD (Eq. 2). Therefore, instead of naively quantize each component in Eq. 5, we propose to only quantize the shift-translation µ and covariance matrix C, while derive the affine transformation A with SVD. The only constraint is to guarantee that the covariance matrix C is symmetric positive definite. To satisfy such requirement, we use the unique cholesky decomposition theorem [52]: Theorem 1. For any real symmetric positive definite matrix C Sn ++, there exists a unique lower triangular matrix L with positive diagonal entries, such that C = LLT. In this way, we turn to quantize the lower triangular matrix L = l1 0 l2 l3 , where the constraint is much simpler as l1, l3 > 0. The updated quantization scheme with cholesky decomposition is: eq µ = arg min dµ Dµ ||eµ dµ|| | {z } quantize shift-translation µ , eq L = arg min d L DL ||e L d L|| | {z } quantize the lower triangular matrix L where e L, eq L, DL and d L denote the encoded feature, quantized feature, codebook and codebook entry for factorial covariance L, respectively. A simple transformation of l1,3 = Re LU(l1,3) + ϵ guarantees the diagonal entries to be positive, where ϵ is a small positive number. The motion component C and A can be further obtained by LLT and SVD calculation, respectively. Position-Irrelevant Motion Pattern. Another problem arises when we inspect the value of the motion representation: Since M is extracted in the image pixel space, the object location will affect the element in M. For example, if a person poses the same gesture at different image regions, the motion component differs yet the underlying motion pattern remains the same. Thus we focus on a image location invariant motion pattern representation. In particular, due to the linear additiveness, the relative shift-translation µ between adjacent frames can be represented as µj = µj µj 1, and the relative change of the lower triangular matrix is Lj = Lj Lj 1 for any j = 2, . . . , N. Note that with the uniqueness of cholesky decomposition, (L + L) corresponds to the sole covariance matrix C = (L + L)(L + L)T, which further determines affine matrix A by SVD. In this way, the term L is sufficient to represent any relative affine transformation between two frames. We accordingly update the quantization scheme with position-irrelevant motion pattern representation as: eq µ = arg min d µ D µ ||e µ d µ|| | {z } quantize relative shift-translation µ , eq L = arg min d L D L ||e L d L|| | {z } quantize relative lower triangular matrix change L where e{ µ, L}, eq { µ, L}, D{ µ, L} and d{ µ, L} are the encoded feature, quantized feature, codebook and entry for the relative shift-translation µ and factorial covariance L, respectively. Overall Quantized Motion Pattern Learning. With the position-irrelevant motion pattern, the codebook naturally contains reusable common co-speech gesture patterns Mpattern. The encoders E µ, E L and the decoders D µ, D µ are jointly learned with the codebooks D µ and D L via: LVQ = || c µ µ|| + ||sg[e µ] eq µ|| + β1||e µ sg[eq µ]||+ ||d L L|| + ||sg[e L] eq L|| + β2||e L sg[eq L]||, (9) where sg denotes the stop gradient operation, β1 and β2 are two weight balancing coefficients. 3.3 Co-Speech Gesture GPT with Motion Refinement Co-Speech Gesture GPT Network. With the position-irrelevant motion pattern of valid quantization design, each co-speech gesture clip can be transformed into discrete representation. We then learn a cospeech gesture GPT network to map from speech audio a(1:T ) to quantized code sequences eq µ,(1:T ) and eq L,(1:T ). Specifically, we extract audio features aonset (1:T ) with onset strength information, which is more suitable for cross-modal pattern learning [46, 49]. Then, a feature embedding layer with positional embedding is leveraged to obtain the tokens for audio onset features, quantized relative shift-translation and qunatized relative factorial covariance. Further, we encode cross-attention information with a series of transformer layers. Finally, followed by a linear transformation with softmax activation, the M-dimensional output denotes the probability of each quantization code at that time step. The whole co-speech gesture GPT is trained with cross-entropy loss LCE. Such design enables us to predict and sample future quantization code eq µ and eq L with speech audio. Motion Refinement by Learning Residuals. Now that we can reconstruct the relative shifttranslation c µ and factorial covariance change d L by VQ-VAE decoding. Given µ1 and L1 extracted from the initial image frame I(1), the absolute shift-translation and factorial covariance for the j-th frame can be calculated as c µj = µ1 + Pj i=2 d µi and c Lj = L1 + Pj i=2 d Li, respectively. However, since the quantized codebook is designed to only represent the large-scale common motion pattern information, while fine-grained rhythmic details are omitted. Therefore, we propose to refine the co-speech movements by learning residual terms. Concretely, we extract the audio mfcc features amfcc (1:T ) to encode more contextual audio cues for prosodic dynamics learning. Then a bi-directional LSTM is used to predict the per-frame motion representation residuals R to the main motion pattern result bµ(1:T ) and b L(1:T ), i.e., c Mrhythmic (1:T ) = h R(bµ(1:T ); amfcc (1:T )); R(b L(1:T ); amfcc (1:T )) i . By adding residual terms, the overall co-speech gesture GPT with motion refinement learning can be formulated as: LResidual = ||M(1:T ) c M(1:T )(a)||, where c M(1:T )(a) = c Mpattern (1:T ) (aonset) + c Mrhythmic (1:T ) (amfcc). (10) In this way, we capture both major gesture patterns and subtle rhythmic dynamics for vivid results. Table 1: The quantitative results on PATS Image Dataset. We compare the proposed Audio-drive N Gesture v Ideo g Eneration (ANGIE) against recent SOTA methods [21, 33, 39, 62] and ground truth on four speakers subsets. For FGD the lower the better, and the higher the better for other metrics. Oliver Seth Kubinec Jon Methods FGD BC Div. FGD BC Div. FGD BC Div. FGD BC Div. GT 0.00 0.76 54.6 0.00 0.71 49.3 0.00 0.84 38.9 0.00 0.73 62.8 S2G [21] 8.57 0.59 46.1 5.75 0.62 38.2 4.76 0.67 31.6 6.07 0.51 47.3 HA2G [33] 3.28 0.75 49.2 4.06 0.72 40.1 2.98 0.79 32.3 3.74 0.64 50.2 SDT [39] 1.04 0.61 52.9 1.97 0.58 46.7 1.15 0.77 36.1 1.63 0.60 57.4 Tri Con [62] 3.63 0.53 48.3 3.79 0.52 40.3 3.27 0.77 35.7 3.98 0.61 49.7 ANGIE 0.88 0.72 53.5 1.83 0.69 46.7 1.10 0.81 36.5 1.57 0.65 60.9 4 Experiments 4.1 Experimental Settings Dataset and Preprocessing. Pose, Audio, Transcript, Style (PATS) is a large-scale dataset of 25 speakers with aligned pose, audio and transcripts [1, 2, 21]. The training corpus contains 251 hours of data with around 84,000 intervals of mean length 10.7s. Notably, the PATS dataset contains three modalities of audio log-mel spectrograms, speech transcripts and per-frame skeletons labeled with Open Pose [9]. To bypass the error accumulation in pose annotation and facilitate co-speech gesture image generation task, we extend PATS with more features: 1) preprocessed image frames and 2) onset strength audio features which are more suitable for co-speech gesture pattern learning. We conduct the experiments on four speakers co-speech video subsets, including Oliver, Seth, Kubinec and Jon. Concretely, 2D skeletons of the image frames are obtained by Open Pose [9] for baseline methods training. The frames are cropped to make the speaker locate at the image center. Since the time span of a meaningful co-speech gesture unit sequence ranges from 4s to 14s [27, 47], we trim invalid videos and finally obtain 1306, 990, 1294 and 1284 clips for four subsets, respectively. The overall mean clip length is 9.8s. We randomly split the segments into 90% for training and 10% for evaluation. The image frames are sampled at 25 fps and further resized to 256 256. Comparison Methods. We compare with recent SOTA works: 1) Speech to Gesture (S2G) [21], a GAN-based pipeline that maps audio to 2D keypoints with a U-Net; 2) Hierarchical Audio to Gesture (HA2G) [33] which captures the hierarchical associations between multi-level audio features and treelike human skeletons; 3) Speech Drives Template (SDT) [39] which relieves the one-to-many mapping ambiguity by a set of continuous gesture template vectors; 4) Trimodal Context (Tri Con) [62], a representative framework that considers the trimodal context of audio, text and speaker identity. Note that all methods could drive 2D human skeletons with speech audio. We train baselines on the PATS image dataset and tune the hyper-parameters by grid search for the best evaluation result. In particular, we also show direct evaluations on the Ground Truth (GT) skeletons for clearer comparison. Implementation Details. We sample T = 96 frame clips with stride 32 for training. 1) For the VQMotion Extractor: the co-speech gesture pattern codebook size M for both relative shift-translation µ and factorial covariance change L are set to 512. The encoders E µ, E L and the decoders D µ, D µ are based on 1D-convolution structure. The channel dimension ℓof each codebook entry d µ, d L as well as the encoded latent features e µ, e L are 512, while the temporal dimension T is set as T/8 = 12 to encode contextual features with downsampling rate of 8. The ϵ is set as 1 10 5 to guarantee the positiveness of diagonal entries in factorial covariance L. The commit loss trade-offs in LVQ are empirically set as β1 = β2 = 0.1. We optimize the gesture pattern VQ-VAE with Adam optimizer [28] of learning rate 3 10 5. 2) For the Co-Speech GPT: the Transformer channel dimension is 768, and the attention layer is implemented in 12 heads with dropout probability of 0.1. The onset strength audio features aonset R426 are extracted by Librosa, while the audio mfcc features amfcc R28 12 are computed with the window size of 10 ms. During the GPT training, the eq µ,(1:11), eq L,(1:11) are used as input while eq µ,(2:12), eq L,(2:12) serve as supervision labels. 3) For the motion representation M and image generator G: we implement as MRAA [45] to use K = 20 regions. The motion estimator is pretrained for knowledge distillation. The overall framework is implemented in Py Torch [37] and trained on one 16G Tesla V100 GPU for three days. Table 2: User study results on co-speech gesture generation quality. The rating scale is 1-5, with the larger the better. We compare the Realness, Synchrony and Diversity to baselines [21, 33, 39, 62]. Methods GT S2G [21] HA2G [33] SDT [39] Tri Con [62] ANGIE (Ours) Realness 4.29 3.27 3.92 4.01 3.74 4.08 Synchrony 4.36 3.48 4.01 3.97 3.85 4.11 Diversity 3.97 2.49 3.31 3.88 3.02 3.95 Figure 3: Image sequence results of ANGIE. We show the co-speech gesture image generation results of Kubinec, Seth and Jon, respectively. More qualitative results can be found in demo video. 4.2 Quantitative Evaluation Evaluation Metrics. We adopt 1) Fréchet Gesture Distance (FGD) [62] to evaluate the distance between the real and synthetic gesture distribution. We train an auto-encoder on the PATS image dataset and use the encoder to compute fréchet distance between the real and synthetic gesture in feature space. We also use the 2) Beat Consistency Score (BC) and 3) Diversity (Div.) [31, 33] to account for the speech-motion alignment and the diversity among generated gestures. Specifically, BC is computed as the average temporal distance between each audio beat and its closest gesture beat, and Diversity indicates the difference of multiple audios corresponding gestures in the latent space. Note that since all metrics are skeleton-based, we downgrade our method to operate on skeleton data for fair comparison, i.e., we use VQ-VAE w/o cholesky scheme to create 2D skeletons for evaluation. Evaluation Results. The results are reported in Table 1. It can be seen that the proposed ANGIE achieves the best evaluation results on most metrics. Since our method summarizes reusable cospeech gesture patterns into quantized codebooks and complements subtle rhythmic dynamics, we can cover richer gesture patterns and create diverse results. Note that HA2G [33] tends to generate over-expressive gestures with multi-level audio features, which makes their results on BC even better than the ground truth in some cases. Despite of this, we perform comparable to ground truth on BC metric with stable motion results, showing that we can generate audio-aligned gestures. Besides, we can find that both SDT [39] and ours perform better on FGD and Diversity metrics than other methods due to the explicit modeling of co-speech gesture patterns. However, since the gesture pattern is finite and discrete, our quantized codebook design is more suitable than continuous representation. 4.3 Qualitative Analysis User Study. We further conduct a user study to reflect the quality of audio-driven gestures. Concretely, we sample 25 audio clips from the PATS image test set for all methods to generate skeleton results, then involve 18 participants for user study. The Mean Opinion Scores protocol is adopted, which requires the participants to rate three aspects: (1) Realness; (2) Synchrony; (3) Diversity. The rating scale is 1 to 5, with 1 being the poorest and 5 being the best. The results are reported in Table 2, where our method performs the best on all three aspects. Notably, with the help of motion pattern codebook, Table 3: Ablation study of VQ-Motion Extractor and Co-Speech GPT with Motion Refinement. Ablation Settings FGD BC Diversity L1 Perceptual AED w/o Quantization 5.86 0.54 35.6 0.071 79.2 0.095 Quantize µ, L - - - 0.058 67.4 0.086 Quantize µ, L - - - 0.043 52.3 0.069 Quantize µ, L - - - 0.052 63.6 0.083 w/o Motion Refinement 1.39 0.58 48.3 0.041 48.1 0.063 ANGIE (Ours) 1.35 0.72 49.4 0.037 42.9 0.063 Drive the Same Image with Different Patterns Drive the Same Image with Different Patterns Drive Different Images with the Same Pattern Drive Different Images with the Same Pattern Figure 4: Codebook Analysis. We validate that the codebooks contain meaningful motion patterns. we achieve comparable diversity result to the ground truth, demonstrating that the quantization design helps to capture common gesture patterns. Video Generation Results. With the Co-Speech GPT and motion refinement network, we could predict the future quantization code and complement rhythmic motion details based on audio features, which enables us to generate video results. As shown in Fig. 3, the synthesized image sequence results from left to right contain diverse and meaningful gestures that are aligned with speech audio. Codebook Analysis for Co-Speech Gesture Pattern. We analyze the meaning of the quantized codebooks by asking the question: Does each codebook entry really correspond to a certain type of motion pattern? To answer this, we validate two things in Fig. 4: 1) Different entries represent diverse gesture patterns, hence driving the same image with different quantized codes leads to different gestures (left). 2) Each codebook entry denotes a fixed motion pattern, hence driving different images with the same quantized code show same motions (right). Please refer to demo video for more results. 4.4 Ablation Study In this section, we present ablation study of two modules in our framework. Note that except for the skeleton-based metrics, we also use video reconstruction accuracy as a proxy for image quality, including L1 and perceptual loss [26] between the reconstructed and GT image; Average Euclidean Distance (AED) evaluates identity preservation by pretrained re-identification networks [24, 45]. VQ-Motion Extractor. We conduct ablation experiments on five settings: 1) w/o Quantization, where we directly infer motion representation with audio; 2) Quantize µ, L; 3) Quantize µ, L; 4) Quantize µ, L, where we quantize the absolute or relative difference of the motion components; 5) w/o Cholesky Decomposition, which means the covariance matrix is not guaranteed to be symmetric positive definite. The results are shown in Table 3, which proves that the quantization design with position-irrelevant relative motion pattern could improve the performance. We also find that quantizing µ is more effective than L, since the shift translation is more correlated with region position. Note that when motion component is invalid, the pipeline fails to generate image due to the numerical instability in calculating affine matrix. Hence the results for the setting 5 are not reported. Motion Refinement Module. To verify the efficacy of motion residuals, we eliminate the motion refinement for ablation. The results in Table 3 suggest that this module improves beat consistency by capturing subtle rhythmic movements, while the improvement on FGD majorly derives from the quantization design. Such module complements the motion pattern for fine-grained results. 5 Discussion Conclusion. In this paper, we propose a novel framework ANGIE to generate audio-driven co-speech gesture video in the image domain. To summarize valid common co-speech gesture patterns, we propose VQ-Motion Extractor with cholesky decomposition based quantization scheme and positionirrelevant design to represent relative motion patterns. Then we propose Co-Speech GPT to refine subtle rhythmic movement details for fine-grained results. Extensive experiments demonstrate that our framework renders realistic and vivid co-speech gesture video generation results. Ethical Consideration. Generating co-speech gesture images facilitates applications such as digital human. However, it could be misused for malicious purposes like forgery generation. We believe that the proper use of this technique will enhance the machine learning research and digital entertainment. Limitation and Future Work. As an early work towards audio-driven co-speech video generation without structural prior, we notice that our method fails for some challenging cases. For example, if the source image is in a large pose, it is hard to generalize well in such an out-of-domain data. We will explore how to develop a model with higher generalization ability in future work. 6 Acknowledgment This work is in part supported by GRF 14205719, TRS T41-603/20-R, Centre for Perceptual and Interactive Intelligence, and CUHK Interdisciplinary AI Research Institute; and in part supported by NTU NAP, MOE Ac RF Tier 2 (T2EP20221-0033), and under the RIE2020 Industry Alignment Fund Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner(s). [1] Chaitanya Ahuja, Dong Won Lee, Ryo Ishii, and Louis-Philippe Morency. No gestures left behind: Learning relationships between spoken language and freeform gestures. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, pages 1884 1895, 2020. 2, 3, 7 [2] Chaitanya Ahuja, Dong Won Lee, Yukiko I Nakano, and Louis-Philippe Morency. Style transfer for co-speech gesture animation: A multi-speaker conditional-mixture approach. In European Conference on Computer Vision, pages 248 265. Springer, 2020. 3, 7 [3] Chaitanya Ahuja and Louis-Philippe Morency. Language2pose: Natural language grounded pose forecasting. In 2019 International Conference on 3D Vision (3DV), pages 719 728. IEEE, 2019. 2, 3 [4] Simon Alexanderson, Gustav Eje Henter, Taras Kucherenko, and Jonas Beskow. Style-controllable speechdriven gesture synthesis using normalising flows. In Computer Graphics Forum, volume 39, pages 487 496. Wiley Online Library, 2020. 2 [5] Guha Balakrishnan, Amy Zhao, Adrian V Dalca, Fredo Durand, and John Guttag. Synthesizing images of humans in unseen poses. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8340 8348, 2018. 2 [6] Uttaran Bhattacharya, Nicholas Rewkowski, Abhishek Banerjee, Pooja Guhan, Aniket Bera, and Dinesh Manocha. Text2gestures: A transformer-based network for generating emotive body gestures for virtual agents. In 2021 IEEE Virtual Reality and 3D User Interfaces (VR), pages 1 10. IEEE, 2021. 3 [7] Judee K Burgoon, Thomas Birk, and Michael Pfau. Nonverbal behaviors, persuasion, and credibility. Human communication research, 17(1):140 169, 1990. 1 [8] Chen Cao, Qiming Hou, and Kun Zhou. Displaced dynamic expression regression for real-time facial tracking and animation. ACM Transactions on graphics (TOG), 33(4):1 10, 2014. 3 [9] Zhe Cao, Gines Hidalgo, Tomas Simon, Shih-En Wei, and Yaser Sheikh. Openpose: realtime multiperson 2d pose estimation using part affinity fields. IEEE transactions on pattern analysis and machine intelligence, 43(1):172 186, 2019. 1, 7 [10] Justine Cassell, David Mc Neill, and Karl-Erik Mc Cullough. Speech-gesture mismatches: Evidence for one underlying representation of linguistic and nonlinguistic information. Pragmatics & cognition, 7(1):1 34, 1999. 1 [11] Justine Cassell, Catherine Pelachaud, Norman Badler, Mark Steedman, Brett Achorn, Tripp Becket, Brett Douville, Scott Prevost, and Matthew Stone. Animated conversation: rule-based generation of facial expression, gesture & spoken intonation for multiple conversational agents. In Proceedings of the 21st annual conference on Computer graphics and interactive techniques, pages 413 420, 1994. 1 [12] Justine Cassell, Hannes Högni Vilhjálmsson, and Timothy Bickmore. Beat: the behavior expression animation toolkit. In Life-Like Characters, pages 163 185. Springer, 2004. 1 [13] Caroline Chan, Shiry Ginosar, Tinghui Zhou, and Alexei A Efros. Everybody dance now. In IEEE International Conference on Computer Vision (ICCV), 2019. 2 [14] Lele Chen, Ross K Maddox, Zhiyao Duan, and Chenliang Xu. Hierarchical cross-modal talking face generation with dynamic pixel-wise loss. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7832 7841, 2019. 3 [15] Vasileios Choutas, Georgios Pavlakos, Timo Bolkart, Dimitrios Tzionas, and Michael J. Black. Monocular expressive body regression through body-driven attention. In European Conference on Computer Vision (ECCV), 2020. 1 [16] Daniel Cudeiro, Timo Bolkart, Cassidy Laidlaw, Anurag Ranjan, and Michael J Black. Capture, learning, and synthesis of 3d speaking styles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10101 10111, 2019. 1, 2 [17] Jan P De Ruiter, Adrian Bangerter, and Paula Dings. The interplay between gesture and speech in the production of referring expressions: Investigating the tradeoff hypothesis. Topics in cognitive science, 4(2):232 248, 2012. 1 [18] Ylva Ferstl and Rachel Mc Donnell. Investigating the use of recurrent motion modelling for speech gesture generation. In Proceedings of the 18th International Conference on Intelligent Virtual Agents, pages 93 98, 2018. 1, 2 [19] Ylva Ferstl, Michael Neff, and Rachel Mc Donnell. Adversarial gesture generation with realistic gesture phasing. Computers & Graphics, 89:117 130, 2020. 3 [20] Zhenglin Geng, Chen Cao, and Sergey Tulyakov. 3d guided fine-grained face manipulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9821 9830, 2019. 3 [21] Shiry Ginosar, Amir Bar, Gefen Kohavi, Caroline Chan, Andrew Owens, and Jitendra Malik. Learning individual styles of conversational gesture. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3497 3506, 2019. 1, 2, 3, 7, 8 [22] Ikhsanul Habibie, Weipeng Xu, Dushyant Mehta, Lingjie Liu, Hans-Peter Seidel, Gerard Pons-Moll, Mohamed Elgharib, and Christian Theobalt. Learning speech-driven 3d conversational gestures from video. ar Xiv preprint ar Xiv:2102.06837, 2021. [23] Dai Hasegawa, Naoshi Kaneko, Shinichi Shirakawa, Hiroshi Sakuta, and Kazuhiko Sumi. Evaluation of speech-to-gesture generation using bi-directional lstm network. In Proceedings of the 18th International Conference on Intelligent Virtual Agents, pages 79 86, 2018. 2, 3 [24] Alexander Hermans, Lucas Beyer, and Bastian Leibe. In defense of the triplet loss for person reidentification. ar Xiv preprint ar Xiv:1703.07737, 2017. 9 [25] Carlos T. Ishi, Daichi Machiyashiki, Ryusuke Mikata, and Hiroshi Ishiguro. A speech-driven hand gesture generation method and evaluation in android robots. IEEE Robotics and Automation Letters, 3(4):3757 3764, 2018. 2, 3 [26] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and super-resolution. In European conference on computer vision, pages 694 711. Springer, 2016. 9 [27] Adam Kendon. Gesture: Visible action as utterance. Cambridge University Press, 2004. 2, 7 [28] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Yoshua Bengio and Yann Le Cun, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. 7 [29] Taras Kucherenko, Patrik Jonell, Sanne van Waveren, Gustav Eje Henter, Simon Alexandersson, Iolanda Leite, and Hedvig Kjellström. Gesticulator: A framework for semantically-aware speech-driven gesture generation. In Proceedings of the 2020 International Conference on Multimodal Interaction, ICMI 20, page 242 250, New York, NY, USA, 2020. Association for Computing Machinery. 2 [30] Jing Li, Di Kang, Wenjie Pei, Xuefei Zhe, Ying Zhang, Zhenyu He, and Linchao Bao. Audio2gestures: Generating diverse gestures from speech audio with conditional variational autoencoders. ar Xiv preprint ar Xiv:2108.06720, 2021. 2, 3 [31] Ruilong Li, Shan Yang, David A Ross, and Angjoo Kanazawa. Learn to dance with aist++: Music conditioned 3d dance generation. ar Xiv preprint ar Xiv:2101.08779, 2021. 8 [32] Xian Liu, Rui Qian, Hang Zhou, Di Hu, Weiyao Lin, Ziwei Liu, Bolei Zhou, and Xiaowei Zhou. Visual sound localization in the wild by cross-modal interference erasing. ar Xiv preprint ar Xiv:2202.06406, 2022. 3 [33] Xian Liu, Qianyi Wu, Hang Zhou, Yinghao Xu, Rui Qian, Xinyi Lin, Xiaowei Zhou, Wayne Wu, Bo Dai, and Bolei Zhou. Learning hierarchical cross-modal association for co-speech gesture generation. ar Xiv preprint ar Xiv:2203.13161, 2022. 2, 3, 7, 8 [34] Xian Liu, Yinghao Xu, Qianyi Wu, Hang Zhou, Wayne Wu, and Bolei Zhou. Semantic-aware implicit neural audio-driven video portrait generation. ar Xiv preprint ar Xiv:2201.07786, 2022. 1 [35] David Mc Neill. Hand and mind. De Gruyter Mouton, 2011. 1 [36] Evonne Ng, Hanbyul Joo, Liwen Hu, Hao Li, Trevor Darrell, Angjoo Kanazawa, and Shiry Ginosar. Learning to listen: Modeling non-deterministic dyadic facial motion. ar Xiv preprint ar Xiv:2204.08451, 2022. 3 [37] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32:8026 8037, 2019. 7 [38] KR Prajwal, Rudrabha Mukhopadhyay, Vinay P Namboodiri, and CV Jawahar. A lip sync expert is all you need for speech to lip generation in the wild. In Proceedings of the 28th ACM International Conference on Multimedia, pages 484 492, 2020. 3 [39] Shenhan Qian, Zhi Tu, Yi Hao Zhi, Wen Liu, and Shenghua Gao. Speech drives templates: Co-speech gesture synthesis with learned templates. ar Xiv preprint ar Xiv:2108.08020, 2021. 1, 2, 3, 7, 8 [40] Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. Language models are unsupervised multitask learners. Open AI blog, 1(8):9, 2019. 2 [41] Alexander Richard, Michael Zollhöfer, Yandong Wen, Fernando De la Torre, and Yaser Sheikh. Meshtalk: 3d face animation from speech using cross-modality disentanglement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1173 1182, 2021. 3 [42] Najmeh Sadoughi, Yang Liu, and Carlos Busso. Msp-avatar corpus: Motion capture recordings to study the role of discourse functions in the design of intelligent virtual agents. In 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), volume 7, pages 1 6. IEEE, 2015. 2 [43] Aliaksandr Siarohin, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. Animating arbitrary objects via deep motion transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2377 2386, 2019. 3 [44] Aliaksandr Siarohin, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. First order motion model for image animation. Advances in Neural Information Processing Systems, 32, 2019. 2 [45] Aliaksandr Siarohin, Oliver J Woodford, Jian Ren, Menglei Chai, and Sergey Tulyakov. Motion representations for articulated animation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13653 13662, 2021. 2, 3, 4, 7, 9 [46] Li Siyao, Weijiang Yu, Tianpei Gu, Chunze Lin, Quan Wang, Chen Qian, Chen Change Loy, and Ziwei Liu. Bailando: 3d dance generation by actor-critic gpt with choreographic memory. ar Xiv preprint ar Xiv:2203.13055, 2022. 3, 6 [47] Michael Studdert-Kennedy. Hand and mind: What gestures reveal about thought. Language and Speech, 37(2):203 209, 1994. 2, 7 [48] Kenta Takeuchi, Souichirou Kubota, Keisuke Suzuki, Dai Hasegawa, and Hiroshi Sakuta. Creating a gesture-speech dataset for speech-based automatic gesture generation. In International Conference on Human-Computer Interaction, pages 198 202. Springer, 2017. 1, 2 [49] Taoran Tang, Jia Jia, and Hanyang Mao. Dance with melody: An lstm-autoencoder approach to musicoriented dance synthesis. In Proceedings of the 26th ACM international conference on Multimedia, pages 1598 1606, 2018. 6 [50] Justus Thies, Michael Zollhofer, Marc Stamminger, Christian Theobalt, and Matthias Nießner. Face2face: Real-time face capture and reenactment of rgb videos. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2387 2395, 2016. 3 [51] Jackson Tolins, Kris Liu, Yingying Wang, Jean E Fox Tree, Marilyn Walker, and Michael Neff. A multimodal motion-captured corpus of matched and mismatched extravert-introvert conversational pairs. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 16), pages 3469 3476, 2016. 2 [52] Lloyd N Trefethen and David Bau III. Numerical linear algebra, volume 50. Siam, 1997. 5 [53] Aaron Van Den Oord, Oriol Vinyals, et al. Neural discrete representation learning. Advances in neural information processing systems, 30, 2017. 5 [54] Susanne Van Mulken, Elisabeth Andre, and Jochen Müller. The persona effect: how substantial is it? In People and computers XIII, pages 53 66. Springer, 1998. 1 [55] Ekaterina Volkova, Stephan De La Rosa, Heinrich H Bülthoff, and Betty Mohler. The mpi emotional body expressions database for narrative scenarios. Plo S one, 9:e113647, 2014. 2 [56] P. Wagner, Z. Malisz, and S. Kopp. Gesture and speech in interaction: An overview. Speech Communication, 57:209 232, 2014. 1 [57] Olivia Wiles, A Koepke, and Andrew Zisserman. X2face: A network for controlling face generation using images, audio, and pose codes. In Proceedings of the European conference on computer vision (ECCV), pages 670 686, 2018. 3 [58] Jason R Wilson, Nah Young Lee, Annie Saechao, Sharon Hershenson, Matthias Scheutz, and Linda Tickle Degnen. Hand gestures and verbal acknowledgments improve human-robot rapport. In International Conference on Social Robotics, pages 334 344. Springer, 2017. 1 [59] Jing Xu, Wei Zhang, Yalong Bai, Qibin Sun, and Tao Mei. Freeform body motion generation from speech. ar Xiv preprint ar Xiv:2203.02291, 2022. 3 [60] Yanzhe Yang, Jimei Yang, and Jessica Hodgins. Statistics-based motion synthesis for social conversations. In Computer Graphics Forum, volume 39, pages 201 212. Wiley Online Library, 2020. 2 [61] Payam Jome Yazdian, Mo Chen, and Angelica Lim. Gesture2vec: Clustering gestures using representation learning methods for co-speech gesture generation, 2022. 2 [62] Youngwoo Yoon, Bok Cha, Joo-Haeng Lee, Minsu Jang, Jaeyeon Lee, Jaehong Kim, and Geehyuk Lee. Speech gesture generation from the trimodal context of text, audio, and speaker identity. ACM Transactions on Graphics (TOG), 39(6):1 16, 2020. 1, 2, 3, 7, 8 [63] Youngwoo Yoon, Woo-Ri Ko, Minsu Jang, Jaeyeon Lee, Jaehong Kim, and Geehyuk Lee. Robots learn social skills: End-to-end learning of co-speech gesture generation for humanoid robots. In 2019 International Conference on Robotics and Automation (ICRA), pages 4303 4309. IEEE, 2019. 1, 2, 3 [64] Egor Zakharov, Aliaksandra Shysheya, Egor Burkov, and Victor Lempitsky. Few-shot adversarial learning of realistic neural talking head models. In Proceedings of the IEEE/CVF international conference on computer vision, pages 9459 9468, 2019. 3 [65] Hang Zhou, Yasheng Sun, Wayne Wu, Chen Change Loy, Xiaogang Wang, and Ziwei Liu. Pose-controllable talking face generation by implicitly modularized audio-visual representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021. 1 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] The limitations are discussed in Section 5. More details and analysis are provided in the supplemental document. (c) Did you discuss any potential negative societal impacts of your work? [Yes] The potential negative societal impacts are discussed in Section 5. More ethical considerations with potential measures are provided in the supplemental document. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] The assumptions of the unique cholesky decomposition theorem are the real symmetric positive definite characteristic of the matrix, which are stated in the Theorem 1. (b) Did you include complete proofs of all theoretical results? [Yes] The proofs are included in the supplemental document. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [N/A] We include very detailed implementation instructions in Section 4.1 to help reproduce the main experimental results. Besides, although the code and data are not included, as promised we will make the code, models and data publicly available. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] They are specified in the Implementation Details. Please see Section 4.1. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] Please see the supplemental document for evaluation results with error bars. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] They are specified in the Implementation Details. Please see Section 4.1. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] The citations are all included in the reference, and the detailed licenses are included in the supplemental document. (b) Did you mention the license of the assets? [Yes] The citations are all included in the reference, and the detailed licenses are included in the supplemental document. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] The citations are all included in the reference, and the detailed licenses are included in the supplemental document. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] Our work and data are based on publicly available dataset and used for the academic use exclusively. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] The discussions are elaborated in the supplemental document, where all the data and models are for the academic use only. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [Yes] The detailed user study instructions with details are provided in the supplemental document. (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [Yes] Since the user study merely requires the participants to watch and evaluate the co-speech gesture generation results, there is no potential participant risk. (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [Yes] We pay the user study participants for their efforts. The details are provided in the supplemental document.