# crosstalk_reduction__12a2d458.pdf Cross-Talk Reduction Zhong-Qiu Wang1 , Anurag Kumar2 and Shinji Watanabe3 1Southern University of Science and Technology, China 2Meta Reality Labs Research, USA 3Carnegie Mellon University, USA wang.zhongqiu41@gmail.com While far-field multi-talker mixtures are recorded, each speaker can wear a close-talk microphone so that close-talk mixtures can be recorded at the same time. Although each close-talk mixture has a high signal-to-noise ratio (SNR) of the wearer, it has a very limited range of applications, as it also contains significant cross-talk speech by other speakers and is not clean enough. In this context, we propose a novel task named cross-talk reduction (CTR) which aims at reducing cross-talk speech, and a novel solution named CTRnet which is based on unsupervised or weakly-supervised neural speech separation. In unsupervised CTRnet, close-talk and far-field mixtures are stacked as input for a DNN to estimate the close-talk speech of each speaker. It is trained in an unsupervised, discriminative way such that the DNN estimate for each speaker can be linearly filtered to cancel out the speaker s cross-talk speech captured at other microphones. In weaklysupervised CTRnet, we assume the availability of each speaker s activity timestamps during training, and leverage them to improve the training of unsupervised CTRnet. Evaluation results on a simulated two-speaker CTR task and on a real-recorded conversational speech separation and recognition task show the effectiveness and potential of CTRnet. 1 Introduction While far-field mixtures of multiple speakers are recorded, the close-talk mixture of each speaker is often recorded at the same time, by placing a microphone close to each target speaker (e.g., in the AMI [Carletta et al., 2006], CHi ME [Barker et al., 2018], Ali Meeting [Yu et al., 2022], and MISP [Wang et al., 2023b] setup).1 See Fig. 1 for an illustration. Each close-talk mixture consists of close-talk speech, cross-talk speech and environmental noises, all of which are reverberant. Inside close-talk mixtures, the closetalk speech usually has a strong energy, and inside close-talk 1Close-talk mixtures are almost always recorded in conversational speech separation and recognition datasets, since it is much easier for humans to annotate transcriptions and speaker-activity timestamps based on close-talk mixtures than far-field mixtures. Reverberant room Far-field microphone array Cross-talk speech Close-talk mic #1 Close-talk mic #2 Close-talk speech Figure 1: Task illustration. Best viewed in color. speech, the direct-path signal of the target speaker is typically much stronger than reverberation. However, besides close-talk speech, close-talk mixtures often contain significant cross-talk speech produced by other speakers, especially if the other speakers, while talking, are spatially close to the target speaker. The contamination of cross-talk speech dramatically limits the application range of close-talk mixtures. For example, they are seldomly exploited as supervisions for training far-field speech separation models or leveraged as reference signals for evaluating speech separation models. It is a common perception that they are not suitable for these purposes as they are not clean enough [Barker et al., 2018; Haeb-Umbach et al., 2019; Wisdom et al., 2020; Watanabe et al., 2020; Sivaraman et al., 2022; Aralikatti et al., 2023; Cornell et al., 2023; Leglaive et al., 2023]. This paper aims at reducing cross-talk speech and separating close-talk speech in close-talk mixtures. We name this task cross-talk reduction. Solving this task could enable many applications. For example, we can (a) leverage separated close-talk speech as pseudo-labels for training supervised farfield separation models; (b) use it as reference signals to evaluate the separation results of far-field separation models; and (c) present it rather than close-talk mixtures to annotators to reduce their annotation efforts. One possible approach for cross-talk reduction is supervised speech separation [Wang and Chen, 2018; Yin et al., 2020; Jenrungrot et al., 2020; Tang et al., 2020; Nachmani et al., 2020; Rixen and Renz, 2022; Wang et al., 2023c], where synthetic pairs of clean close-talk speech and close-talk (and far-field) mixtures are first synthesized via room simulation and then used to train supervised learning based models to predict the clean close-talk speech from its paired mixtures. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) The trained models, however, are known to suffer from severe generalization issues, as the simulated data used for training is often and inevitably mismatched with real-recorded test data [Wang and Chen, 2018; Pandey and Wang, 2020; Tzinis et al., 2021; Zhang et al., 2021; Tzinis et al., 2022a; Tzinis et al., 2022b; Leglaive et al., 2023]. Given only paired real-recorded far-field and close-talk mixtures, we propose to tackle cross-talk reduction by training unsupervised speech separation models directly on the real-recorded mixtures. This way, we can avoid the effort of data simulation, and the issues incurred when using simulated data. We point out that many existing unsupervised separation algorithms could be employed for cross-talk reduction. In this paper, we design a potentially better unsupervised algorithm by leveraging the fact that, for each speaker, the close-talk speech in close-talk mixtures has a strong energy, which can provide an informative cue about what the target speech is. Our idea is that we can (a) roughly separate the close-talk speech from each close-talk mixture (this could be achieved with a reasonable performance since the SNR of close-talk speech is usually high); and (b) reverberate the separated close-talk speech of each speaker via linear filtering to identify, and then cancel out, the speaker s cross-talk speech captured by the close-talk microphones of the other speakers. Based on this idea, we propose CTRnet, where, during training, close-talk and far-field mixtures are stacked as input for a deep neural network (DNN) to first produce an estimate for each close-talk speech, and then the cross-talk speech in each close-talk mixture is identified (and cancelled out) by linearly filtering the DNN estimates via a linear prediction algorithm such that the filtering results of all the speakers can add up to the mixture. This paper makes three major contributions: We propose a novel task, cross-talk reduction. We propose a novel solution, unsupervised CTRnet. We further propose weakly-supervised CTRnet, where, during training, speaker-activity timestamps are leveraged as a weak supervision to improve unsupervised CTRnet. Evaluation results on a simulated CTR task and on a realrecorded conversational speech separation and recognition task show the effectiveness and potential of CTRnet. A sound demo is provided in the link below.2 2 Related Work To the best of our knowledge, we are the first studying crosstalk reduction. There are tasks with similar names such as cross-talk cancellation [Bleil, 2023] which deals with sound field manipulation in spatial audio, and cross-talk suppression [Tripathi et al., 2022] in circuits design. They are not related to our task and algorithms. CTRnet builds upon the UNSSOR algorithm [Wang and Watanabe, 2023], leveraging UNSSOR s capability at unsupervised speech separation. UNSSOR is designed to perform unsupervised separation based on far-field mixtures, assuming compact microphone arrays, while unsupervised CTRnet performs separation based on not only far-field but also closetalk mixtures, dealing with distributed-array cases for a different task: cross-talk reduction. We further propose weakly- 2See https://zqwang7.github.io/demos/CTRnet demo/index.html supervised CTRnet, which leverages speaker-activity timestamps as weak supervision and shows strong performance on real-recorded data. In contrast, UNSSOR deals with unsupervised separation and is only validated on simulated data. 3 Problem Formulation In a reverberant enclosure with C speakers (each wearing a close-talk microphone) and a P-microphone far-field array (see Fig. 1 for an illustration), each of the recorded closedtalk and far-field mixtures can be respectively formulated in the short-time Fourier transform (STFT) domain as follows: Yc(t, f) = XC c =1 Xc(c , t, f) + εc(t, f), (1) Yp(t, f) = XC c=1 Xp(c, t, f) + εp(t, f), (2) where t indexes T frames, f indexes F frequencies, c indexes C speakers (and close-talk microphones), and p indexes P far-field microphones. Yc(t, f), Xc(c , t, f) and εc(t, f) in (1) respectively denote the STFT coefficients of the close-talk mixture, reverberant image of speaker c , and non-speech signals captured by the close-talk microphone of speaker c at time t and frequency f. Notice that we use subscript c to index the C close-talk microphones. Similarly, Yp(t, f), Xp(c, t, f) and εp(t, f) in (2) respectively denote the STFT coefficients of the far-field mixture, reverberant image of speaker c, and non-speech signals captured at far-field microphone p. In the rest of this paper, we refer to the corresponding spectrograms when dropping p, c, t or f in notations. ε is assumed as a weak and stationary noise term. Based on the close-talk and far-field mixtures, we aim at estimating Xc(c), the close-talk speech of each speaker c.3 Xc(c) is very clean and can be roughly viewed as the dry source signal.4 While speaker c is speaking, Xc(c) inside Yc is usually stronger than the cross-talk speech Xc(c ) by any other speaker c ( = c). With these understandings, we formulate the above physical models as (3) and (4), and formulate cross-talk reduction as a blind deconvolution problem in (5). In detail, let Z(c) = Xc(c) denote the close-talk speech of speaker c, we re-formulate (1) as Yc(t, f) = Z(c, t, f) + c =1,c =c Xc(c , t, f) + εc(t, f) = Z(c, t, f) + gc(c , f)H e Z(c , t, f) + ε c(t, f), (3) where, in the second row, e Z(c , t, f) = [Z(c , t A + 1, f), ..., Z(c , t, f), ..., Z(c , t + B, f)]T CA+B stacks a window of A + B T-F units, gc(c , f) CA+B is a subband filter, and ( )H computes Hermitian transpose. In (3), we leverage narrow-band approximation [Talmon et al., 2009; 3In Xc(c), symbol c in the subscript denotes the close-talk microphone of speaker c, and c in the parenthesis indicates that the notation represents the reverberant image of speaker c. 4Some reverberation of the source signal still exists in Xc(c), but it is much weaker than the direct-path signal. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Close-talk mic 𝑐, where 𝑐 1, , 𝐶 '𝑋!"#$ 1 , '𝑋!"#$ 𝐶 , , *𝑍𝑐, *𝑍𝐶 *𝑍1 , , , Far-field mic 𝑝, where 𝑝 1, , 𝑃 '𝑋'"#$ 1 , '𝑋'"#$ 𝐶 , *𝑍(𝐶) *𝑍1 , , '𝑋!"#$ 𝑐+ 1 , *𝑍𝑐+ 1 , *𝑍𝑐 1 , 𝑌! for 𝑐 1, , 𝐶; 𝑌' for 𝑝 1, , 𝑃 *𝑍1 , , *𝑍𝑐, , *𝑍𝐶 Figure 2: Illustration of unsupervised CTRnet (see first paragraph of Section 4 for detailed description). Gannot et al., 2017; Wang, 2024] to approximate Xc(c ) (i.e., cross-talk speech of speaker c captured by the close-talk microphone of speaker c) as a linear convolution between a filter gc(c , ) and the close-talk speech Z(c ) of speaker c , i.e., Xc(c , t, f) gc(c , f)H e Z(c , t, f). This is reasonable since Z(c ) can be roughly viewed as the dry source signal, and, in this case, we can interpret the filter as the acoustic transfer function from speaker c to close-talk microphone c. In (3), ε contains the non-speech signal ε and absorbs the modeling errors incurred by using narrow-band approximation. Similarly, for each far-field mixture, we re-formulate (2) as Yp(t, f) = XC c=1 gp(c, f)H e Z(c, t, f) + ε p(t, f), (4) where gp(c, f) can be interpreted as the acoustic transfer function from speaker c to far-field microphone p. Assuming that ε is weak, time-invariant and Gaussian, we can realize cross-talk reduction by solving, e.g., the problem below, which finds the sources, Z( , , ), and filters, g ( , ), most consistent with the physical models in (3) and (4): argmin g ( , ),Z( , , ) Yc(t, f) Z(c, t, f) gc(c , f)H e Z(c , t, f) 2 c=1 gp(c, f)H e Z(c, t, f) 2 , (5) where | | computes magnitude. This is a blind deconvolution problem [Levin et al., 2011] in artificial intelligence and machine learning, which is highly non-convex and not solvable if no prior knowledge is assumed about the sources or the filters, since all of them are unknown and need to be estimated. Motivated by the classic expectation-maximization algorithm in machine learning [Bishop, 2006] and the recent UNSSOR algorithm [Wang and Watanabe, 2023], we propose to tackle this problem by leveraging unsupervised and weaklysupervised deep learning, where (a) a DNN is trained to first produce an estimate for each source (hence modeling source priors); (b) with the sources estimated, filter estimation in (5) then becomes a much simpler linear regression problem, where a closed-form solution exists and can be readily computed; and (c) loss functions similar to the objective in (5) can be designed to regularize the DNN estimates to have them respectively approximate the close-talk speech of each speaker. 4 Unsupervised CTRnet Fig. 2 illustrates unsupervised CTRnet. Given close-talk and far-field mixtures that are reasonably time-synchronized, it stacks close-talk and far-field mixtures as input, and produces, for each speaker c, an estimate ˆZ(c), which, we will show, is regularized to be an estimate for close-talk speech Xc(c). Since Xc(c) can be viewed as the dry source signal, we can reverberate its estimate ˆZ(c) via a linear filtering algorithm named forward convolutive prediction (FCP) [Wang et al., 2021b] to identify speaker c s cross-talk speech at the close-talk microphones of the other speakers, as well as its reverberant images at far-field microphones. We realize this identification by training the DNN to optimize a combination of two loss functions, which encourage the DNN estimates and their linear filtering results to sum up to the mixture at each microphone (so that the linear filtering results can correctly identify the cross-talk speech which we aim at reducing). At run time, ˆZ(c) is used as the estimate of the closetalk speech Xc(c) for each speaker c. This section describes the DNN configurations, loss functions, and FCP filtering. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) 4.1 DNN Configurations We stack the real and imaginary (RI) components of closetalk and far-field mixtures as input features for the DNN to predict ˆZ(c) for each speaker c. We can use complex spectral mapping [Tan and Wang, 2020; Wang et al., 2021a], where the DNN is trained to directly predict the RI components of ˆZ(c) for each speaker c, or complex ratio masking [Williamson et al., 2016], where the DNN is trained to predict the RI components of a complex-valued mask ˆ M(c) for each speaker c and ˆZ(c) is then computed via ˆZ(c) = ˆ M(c) Yc, with denoting point-wise multiplication. The details of the DNN architecture are provided in Section 6.3, and the loss functions in the next subsection. 4.2 Mixture-Constraint Loss We propose the following mixture-constraint (MC) loss: c=1 LMC,c + α XP p=1 LMC,p, (6) where LMC,c is the MC loss at close-talk microphone c, LMC,p at far-field microphone p, and α R>0 a weighting term. Following the physical model in (3) and the first term in (5), at close-talk microphone c we define LMC,c as t,f F Yc(t, f), ˆYc(t, f) t,f F Yc(t, f), ˆXc(c, t, f) + c =1,c =c ˆXFCP c (c , t, f) t,f F Yc(t, f), ˆZ(c, t, f)+ c =1,c =c ˆgc(c , f)H eˆZ(c , t, f) , (7) where, in the third row, eˆZ(c , t, f) = [ ˆZ(c , t I + 1, f), ..., ˆZ(c , t, f), ..., ˆZ(c , t + J, f)]T CI+J stacks a window of I + J T-F units, and ˆgc(c , f) CI+J is a subband FCP filter to be described later in Section 4.3. From row 2 to 3, we constrain the DNN estimate ˆZ(c) of each speaker c to be a gainand time-aligned estimate of the close-talk speech (i.e., ˆXc(c) = ˆZ(c)); and, for the reverberant image of each of the other speakers (i.e., cross-talk speech Xc(c ) with c = c), we approximate it by linearly filtering ˆZ(c ) (i.e., ˆXFCP c (c , t, f) = ˆgc(c , f)H eˆZ(c , t, f)). Since ˆZ(c) is constrained to approximate Xc(c), which can be viewed as the dry source signal, it is very reasonable to linearly filter ˆZ(c) to estimate the image of speaker c at the close-talk microphone of a different speaker and at far-field microphones. The estimated speaker images are then added together to reconstruct the mixture: ˆYc = ˆXc(c) + PC c =1,c =c ˆXFCP c (c ), which is reasonable under the assumption that the non-speech signal ε is weak. Finally, we compute a loss between the reconstructed mixture ˆYc and the mixture Yc, by using a function F( , ), which computes an absolute loss on the estimated RI components and their magnitude [Wang et al., 2023c]: F Yc(t, f), ˆYc(t, f) = P O Ω O(Yc(t, f)) O( ˆYc(t, f)) P t ,f |Yc(t , f )| , where Ω= {R, I, A} is a set of functions with R( ) extracting the real part, I( ) the imaginary part and A( ) the magnitude of a complex number, and the denominator balances the losses across different microphones and training mixtures. Similarly, at far-field microphone p we define LMC,p as t,f F Yp(t, f), ˆYp(t, f) t,f F Yp(t, f), XC c=1 ˆXFCP p (c, t, f) t,f F Yp(t, f), c=1 ˆgp(c, f)H eˆZ(c, t, f) , (8) where we linearly filter the DNN estimate ˆZ(c) for each speaker c using ˆgp(c, f) so that their summation can approximate the mixture Yp captured by far-field microphone p. 4.3 FCP for Filter Estimation To compute LMC, we need to first estimate the linear filters, each of which is the relative transfer function relating the close-talk speech of a speaker to the speaker s reverberant image captured by a distant microphone. Following UNSSOR [Wang and Watanabe, 2023], we employ FCP [Wang et al., 2021b] to estimate them. Assuming speakers are non-moving within each utterance, we estimate the filters by solving the following problem: ˆgr(c, f) = argmin gr(c,f) Yr(t, f) gr(c, f)H eˆZ(c, t, f) 2 ˆλr(c, t, f) , (9) where symbol r is used to index the P far-field and C close- talk microphones, and ˆgr(c, f) and eˆZ(c, t, f) are defined below (7). ˆλ is a weighting term balancing the importance of each T-F unit, and following [Wang et al., 2021b], for each close-talk microphone c and speaker c , it is defined as ˆλc(c , t, f) = ξ max(|Yc|2) + |Yc(t, f)|2, (10) where ξ floors the weighting term and max( ) extracts the maximum value of a power spectrogram; and for each farfield microphone p and speaker c, it is defined as ˆλp(c, t, f) = ξ max(|Yp|2) + Yp(t, f). (11) We compute the weighting term individually for each microphone, considering that each speaker has different energy levels at different microphones. (9) is a weighted linear regression problem, where a closed-form solution can be computed: ˆgr(c, f) = eˆZ(c, t, f)eˆZ(c, t, f)H ˆλr(c, t, f) eˆZ(c, t, f)(Yr(t, f)) ˆλr(c, t, f) , Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) where ( ) computes complex conjugate. We then plug it into (7) and (8), compute the losses, and train the DNN. Although, in (9), ˆZ(c) is filtered to approximate Yr, earlier studies [Wang et al., 2021b] have suggested that the filter- ing result ˆgr(c, f)H eˆZ(c, t, f) would approximate Xr(c, t, f) (see the derivation in Appendix C of [Wang and Watanabe, 2023]), if ˆZ(c) is sufficiently accurate, which can be reasonably satisfied as the close-talk speech in close-talk mixtures already has a high input SNR. The estimated speaker image is named FCP-estimated image [Wang and Watanabe, 2023]: ˆXFCP r (c, t, f) = ˆgr(c, f)H eˆZ(c, t, f). (12) We can hence sum up the FCP-estimated images and compare the summation with Yr in (7) and (8). Notice that the FCP-estimated images at close-talk microphones represent the identified cross-talk speech we aim to reduce. 5 Weakly-Supervised CTRnet We propose to leverage speaker-activity timestamps as a weak supervision to improve the training of unsupervised CTRnet. For each speaker c, its activity timestamps are denoted as a binary vector d(c) {0, 1}N, where N is signal length in samples, and value 1 means that speaker c is active at a sample and 0 otherwise. Note that speaker-activity timestamps are provided in almost every conversational speech recognition dataset. They are annotated by having human annotators listen to close-talk mixtures. This section describes why and how the timestamps can help training CTRnet. 5.1 Motivation In human conversations where concurrent speech naturally happens, speaker overlap is often sparse. See Appendix B for an example. This sparsity poses major difficulties for speech separation [Chen et al., 2020; Cosentino et al., 2020], as the number of concurrent speakers is time-varying. Modern separation models usually assume a maximum number of speakers the models can separate, but the models could overor under-separate the speakers due to errors in speaker counting. Assuming that there are at maximum C speakers in each processing segment, unsupervised CTRnet, in essence, learns to produce C output spectrograms that can be filtered to best explain the recorded C close-talk and P far-field mixtures. One issue is that, in some training segments, if there are fewer than C speakers, unsupervised CTRnet would always overseparate the speakers to obtain a smaller LMC loss.5 Therefore, the knowledge about the accurate number of speakers, which can be provided by speaker-activity timestamps, can be very useful to improve the training of unsupervised CTRnet. Meanwhile, we find that unsupervised CTRnet usually cannot produce zero predictions in the silent ranges of each speaker. In these ranges, the predicted spectrogram for each speaker often contains weak but non-negligible and intelligible signals from the other speakers. This is a common problem in conversational speech separation [Wang et al., 2021a; 5This is similar to unsupervised clustering. Hypothesizing more clusters for clustering can almost always produce a smaller objective but some clusters are split into smaller ones [Bishop, 2006]. Morrone et al., 2023]. In unsupervised CTRnet, this often results in a smaller LMC loss, as the FCP filtering procedure could better approximate close-talk and far-field mixtures by filtering the non-zero predictions in silent regions. With these understandings, we leverage speaker-activity timestamps in the following two ways to better train CTRnet. 5.2 Muting during Training During training, we propose to mute all or part of the frames in each ˆZ(c) based on the speaker-activity timestamps d(c), before computing the filters and loss. In detail, we compute ˆR(c, t, f) = ˆZ(c, t, f) D(c, t) E(c), (13) where D(c, t) {0, 1}, defined based on d(c), is 1 if the STFT window corresponding to frame t contains any active speech samples of speaker c and is 0 otherwise, and E(c) {0, 1} is 1 if d(c) contains at least 0.1 second of active speech in the training segment and is 0 otherwise. Using D(c, t) for muting frames and E(c) for muting speakers, we can avoid using non-zero predictions in silent ranges for FCP. We then replace ˆZ(c) with ˆR(c) in (7), (8) and (9) to compute FCP filters and LMC for DNN training. This way, CTRnet could more robustly reduce cross-talk speech for up to C speakers. 5.3 Speaker-Activity Loss We propose a speaker-activity (SA) loss LSA,c that can encourage the DNN-estimated signal for each speaker c to be zero in silent ranges marked by speaker-activity timestamps: LSA,c = ˆz(c) (1 d(c)) 1 yc (1 d(c)) 1 N d(c) 1 where d(c) and N are defined in the first paragraph of Section 5, 1 computes the L1 norm, ˆz(c) = i STFT( ˆZ(c)) RN is the re-synthesized time-domain signal of ˆZ(c) obtained by applying inverse STFT (i STFT) to ˆZ(c), yc RN is the timedomain close-talk mixture signal of speaker c, and the second term is a scaling factor accounting for the fact that each speaker usually has different length of silence in each training segment. We combine it with LMC in (6) for training: LMC+SA = LMC + β LSA = LMC + β c=1 LSA,c, (15) where β R>0 is a weighting term. LMC penalizes prediction errors in non-silent ranges of each speaker when muting is used, while LSA penalizes predictions in silent ranges. 6 Experimental Setup There are no existing studies on cross-talk reduction. We first evaluate unsupervised CTRnet on a simulated dataset named SMS-WSJ-FF-CT, where clean signals can be available for metric computation, and then evaluate weakly-supervised CTRnet on a real-recorded conversational automatic speech recognition (ASR) dataset named CHi ME-7, using ASR metrics for evaluation. This section describes the datasets, system setups, comparison systems, and evaluation metrics. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) 6.1 SMS-WSJ-FF-CT and Evaluation Setup SMS-WSJ-FF-CT, with FF meaning far-field and CT close-talk, is built upon a simulated dataset named SMSWSJ [Drude et al., 2019] which consists of 2-speaker fullyoverlapped noisy-reverberant mixtures, by simulating a closetalk microphone for each speaker. Fig. 1 shows this setup. This evaluation serves as a proof of concept to validate whether CTRnet can work well in ideal cases, where the hypothesized physical models in (3) and (4) are largely satisfied. SMS-WSJ [Drude et al., 2019], a popular corpus for evaluating 2-speaker separation algorithms in reverberant conditions, has 33, 561 ( 87.4 h), 982 ( 2.5 h) and 1, 332 ( 3.4 h) 2-speaker mixtures for training, validation and testing. The clean speech is sampled from the WSJ0 and WSJ1 corpus. The simulated far-field microphone array has 6 microphones uniformly placed on a circle with a diameter of 20 cm. For each mixture, the speaker-to-array distance is sampled from the range [1.0, 2.0] m, and the reverberation time (T60) from [0.2, 0.5] s. A weak white noise is added to simulate microphone self-noises, at an energy level between the summation of the reverberant speech signals and the noise sampled from the range [20, 30] d B. The sampling rate is 8 k Hz. SMS-WSJ-FF-CT is simulated by adding a close-talk microphone for each of the speakers in each SMS-WSJ mixture. In each mixture, the distance from each speaker to its close-talk microphone is uniformly sampled from the range [0.1, 0.3] m, and the angle from [ π, π]. All the other setup remains unchanged. This way, we can obtain the close-talk mixture of each speaker, and the far-field mixtures are exactly the same as those in SMS-WSJ. Training and Inference of CTRnet. For training, in default we sample an L-second segment from each mixture in each epoch, and the batch size is H. For inference, we feed each test mixture in its full length to CTRnet. Evaluation Metrics. For each speaker c, we use its timedomain close-talk speech corresponding to Xc(c) as the reference signal for evaluation. The evaluation metrics include signal-to-distortion ratio (SDR) [Vincent et al., 2006], scaleinvariant SDR (SI-SDR) [Le Roux et al., 2019], perceptual evaluation of speech quality (PESQ) [Rix et al., 2001], and extended short-time objective intelligibility (e STOI) [Taal et al., 2011]. All of them are widely-used in speech separation. Comparison Systems include signal processing based unsupervised speech separation algorithms, c ACGMM-based spatial clustering (SC) [Boeddeker et al., 2021] and independent vector analysis (IVA) [Sawada et al., 2019]. Both are very popular. We stack close-talk and far-field mixtures as input. For SC, we use a public implementation [Boeddeker, 2019] provided in the pb bss toolkit; and for IVA, we use the torchiva toolkit [Scheibler and Saijo, 2022]. The STFT window and hop sizes are tuned to 128 and 16 ms for SC, and to 256 and 32 ms for IVA. For IVA, we use the Gaussian model in torchiva to model source distribution. A garbage source is used in both models to absorb modeling errors. 6.2 CHi ME-7 and Evaluation Setup To show that CTRnet can work on realistic data, we train and evaluate CTRnet using the real-recorded CHi ME-7 dataset, following the setup of the CHi ME-7 DASR challenge [Cornell et al., 2023]. CHi ME-7, built upon CHi ME-{5,6} [Watanabe et al., 2020; Barker et al., 2018], is a notoriously difficult dataset in conversational speech separation and recognition, mainly due to its realisticness, which is representative of common problems deployed systems could run into in practice, such as microphone synchronization errors, signal clipping, frame dropping, microphone failures, moving arrays, moving speakers, varying degrees of speaker overlap, and challenging environmental noises. So far, the most successful separation algorithm for CHi ME7 is still based on guided source separation (GSS) [Boeddecker et al., 2018], a signal processing algorithm, and supervised DNN-based approaches have nearly no success on this dataset. All the top teams in CHi ME-7 [Wang et al., 2023a; Ye et al., 2023] and in similar challenges or datasets, e.g., Ali Meeting [Yu et al., 2022; Liang et al., 2023] and AMI [Raj et al., 2023], all adopt GSS as the only separation module. This section describes the dataset and our evaluation setup. CHi ME-7 Dataset contains real-recorded conversational sessions, each with 4 participants speaking spontaneously in a domestic, dinner-party scenario, where concurrent speech can naturally happen. Each speaker (participant) wears a binaural close-talk microphone, and there are 6 Kinect devices, each with 4 microphones, placed in a strategic way in the room to record each entire session, which is 1.5 to 2.5 hours long. The recorded close-talk mixtures contain severe cross-talk speech. Realistic noises typical in dinner parties are recorded at the same time along with speech. There are 14 ( 34 h), 2 ( 2 h) and 4 ( 5 h) recorded sessions respectively for training, validation and testing. The sampling rate is 16 k Hz. Experiment Design. Based on the CHi ME-7 dataset, we design an experiment for cross-talk reduction. We consider the mixture recorded at the right ear of each binaural microphone as the close-talk mixture and the one at the left ear as far-field mixture, meaning that we have C = 4 close-talk and P = 4 far-field mixtures. Training and Inference of CTRnet. To train CTRnet, we cut each (long) session into 8-second segments with 50% overlap between consecutive segments, and train CTRnet on these segments. At inference time, we apply the trained CTRnet in a block-wise way to process each (long) session. The final separated speech for each speaker has the same length as the (long) input session. See Appendix A for the details. Evaluation Metrics. With the separated signal of each speaker, following the CHi ME-7 DASR challenge [Cornell et al., 2023] we compute diarization-assigned word error rates (DA-WER). In detail, the separated signal of each speaker is split to short utterances by using oracle speaker-activity timestamps, and the default, pre-trained ASR model provided by the challenge is used to recognize and score each utterance. The pre-trained ASR model6 is a strong end-to-end model based on a transformer encoder/decoder architecture trained with joint CTC/attention, using Wav LM features. Comparison Systems. We use GSS [Boeddecker et al., 2018] for comparison, by following the implementation pro- 6https://huggingface.co/popcornell/chime7 task1 asr1 baseline Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Row Systems I J C P Masking/Mapping α H/L SI-SDR (d B) SDR (d B) PESQ e STOI 0 Unprocessed mixture - - - - - - - 14.7 14.7 2.92 0.875 1a Unsupervised CTRnet 20 0 2 6 Mapping 1.0 4 / 4 24.0 24.3 3.77 0.964 1b Unsupervised CTRnet 30 0 2 6 Mapping 1.0 4 / 4 24.1 24.5 3.81 0.965 1c Unsupervised CTRnet 40 0 2 6 Mapping 1.0 4 / 4 23.9 24.2 3.86 0.965 1d Unsupervised CTRnet 50 0 2 6 Mapping 1.0 4 / 4 23.4 23.7 3.86 0.962 2 Unsupervised CTRnet 30 0 2 6 Masking 1.0 4 / 4 23.4 23.7 3.90 0.964 3a Unsupervised CTRnet 30 0 2 6 Mapping C/P 4 / 4 25.8 26.1 3.91 0.971 3b Unsupervised CTRnet 30 0 2 6 Mapping 1/P 4 / 4 26.0 26.3 3.90 0.971 3c Unsupervised CTRnet 30 0 2 6 Mapping 1/(2 P ) 4 / 4 25.9 26.2 3.90 0.971 3d Unsupervised CTRnet 30 0 2 6 Mapping 1/(3 P ) 4 / 4 25.9 26.2 3.87 0.970 4a Unsupervised CTRnet 29 1 2 6 Mapping 1/P 4 / 4 25.9 26.1 3.90 0.971 4b Unsupervised CTRnet 28 2 2 6 Mapping 1/P 4 / 4 25.7 26.0 3.90 0.971 4c Unsupervised CTRnet 27 3 2 6 Mapping 1/P 4 / 4 25.9 26.2 3.88 0.971 5a Unsupervised CTRnet 30 0 2 3 Mapping 1/P 4 / 4 25.9 26.2 3.87 0.971 5b Unsupervised CTRnet 30 0 2 2 Mapping 1/P 4 / 4 25.9 26.2 3.87 0.970 5c Unsupervised CTRnet 30 0 2 1 Mapping 1/P 4 / 4 24.8 25.0 3.87 0.968 6 Unsupervised CTRnet 30 0 2 6 Mapping 1/P 1 / 26.5 26.8 3.88 0.973 7a SC [Boeddeker, 2019] - - - 6 - - - 1.9 7.1 2.27 0.561 7b IVA [Scheibler and Saijo, 2022] - - - 6 - - - 22.6 23.7 3.66 0.948 Table 1: Averaged separation results of unsupervised CTRnet on SMS-WSJ-FF-CT. vided in CHi ME-7 DASR Challenge.7 GSS is the most popular and effective separation model so far for modern ASR systems. It first performs dereverberation using the weighted prediction algorithm [Nakatani et al., 2010] and then computes a mask-based beamformer for separation by using posterior time-frequency masks estimated by a spatial clustering module guided by oracle speaker-activity timestamps [Boeddecker et al., 2018]. Notice that at run time GSS requires oracle speaker-activity timestamps, while weakly-supervised CTRnet only needs them for training and, once trained, no longer needs them. Another note is that the training data of the pre-trained ASR model provided by the challenge contains GSS-processed signals, and is hence favorable to GSS. 6.3 Miscellaneous Configurations of CTRnet For STFT, the window size is 16 ms, hop size 8 ms, and the square root of the Hann window is used as the analysis window. TF-Grid Net [Wang et al., 2023c] is employed as the DNN architecture. Using the symbols defined in Table I of [Wang et al., 2023c], we set its hyper-parameters to D = 128, B = 4, I = 1, J = 1, H = 192, L = 4 and E = 4 (please do not confuse these symbols with the ones defined in this paper). The model has around 4.8 million parameters. ξ in (10) and (11) is tuned to 10 3. β in (15) is set to 1.0. 7 Evaluation Results and Discussions 7.1 Results on SMS-WSJ-FF-CT Table 1 configures unsupervised CTRnet in various ways and presents the results on SMS-WSJ-FF-CT. Row 0 reports the scores of unprocessed mixtures. The 14.7 d B SI-SDR indicates that the close-talk mixtures are not clean due to the contamination by cross-talk speech. In 1a-1d, we vary the number of FCP filter taps I + J {20, 30, 40, 50} and configure 7See https://github.com/espnet/espnet/blob/master/egs2/chime7 task1/asr1/local/run gss.sh the FCP filters to be causal by setting J = 0. We observe that the setup in 1b obtains the best SI-SDR. In row 2, complex masking rather than mapping is used to obtain ˆZ(c), but the result is not better than 1b. In 3a-3d, we reduce α in (6) from 1.0 to C/P, 1/P, 1/(2 P) and 1/(3 P) so that the loss on close-talk mixtures is emphasized. This is reasonable since we aim at separating close-talk speech, rather than far-field speaker images. From 1b and 3b, we observe that this change leads to clear improvement (e.g., from 24.1 to 26.0 d B SI-SDR). In 4a-4c, we use non-causal FCP filters by increasing J from 0 to 1, 2 and 3, while fixing the total number of filter taps to 30. This does not yield improvements, likely because ˆZ(c) is regularized to be an estimate of the close-talk speech of speaker c and hence the transfer function relating it to speaker c s reverberant images at other microphones should be largely causal. In 5a-5c, we reduce the number of far-field microphones from 6 to 3, 2 and 1. That is, we only use microphone (a) 1; (b) 1 and 4; and (c) 1, 3 and 5 of the far-field six-microphone array to simulate the cases when the far-field array only has a limited number of microphones. Compared with 3b, the performance drops, indicating the benefits of using more far-field mixtures as network input and for loss computation, while the improvement over the mixture scores in row 0 is still large. In default, our models are trained using mini-batches of 4 four-second segments, while, in row 6, we train the model by using a batch size of 1 and using each training mixture in its full length. This improves SI-SDR over 3b, likely because better FCP filters can be computed during training by using all the frames in each mixture. In row 7a and 7b, SC and IVA perform worse than CTRnet. SC does not work well, possibly because, in this distributedmicrophone scenario where each speaker signal can have very different SNRs at different microphones, the target T-F masks at different microphones are significantly different. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Row Systems Muting? I J C P Val. Test 0 Unprocessed mixture - - - 4 - 28.3 27.8 1 Unsupervised CTRnet - 19 1 4 4 22.5 25.1 2 Weakly-supervised CTRnet 19 1 4 4 79.1 73.0 3 Weakly-supervised CTRnet 19 1 4 4 20.5 22.6 4 GSS [Boeddecker et al., 2018] - - - 4 4 26.2 26.6 Table 2: ASR results of CTRnet on CHi ME-7 close-talk mixtures. 7.2 Results on CHi ME-7 Table 2 reports the ASR results of CTRnet on the close-talk mixtures of CHi ME-7. The filter taps I and J are tuned to 19 and 1. Notice that, here, one future tap is used, as the real-recorded data in CHi ME-7 exhibits non-negligible synchronization errors among different microphone signals and we find that allowing one future tap can mitigate the synchronization issues. Complex spectral mapping is used in default. In row 0, the mixture DA-WER is high even though a strong pre-trained ASR model is used for recognition. This is because the close-talk mixtures contain very strong cross-talk speech, which confuses the ASR model on which speaker to recognize. In row 1, using unsupervised CTRnet to reduce cross-talk speech produces clear improvement (from 27.8% to 25.1% DA-WER). In row 3, weakly-supervised CTRnet with muting during training further improves the performance to 22.6%. In row 2, muting is not applied when training weakly-supervised CTRnet, and we observe much worse DAWER. We found that this is because, without using muting, the LSA loss in (15) tends to push ˆZ(c) of each speaker c towards zeros to obtain a smaller loss value. Compared with GSS, weakly-supervised CTRnet, through learning, obtains clearly better DA-WER (i.e., 22.6% vs. 26.6%). In addition, clear improvement is obtained over the unprocessed mixtures (i.e., 22.6% vs. 27.8%). These results indicate the effectiveness of CTRnet for cross-talk reduction on real-recorded data. 8 Conclusion We have proposed a novel task, cross-talk reduction, and a novel solution, CTRnet, with or without leveraging speakeractivity timestamps for model training. A key contribution of this paper, we emphasize, is that the proposed CTRnet can be trained directly on real-recorded pairs of far-field and close-talk mixtures, and is capable of effectively reducing cross-talk speech, especially on the notoriously difficult real-recorded CHi ME-7 data. This contribution suggests a promising way towards addressing a fundamental limitation of real-recorded close-talk speech. That is, the contamination by cross-talk speech, which makes close-talk mixtures not sufficiently clean. With CTRnet producing reasonablygood cross-talk reduction, we expect many applications to be enabled, and we will investigate them in our future work. In closing, another key contribution of this paper, we point out, is that the proposed weakly-supervised deep learning based methodology for blind deconvolution can work well on challenging real-recorded data. This contribution, we believe, would motivate a new stream of research towards real- izing robust neural speech separation in realistic conditions, and generate broader impact beyond CTR and speech separation, especially in many machine learning and artificial intelligence applications where the sensors would not only capture target signals but also interference signals very detrimental to machine perception. A Run-Time Separation of Long Sessions In CHi ME-7, the duration of each recorded session ranges from 1.5 to 2.5 hours. At run time, to separate the close-talk speech of an entire session, we run CTRNet in a block-wise way, using the pseudo-code below at each processing block: Yc := Yc/σc, for c {1, ..., C}, (16) Yp := Yp/σp, for p {1, ..., P}, (17) ˆZ(1), ..., ˆZ(C) = CTRnet(Y1, ..., YC, Y1, ..., YP ), (18) ˆZ(c) := ˆZ(c) σc, for c {1, ..., C}, (19) where σc and σp R are respectively the sample-level standard deviations of the time-domain close-talk mixture yc and far-field mixture yp. After obtaining ˆZ(c) at each block, we stack ˆZ(c) of all the processing blocks and then revert the stacked spectrogram to time domain via inverse STFT. Notice that LMC,c in (7) constrains each close-talk speech estimate ˆZ(c) to have the same gain level as the close-talk mixture Yc, by not filtering ˆZ(c). As a result, in (18), each output ˆZ(c) is expected to have the same gain level as Yc. For our experiments on CHi ME-7, the processing block size is set to 8 seconds, the same as the segment length used during training. We configure the blocks to be slightly overlapped, where we consider the first and the last 0.96 seconds as context, and output the DNN estimates in the center 6.08 (= 8 0.96 0.96) seconds at each block. B Illustration of Sparse Speaker Overlap Fig. 3 illustrates sparse speaker overlap in human conversations. This sparsity is because, for conversations in, e.g., meeting or dinner-party scenarios, people often take turns to speak and tend to not always speak at the same time. As a result, in different (short) processing segments of separation systems, the number of active speakers and the degrees of speaker overlap can vary a lot. Figure 3: Illustration of sparse speaker overlap in human conversations. Best viewed in color. Each colored band means that the corresponding speaker is talking in the time range. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Acknowledgments Experiments of this work used the Bridges2 system at PSC and Delta at NCSA through allocation CIS210014 and IRI120008P from the Advanced Cyberinfrastructure Coordination Ecosystem: Services and Support (ACCESS) program, which is supported by National Science Foundation grants #2138259, #2138286, #2138307, #2137603 and #2138296. References [Aralikatti et al., 2023] Rohith Aralikatti, Christoph Boeddeker, Gordon Wichern, et al. Reverberation as Supervision for Speech Separation. In ICASSP, 2023. [Barker et al., 2018] Jon Barker, Shinji Watanabe, Emmanuel Vincent, et al. The Fifth CHi ME Speech Separation and Recognition Challenge: Dataset, Task and Baselines. In Interspeech, pages 1561 1565, 2018. [Bishop, 2006] C. M. Bishop. Pattern Recognition and Machine Learning. Springer, 2006. [Bleil, 2023] Jacob Bleil. Systematic Development of a Phased Speaker Array for Optimal 3D Audio Reproduction. Ph D thesis, Princeton University, 2023. [Boeddecker et al., 2018] Christoph Boeddecker, Jens Heitkaemper, Joerg Schmalenstroeer, Lukas Drude, Jahn Heymann, et al. Front-End Processing for The CHi ME-5 Dinner Party Scenario. In CHi ME, pages 35 40, 2018. [Boeddeker et al., 2021] Christoph Boeddeker, Frederik Rautenberg, et al. A Comparison and Combination of Unsupervised Blind Source Separation Techniques. In Speech Communication, pages 129 133, 2021. [Boeddeker, 2019] Christoph Boeddeker. pb bss. https://github.com/fgnt/pb bss/blob/master/examples/ mixture model example.ipynb, 2019. Accessed: 202301-18. [Carletta et al., 2006] Jean Carletta, Simone Ashby, Sebastien Bourban, Mike Flynn, Mael Guillemot, et al. The AMI Meeting Corpus: A Pre-Announcement. In Machine Learning for Multimodal Interaction, pages 28 39, 2006. [Chen et al., 2020] Zhuo Chen, Takuya Yoshioka, Liang Lu, Tianyan Zhou, Zhong Meng, Yi Luo, Jian Wu, Xiong Xiao, and Jinyu Li. Continuous Speech Separation: Dataset and Analysis. In ICASSP, pages 7284 7288, 2020. [Cornell et al., 2023] Samuele Cornell, Matthew S. Wiesner, Shinji Watanabe, et al. The CHi ME-7 DASR Challenge: Distant Meeting Transcription with Multiple Devices in Diverse Scenarios. In CHi ME, pages 1 6, 2023. [Cosentino et al., 2020] Joris Cosentino, Manuel Pariente, Samuele Cornell, Antoine Deleforge, et al. Libri Mix: An Open-Source Dataset for Generalizable Speech Separation. In ar Xiv preprint ar Xiv:2005.11262, 2020. [Drude et al., 2019] Lukas Drude, Jens Heitkaemper, Christoph Boeddeker, and Reinhold Haeb-Umbach. SMS-WSJ: Database, Performance Measures, and Baseline Recipe for Multi-Channel Source Separation and Recognition. In ar Xiv preprint ar Xiv:1910.13934, 2019. [Gannot et al., 2017] Sharon Gannot, Emmanuel Vincent, Shmulik Markovich-Golan, and Alexey Ozerov. A Consolidated Perspective on Multi-Microphone Speech Enhancement and Source Separation. IEEE/ACM Trans. Audio, Speech, Lang. Proc., 25:692 730, 2017. [Haeb-Umbach et al., 2019] Reinhold Haeb-Umbach, Shinji Watanabe, Tomohiro Nakatani, Michiel Bacchiani, et al. Speech Processing for Digital Home Assistants: Combining Signal Processing with Deep-Learning Techniques. IEEE Signal Processing Magazine, 36(6):111 124, 2019. [Jenrungrot et al., 2020] Teerapat Jenrungrot, Vivek Jayaram, Steve Seitz, and Ira Kemelmacher-Shlizerman. The Cone of Silence: Speech Separation by Localization. In Neur IPS, 2020. [Le Roux et al., 2019] Jonathan Le Roux, Scott Wisdom, Hakan Erdogan, and John R Hershey. SDR - Half-Baked or Well Done? In ICASSP, pages 626 630, 2019. [Leglaive et al., 2023] Simon Leglaive, L eonie Borne, Efthymios Tzinis, Mostafa Sadeghi, et al. The CHi ME-7 UDASE Task: Unsupervised Domain Adaptation for Conversational Speech Enhancement. In CHi ME, 2023. [Levin et al., 2011] Anat Levin, Yair Weiss, Fredo Durand, and William T. Freeman. Understanding Blind Deconvolution Algorithms. IEEE Trans. Pattern Anal. Mach. Intell., 33(12):2354 2367, 2011. [Liang et al., 2023] Yuhao Liang, Mohan Shi, Fan Yu, Yangze Li, et al. The Second Multi-Channel Multi-Party Meeting Transcription Challenge (M2Me T 2.0): A Benchmark for Speaker-Attributed ASR. In ASRU, 2023. [Morrone et al., 2023] Giovanni Morrone, Samuele Cornell, Desh Raj, Luca Serafini, Enrico Zovato, et al. Low Latency Speech Separation Guided Diarization for Telephone Conversations. In SLT, pages 641 646, 2023. [Nachmani et al., 2020] Eliya Nachmani, Yossi Adi, and Lior Wolf. Voice Separation with An Unknown Number of Multiple Speakers. In ICML, pages 7121 7132, 2020. [Nakatani et al., 2010] Tomohiro Nakatani, Takuya Yoshioka, et al. Speech Dereverberation Based on Variance Normalized Delayed Linear Prediction. IEEE Trans. Audio, Speech, Lang. Proc., 18(7):1717 1731, 2010. [Pandey and Wang, 2020] Ashutosh Pandey and De Liang Wang. On Cross-Corpus Generalization of Deep Learning Based Speech Enhancement. IEEE/ACM Trans. Audio, Speech, Lang. Proc., 28:2489 2499, 2020. [Raj et al., 2023] Desh Raj, Daniel Povey, et al. GPUAccelerated Guided Source Separation for Meeting Transcription. In Interspeech, pages 3507 3511, 2023. [Rix et al., 2001] Antony W. Rix, John G. Beerends, et al. Perceptual Evaluation of Speech Quality (PESQ)-A New Method for Speech Quality Assessment of Telephone Networks and Codecs. In ICASSP, pages 749 752, 2001. [Rixen and Renz, 2022] Joel Rixen and Matthias Renz. SFSRNet: Super-Resolution for Single-Channel Audio Source Separation. In AAAI, 2022. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) [Sawada et al., 2019] Hiroshi Sawada, Nobutaka Ono, Hirokazu Kameoka, Daichi Kitamura, et al. A Review of Blind Source Separation Methods: Two Converging Routes to ILRMA Originating from ICA and NMF. APSIPA Trans. Signal and Info. Proc., 8:1 14, 2019. [Scheibler and Saijo, 2022] Robin Scheibler and Kohei Saijo. Torchiva. https://github.com/fakufaku/torchiva, 2022. Accessed: 2023-01-18. [Sivaraman et al., 2022] Aswin Sivaraman, Scott Wisdom, Hakan Erdogan, and John R. Hershey. Adapting Speech Separation To Real-World Meetings Using Mixture Invariant Training. In ICASSP, pages 686 690, 2022. [Taal et al., 2011] Cees H. Taal, Richard C. Hendriks, et al. An Algorithm for Intelligibility Prediction of Time Frequency Weighted Noisy Speech. IEEE Trans. Audio, Speech, Lang. Proc., 19(7):2125 2136, 2011. [Talmon et al., 2009] Ronen Talmon, Israel Cohen, et al. Relative Transfer Function Identification using Convolutive Transfer Function Approximation. IEEE Trans. Audio, Speech, Lang. Proc., 17(4):546 555, 2009. [Tan and Wang, 2020] Ke Tan and De Liang Wang. Learning Complex Spectral Mapping With Gated Convolutional Recurrent Networks for Monaural Speech Enhancement. IEEE/ACM Trans. Audio, Speech, Lang. Proc., 28:380 390, 2020. [Tang et al., 2020] Chuanxin Tang, Chong Luo, Zhiyuan Zhao, Wenxuan Xie, and Wenjun Zeng. Joint Time Frequency and Time Domain Learning for Speech Enhancement. In IJCAI, pages 3816 3822, 2020. [Tripathi et al., 2022] Vinay Tripathi, Huo Chen, Mostafa Khezri, Ka Wa Yip, et al. Suppression of Crosstalk in Superconducting Qubits using Dynamical Decoupling. Physical Review Applied, 18(2), 2022. [Tzinis et al., 2021] Efthymios Tzinis, Scott Wisdom, Aren Jansen, Shawn Hershey, Tal Remez, et al. Into the Wild with Audio Scope: Unsupervised Audio-Visual Separation of On-Screen Sounds. In ICLR, 2021. [Tzinis et al., 2022a] Efthymios Tzinis, Yossi Adi, Vamsi K. Ithapu, Buye Xu, Paris Smaragdis, and Anurag Kumar. Remix IT: Continual Self-Training of Speech Enhancement Models via Bootstrapped Remixing. IEEE J. of Selected Topics in Signal Proc., 16(6):1329 1341, 2022. [Tzinis et al., 2022b] Efthymios Tzinis, Scott Wisdom, Tal Remez, and John R. Hershey. Audio Scope V2: Audio Visual Attention Architectures for Calibrated Open Domain On-Screen Sound Separation. In ECCV, pages 368 385, 2022. [Vincent et al., 2006] Emmanuel Vincent, R emi Gribonval, and C edric F evotte. Performance Measurement in Blind Audio Source Separation. IEEE Trans. Audio, Speech, Lang. Proc., 14(4):1462 1469, 2006. [Wang and Chen, 2018] De Liang Wang and Jitong Chen. Supervised Speech Separation Based on Deep Learning: An Overview. IEEE/ACM Trans. Audio, Speech, Lang. Proc., 26(10):1702 1726, 2018. [Wang and Watanabe, 2023] Zhong-Qiu Wang and Shinji Watanabe. UNSSOR: Unsupervised Neural Speech Separation by Leveraging Over-determined Training Mixtures. In Neur IPS, 2023. [Wang et al., 2021a] Zhong-Qiu Wang, Peidong Wang, and De Liang Wang. Multi-Microphone Complex Spectral Mapping for Utterance-Wise and Continuous Speech Separation. IEEE/ACM Trans. Audio, Speech, Lang. Proc., 29:2001 2014, 2021. [Wang et al., 2021b] Zhong-Qiu Wang, Gordon Wichern, and Jonathan Le Roux. Convolutive Prediction for Monaural Speech Dereverberation and Noisy-Reverberant Speaker Separation. IEEE/ACM Trans. Audio, Speech, Lang. Proc., 29:3476 3490, 2021. [Wang et al., 2023a] Ruoyu Wang, Maokui He, Jun Du, et al. The USTC-NERCSLIP Systems for the CHi ME-7 DASR Challenge. In CHi ME, pages 13 18, 2023. [Wang et al., 2023b] Zhe Wang, Shilong Wu, Hang Chen, et al. The Multimodal Information Based Speech Processing (MISP) 2022 Challenge: Audio-Visual Diarization And Recognition. In ICASSP, pages 1 5, 2023. [Wang et al., 2023c] Zhong-Qiu Wang, Samuele Cornell, Shukjae Choi, et al. TF-Grid Net: Integrating Fulland Sub-Band Modeling for Speech Separation. IEEE/ACM Trans. Audio, Speech, Lang. Proc., 31:3221 3236, 2023. [Wang, 2024] Zhong-Qiu Wang. USDnet: Unsupervised Speech Dereverberation via Neural Forward Filtering. ar Xiv preprint ar Xiv:2402.00820, 2024. [Watanabe et al., 2020] Shinji Watanabe, Michael Mandel, Jon Barker, et al. CHi ME-6 Challenge: Tackling Multispeaker Speech Recognition for Unsegmented Recordings. In ar Xiv preprint ar Xiv:2004.09249, 2020. [Williamson et al., 2016] Donald S. Williamson, Yuxuan Wang, and De Liang Wang. Complex Ratio Masking for Monaural Speech Separation. IEEE/ACM Trans. Audio, Speech, Lang. Proc., pages 483 492, 2016. [Wisdom et al., 2020] Scott Wisdom, Efthymios Tzinis, Hakan Erdogan, Ron J. Weiss, Kevin Wilson, and John R. Hershey. Unsupervised Sound Separation using Mixture Invariant Training. In Neur IPS, 2020. [Ye et al., 2023] Lingxuan Ye, Haitian Lu, Gaofeng Cheng, Yifan Chen, et al. The IACAS-Thinkit System for CHi ME-7 Challenge. In CHi ME, pages 23 26, 2023. [Yin et al., 2020] Dacheng Yin, Chong Luo, Zhiwei Xiong, et al. PHASEN: A Phase-and-Harmonics-Aware Speech Enhancement Network. In AAAI, pages 9458 9465, 2020. [Yu et al., 2022] Fan Yu, Shiliang Zhang, Pengcheng Guo, Yihui Fu, Zhihao Du, et al. Summary on The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Grand Challenge. In ICASSP, pages 9156 9160, 2022. [Zhang et al., 2021] Wangyou Zhang, Jing Shi, Chenda Li, Shinji Watanabe, et al. Closing The Gap Between Time Domain Multi-Channel Speech Enhancement on Real and Simulation Conditions. In WASPAA, pages 146 150, 2021. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24)