# unpaired_imagetoimage_translation_via_neural_schrödinger_bridge__a83fc8e3.pdf Published as a conference paper at ICLR 2024 UNPAIRED IMAGE-TO-IMAGE TRANSLATION VIA NEURAL SCHRÖDINGER BRIDGE Beomsu Kim 1 Gihyun Kwon 2 Kwanyoung Kim1 Jong Chul Ye1 Equal contribution 1Kim Jaechul Graduate School of AI, KAIST 2Department of Bio and Brain Engineering, KAIST {beomsu.kim,cyclomon,cubeyoung,jong.ye}@kaist.ac.kr Diffusion models are a powerful class of generative models which simulate stochastic differential equations (SDEs) to generate data from noise. While diffusion models have achieved remarkable progress, they have limitations in unpaired imageto-image (I2I) translation tasks due to the Gaussian prior assumption. Schrödinger Bridge (SB), which learns an SDE to translate between two arbitrary distributions, have risen as an attractive solution to this problem. Yet, to our best knowledge, none of SB models so far have been successful at unpaired translation between high-resolution images. In this work, we propose Unpaired Neural Schrödinger Bridge (UNSB), which expresses the SB problem as a sequence of adversarial learning problems. This allows us to incorporate advanced discriminators and regularization to learn a SB between unpaired data. We show that UNSB is scalable and successfully solves various unpaired I2I translation tasks. Code: https://github.com/cyclomon/UNSB 1 INTRODUCTION Diffusion models (Sohl-Dickstein et al., 2015; Ho et al., 2020; Song et al., 2021a;b), a class of generative model which generate data by simulating stochastic differential equations (SDEs), have achieved remarkable progress over the past few years. They are capable of diverse and high-quality sample synthesis (Xiao et al., 2022), a desiderata which was difficult to obtain for several widely acknowledged models such as Generative Adversarial Networks (GANs) (Goodfellow et al., 2014), Variational Autoencoders (VAEs) (Kingma & Welling, 2014), and flow-based models (Dinh et al., 2015). Furthermore, the iterative nature of diffusion models proved to be useful for a variety of downstream tasks, e.g., image restoration (Chung et al., 2023) and conditional generation (Rombach et al., 2022). However, unlike GANs and their family which are free to choose any prior distribution, diffusion models often assume a simple prior, such as the Gaussian distribution. In other words, the initial condition for diffusion SDEs is generally fixed to be Gaussian noise. The Gaussian assumption prevents diffusion models from unlocking their full potential in unpaired image-to-image translation tasks such as domain transfer, style transfer, or unpaired image restoration. Schrödinger bridges (SBs), a subset of SDEs, present a promising solution to this issue. They solve the entropy-regularized optimal transport (OT) problem, enabling translation between two arbitrary distributions. Their flexibility have motivated several approaches to solving SBs via deep learning. For instance, Bortoli et al. (2021) extends the Iterative Proportional Fitting (IPF) procedure to the continuous setting, Chen et al. (2022) proposes likelihood training of SBs using Forward-Backward SDEs theory, and Tong et al. (2023) solves the SB problem with a conditional variant of flow matching. Some methods have solved SBs assuming one side of the two distributions is simple. Wang et al. (2021a) developed a two-stage unsupervised procedure for estimating the SB between a Dirac delta and data. I2SB (Liu et al., 2023) and In DI (Delbracio & Milanfar, 2023) uses paired data to learn SBs between Dirac delta and data. Su et al. (2023) discovered DDIMs (Song et al., 2021a) as SBs between Published as a conference paper at ICLR 2024 data and Gaussian distributions. Hence, they were able to perform image-to-image translation by concatenating two DDIMs, i.e., by passing through an intermediate Gaussian distribution. Yet, to the best of our knowledge, no work so far has successfully trained SBs for direct translation between high-resolution images in the unpaired setting. Most methods demand excessive computation, and even if it is not the case, we observe poor results. In fact, all representative SB methods fail even on the simple task of translating points between two concentric spheres as dimension increases. In this work, we first identify the main culprit behind the failure of SB for unpaired image-to-image translation as the curse of dimensionality. As the dimension of the considered data increases, the samples become increasingly sparse, failing to capture the shape of the underlying image manifold. This sampling error then biases the SB optimal transport map, leading to an inaccurate SB. Based on the self-similarity of SB, referring to the intriguing property that SB restricted to a subinterval of its time domain also a SB, we therefore propose the Unpaired Neural Schrödinger Bridge (UNSB), which formulates the SB problem as a sequence of transport cost minimization problems under the constraint on the KL divergence between the true target distribution and the model distribution. We show that its Lagrangian formulation naturally expresses the SB a composition of generators learned via adversarial learning (Goodfellow et al., 2014). One of the important advantages of the UNSB formulation of SB is that it allows us to mitigate the curse of dimensionality on two levels: we can extend the discriminator in adversarial learning to a more advanced one, and we can add regularization to enforce the generator to learn a mapping which aligns with our inductive bias. Furthermore, all components of UNSB are scalable, so UNSB is naturally scalable as well to large scale image translation problems. Experiments on toy and practical image-to-image translation tasks demonstrate that UNSB opens up a new research direction for applying diffusion models to large scale unpaired translation tasks. Our contributions can be summarized as follows. We identify the cause behind the failure of previous SB methods for image-to-image translation as the curse of dimensionality. We empirically verify this by using a toy task as a sanity check for whether an OT-based method is robust to the curse of dimensionality. We propose UNSB, which formulates SB as Lagrangian formulation under the constraint on the KL divergence between the true target distribution and the model distribution. This leads to the composition of generators learned via adversarial learning that overcome the curse of dimensionality with advanced discriminators. UNSB improves upon the Denoising Diffusion GAN (Xiao et al., 2022) by enabling translation between arbitrary distributions. Furthermore, based on the comparison with other existing unpaired translation methods, we demonstrate that UNSB is indeed a generalization of them by overcoming their shortcomings. 2 RELATED WORKS Schrödinger Bridges. The Schrödinger bridge (SB) problem, commonly referred to as the entropyregularized Optimal Transport (OT) problem (Schrödinger, 1932; Léonard, 2013), is the task of learning a stochastic process that transitions from an initial probability distribution to a terminal distribution over time, while subject to a reference measure. It is closely connected to the field of stochastic control problems (Caluya & Halder, 2021; Chen et al., 2021). Recently, the remarkable characteristic of the SB problem, which allows for the choice of arbitrary distributions as the initial and terminal distributions, has facilitated solutions to various generative modeling problems. In particular, novel algorithms leveraging Iterative Proportional Fitting (IPF) (Bortoli et al., 2021; Vargas et al., 2021) have been proposed to approximate score-based diffusion. Building upon these algorithms, several variants have been introduced and successfully applied in diverse fields such as, inverse problems (Shi et al., 2022), Mean-Field Games (Liu et al., 2022), constrained transport problems (Tamir et al., 2023), Riemannian manifolds (Thornton et al., 2022), and path samplers (Zhang & Chen, 2021; Ruzayqat et al., 2023). Some methods have solved SBs assuming one side of the two distributions is simple. In the unsupervised setting, Wang et al. (2021a) first learns a SB between a Dirac delta and noisy data, and then denoises the noisy samples. In the supervised setting, I2SB (Liu et al., 2023) and In DI (Delbracio & Milanfar, 2023) used paired data to learn SBs Published as a conference paper at ICLR 2024 Figure 1: Left: Illustration of trajectories for Vanilla SB and UNSB. Due to the curse of dimensionality, observed data in high dimensions become sparse and fail to describe image manifolds accurately. Vanilla SB learns optimal transport between observed data, leading to undesirable mappings. UNSB employs adversarial learning and regularization to learn an optimal transport mapping which successfully generalizes beyond observed data. Right: UNSB can be interpreted as successively refining the predicted target domain image, enabling the model to modify fine details while preserving semantics. See Section 4. Here, NFE stands for the number of function evaluations. between Dirac delta and data. DDIB (Su et al., 2023) concatenates two SBs between data and the Gaussian distribution for image-to-image translation. Recent works (Gushchin et al., 2023a; Shi et al., 2023) have achieved unpaired image-to-image translation for 128 128 resolution images. But, they are computationally intensive, often taking several days to train. To our best knowledge, our work represents the first endeavor to efficiently learn SBs between higher resolution unpaired images. Unpaired image-to-image translation. The aim of image-to-image translation (I2I) is to produce an image in the target domain that preserves the structural similarity to the source image. A seminal work in I2I is pix2pix (Isola et al., 2017), which undertook the task with paired training images, utilizing a straightforward pixel-wise regularization strategy. However, this approach is not applicable when dealing with unpaired training settings. In the context of unpaired data settings, early methods like (Zhu et al., 2017; Huang et al., 2018) maintained the consistency between the source and output images through a two-sided training strategy, namely cycle-consistency. However, these models are plagued by inefficiencies in training due to the necessity of additional generator training. In order to circumvent this issue, recent I2I models have shifted their focus to one-sided I2I translation. They aim to preserve the correspondence between the input and output through various strategies, such as geometric consistency (Fu et al., 2019) and mutual information regularization (Benaim & Wolf, 2017). Recently, Contrastive Unpaired Translation (CUT) and its variants (Park et al., 2020; Jung et al., 2022; Wang et al., 2021b; Zheng et al., 2021) have demonstrated improvements in I2I tasks by refining patch-wise regularization strategies. Despite the impressive performance demonstrated by previous GAN-based models in the I2I domain, this paper demonstrates that our SB-based approach paves the way for further improvements in the I2I task by introducing the iterative refinement through SB that overcomes the potential mode collapsing issue from a single-step generator from GAN. We believe that it represents a new paradigm in the ongoing advancement of image-to-image translation. 3 SCHRÖDINGER BRIDGES AND THE CURSE OF DIMENSIONALITY Given two distributions π0, π1 on Rd, the Schrödinger Bridge problem (SBP) seeks the most likely random process {xt : t [0, 1]} that interpolates π0 and π1. Specifically, let Ωbe the path space on Rd, i.e., the space of continuous functions from [0, 1] to Rd, and let P(Ω) be the space of probability measures on Ω. Then, the SBP solves QSB = arg min Q P(Ω) DKL(Q Wτ) s.t. Q0 = π0, Q1 = π1 (1) where Wτ is the Wiener measure with variance τ, and Qt denotes the marginal of Q at time t. We call QSB the Schrödinger Bridge (SB) between π0 and π1. Published as a conference paper at ICLR 2024 The SBP admits multiple alternative formulations, and among those, the stochastic control formulation and the static formulation play crucial roles in our work. Both formulations will serve as theoretical bases for our Unpaired Neural Schrödinger Bridge algorithm, and the static formulation will shed light on why previous SB methods have failed on unpaired image-to-image translation tasks. Stochastic control formulation. The stochastic control formulation (Pra, 1991) shows that {xt} QSB can be described by an Itô SDE dxt = u SB t dt + τ dwt (2) where the time-varying drift u SB t is a solution to the stochastic control problem u SB t = arg min u E Z 1 1 2 ut 2 dt s.t. dxt = ut dt + τ dwt x0 π0, x1 π1 (3) assuming u satisfies certain regularity conditions. Eq. (3) says that among the SDEs of the form Eq. (2) with boundaries π0 and π1, the drift for the SDE describing the SB has minimum energy. This formulation also reveals two useful properties of SBs. First, {xt} is a Markov chain, and second, {xt} converges to the optimal transport ODE trajectory as τ 0. Intuitively, τ controls the amount of randomness in the trajectory {xt}. Static formulation. The static formulation of SBP shows that sampling from QSB is extremely simple, assuming we know the joint distribution of the SB at t {0, 1}, denoted as QSB 01 . Concretely, for {xt} QSB, conditioned upon initial and terminal points x0 and x1, the density of xt can be described by a Gaussian density (Tong et al., 2023) p(xt|x0, x1) = N(xt|tx1 + (1 t)x0, t(1 t)τI). (4) Hence, to simulate the SB given an initial point x0, we may sample xt according to p(xt|x0) = R p(xt|x0, x1) d QSB 1|0(x1|x0) (5) where QSB 1|0 denotes the conditional distribution of x1 given x0. Surprisingly, QSB 01 is shown to be a solution to the entropy-regularized optimal transport problem QSB 01 = arg min γ Π(π0,π1) E(x0,x1) γ[ x0 x1 2] 2τH(γ) (6) where Π(π0, π1) is the collection of joint distributions whose marginals are π0 and π1, and H denotes the entropy function. For discrete π0 and π1, it is possible to find QSB 01 via the Sinkhorn-Knopp algorithm, and this observation has inspired several algorithms (Tong et al., 2023; Pooladian et al., 2023) for approximating the SB. Figure 2: Curse of dimensionality. Curse of dimensionality. According to the static formulation of the SBP, any algorithm for solving the SBP can be interpreted as learning an interpolation according to the entropy-regularized optimal transport Eq. (6) of the marginals π0 and π1. However, in practice, we only have a finite number of samples {xn 0}N n=1 π0 and {xm 1 }M m=1 π1. This means, the SB is trained to transport samples between the empirical distributions 1 N PN n=1 δxn 0 and 1 M PM m=1 δxm 1 . Due to the curse of dimensionality, the samples fail to describe the image manifolds correctly in high dimension. Ultimately, QSB 01 yield image pairs that do not meaningfully correspond to one another (see Figure 1 or the result for Neural Optimal Transport (NOT) in Figure 5). To illustrate this phenomenon in the simplest scenario, we consider the case where π0 and π1 are supported uniformly on two concentric d-spheres of radii 1 and 2, respectively. Then, samples from QSB 01 should have near-one cosine similarity, since QSB should transport samples from π0 radially outwards. However, when we draw M = N = 1000 i.i.d. samples from π0 and π1, and calculate the cosine similarity between (x0, x1) QSB 01 where QSB 01 is estimated using the Sinkhorn-Knopp algorithm, we observe decreasing similarity as dimension increases (see Figure 2). Thus, in a high dimension, QSB approximated by Sinkhorn-Knopp will interpolate between nearly orthogonal points. Published as a conference paper at ICLR 2024 4 UNPAIRED NEURAL SCHRÖDINGER BRIDGE (UNSB) We now explain our novel UNSB algorithm which shows that SB can be expressed as a composition of generators learned via adversarial learning. Specifically, given a partition {ti}N i=0 of the unit interval [0, 1] such that t0 = 0, t N = 1, and ti < ti+1, we can simulate SB according to the Markov chain decomposition p({xtn}) = p(xt N |xt N 1)p(xt N 1|xt N 2) p(xt1|xt0)p(xt0). (7) The decomposition Eq. (7) allows us to learn the SB inductively: we learn p(xti+1|xti) assuming we are able to sample from p(xti). Having approximated p(xti+1|xti), we can sample from p(xti+1), so we may learn p(xti+2|xti+1). Since the distribution of xt0 = x0 is already known as π0, our main interest lies in learning the transition probabilities p(xti+1|xti) assuming we can sample from p(xti). This procedure is applied recursively for i = 0, . . . , N 1. Let qϕi(x1|xti) be a conditional distribution parametrized by a DNN with parameter ϕi. Intuitively, qϕi(x1|xti) is a generator which predicts the target domain image for xti. We define qϕi(xti, x1) := qϕi(x1|xti)p(xti), qϕi(x1) := Ep(xti)[qϕi(x1|xti)]. (8) The following theorem shows how to optimize ϕi and sample from p(xti+1|xti). Theorem 1. For any ti, consider the following constrained optimization problem min ϕi LSB(ϕi, ti) := Eqϕi(xti,x1)[ xti x1 2] 2τ(1 ti)H(qϕi(xti, x1)) (9) s.t. LAdv(ϕi, ti) := DKL(qϕi(x1) p(x1)) = 0 (10) and define the distributions p(xti+1|x1, xti) := N(xti+1|si+1x1 + (1 si+1)xti, si+1(1 si+1)τ(1 ti)I) (11) where si+1 := (ti+1 ti)/(1 ti) and qϕi(xti+1|xti) := Eqϕi(x1|xti)[p(xti+1|x1, xti)], qϕi(xti+1) := Ep(xti)[qϕi(xti+1|xti)]. (12) If ϕi solves Eq. (9), then we have qϕi(x1|xti) = p(x1|xti), qϕi(xti+1|xti) = p(xti+1|xti), qϕi(xti+1) = p(xti+1). (13) Proof Sketch. Using the stochastic control formulation of SB, we show that the SB satisfies a certain self-similarity property, which states that the SB restricted any sub-interval of [0, 1] is also a SB. Note that Eqs. (11), (12) and (9) under the constraint Eq. (10) are indeed counter-parts of the static formulation Eqs. (4), (5), and (6), respectively, for the restricted domain [ti, 1]. The static formulation of SB restricted to the interval [ti, 1] shows that qϕi(x1|xti) = p(x1|xti) is the solution to Eq. (9). If we assume ϕi solves Eq. (9) such that qϕi(x1|xti) = p(x1|xti), Eq. (13) follows by simple calculation. The detailed proof can be found in the Appendix. In practice, by incorporating the equality constraint in Eq. (10) into the loss with a Lagrange multiplier, we obtain the UNSB objective for a single time-step ti min ϕi LUNSB(ϕi, ti) := LAdv(ϕi, ti) + λSB,ti LSB(ϕi, ti). (14) Since it is impractical to use separate parameters ϕi for each time-step ti and to learn p(xti+1|xti) sequentially for i = 0, . . . , N 1, we replace qϕi(x1|xti), which takes xti as input, with a timeconditional DNN qϕ(x1|xti), which shares a parameter ϕ for all time-steps ti, and takes the tuple (xti, ti) as input. Then, we optimize the sum of LUNSB(ϕ, ti) over i = 0, . . . , N 1. Training. For the training of our UNSB, we first randomly choose a time-step ti to optimize. To calculate LUNSB(ϕ, ti), we sample xti and x1 π1, where π1 denotes the target distribution. The sampling procedure of xti will be described soon. The sample xti is then passed through qϕ(x1|xti) to obtain x1(xti) that refers to the estimated target data sample given xti. The pairs (xti, x1(xti)) and (x1, x1(xti)) are then used to compute LSB(ϕ, ti) and LAdv(ϕ, ti) in Eq. (9) and Eq. (10), respectively. Specifically, we estimate the entropy term in LSB with a mutual information estimator, Published as a conference paper at ICLR 2024 Figure 3: Generation and training process of UNSB for time step ti. sg means stop gradient. using the fact that for a random variable X, I(X, X) = H(X) where I denotes mutual information. We then estimate the divergence in LAdv with adversarial learning. Intuitively, x1 and x1(xti) are real and fake inputs to the discriminator. This process is shown in the training stage of Figure 3. Generation of intermediate and final samples. We now describe the sampling procedure for the intermediate samples for training and inference. We simulate the Markov chain Eq. (7) using qϕ as follows: given xtj qϕ(xtj) (note that xtj = x0 π0 if j = 0), we predict the target domain image x1(xtj) qϕ(x1|xtj). We then sample xtj+1 qϕ(xtj+1) according to Eq. (11) by interpolating x0 and x1(xtj) and adding Gaussian noise. Repeating this procedure for j = 0 . . . , i 1, we get xti qϕ(xti). With optimal ϕ, by Theorem 1, xti qϕ(xti) = p(xti). Thus, the trajectory {x1(xti) : i = 0, . . . , N 1} can be viewed as an iterative refinement of the predicted target domain sample. This process is illustrated within the generation stage of Figure 3. 4.1 COMBATING THE CURSE OF DIMENSIONALITY Advanced discriminator in adversarial learning. One of the most important advantages of the formulation Eq. (14) is that we can replace the KL-divergence in LAdv by any divergence or metric which measures discrepancy between two distributions. Such divergence or metric can be estimated through adversarial learning, i.e., its Kantorovich dual. For example, 2 DJSD(qϕi(x1) p(x1)) log(4) = max D Ep(x1)[log D(x1)] + Eqϕi(x1)[log(1 D(x1))] (15) where D is a discriminator. This allows us to use various adversarial learning techniques to mitigate the curse of dimensionality. For instance, instead of using a standard discriminator which distinguishes generated and real samples on the instance level, we can use a Markovian discriminator which distinguishes samples on the patch level. The Markovian discriminator is effective at capturing high-frequency characteristics (e.g., style) of the target domain data (Isola et al., 2017). Regularization. Furthermore, we augment the UNSB objective with regularization, which enforces the generator network qϕ to satisfy consistency between predicted x1 and the initial point x0: LReg(ϕ, ti) := Ep(x0,xti)Eqϕ(x1|xti)[R(x0, x1)] (16) Here, R is a scalar-valued differentiable function which quantifies an application-specific measure of similarity between its inputs. In other words, R reflects our inductive bias for similarity between two images. Thus, the regularized UNSB objective for time ti is LUNSB(ϕ, ti) := LAdv(ϕ, ti) + λSB,ti LSB(ϕ, ti) + λReg,ti LReg(ϕ, ti), (17) which is the final objective in our UNSB algorithm. 4.2 SANITY CHECK ON TOY DATA Two shells. With the two shells data of Section 3, we show that representative SB methods suffer from the curse of dimensionality, whereas UNSB does not, given appropriate discriminator and regularization. For SB methods, we consider Sinkhorn-Knopp (SK), SB Conditional Flow Matching (SBCFM) Published as a conference paper at ICLR 2024 (Tong et al., 2023), Diffusion SB (DSB) (Bortoli et al., 2021), and SB-FBSDE (Chen et al., 2022). For baselines, we use recommended settings, and for UNSB, we train the discriminator to distinguish real and fake samples by input norms, and choose negative cosine similarity as R in Eq. (16). Figure 4: Results on two shells. Method µ MSE Σ MSE SK 1.5e-5 3.1e-6 SBCFM 4.7e-5 2.3e-5 DSB 4.028 0.139 SB-FBSDE 2.974 2.5e-9 UNSB 0.008 6.4e-7 Table 1: Results on two Gaussians. Only 1k samples from each π0, π1 are used throughout the training. In Figure 4, we see all baselines either fail to transport samples to the target data manifold or fail to maintain high cosine similarity between input and outputs. On the other hand, UNSB is robust to dimension. Two Gaussians. We also verify whether UNSB can learn the SB between two Gaussians. The SB between Gaussian distributions is also Gaussian, and its mean and covariance can be derived in a closed form (Bunne et al., 2023). Thus, we may measure the error between approximated and true solutions exactly. We consider learning the SB between N( 1, I) and N(1, I) in a 50-dimensional space. In Table 1, we see that UNSB recovers the mean and covariance of the ground-truth SB relatively accurately. 4.3 RELATION TO OTHER UNPAIRED TRANSLATION METHODS UNSB vs. GANs. N = 1 version of UNSB is nearly equivalent to GAN-based translation methods such as CUT, i.e., UNSB may be interpreted as a multi-step generalization of GAN methods. Multi-step models are able to fit complex mappings that GANs are unable to by breaking down complex maps into a composition of simple maps. That is why diffusion models often achieve better performance than GANs (Salmona et al., 2022). Similarly, we believe the multi-step nature of UNSB arising from SB allow it to obtain better performance than GAN-based translation methods. UNSB vs. previous diffusion methods. Previous diffusion-based methods such as SDEdit (Meng et al., 2022) also translate images along SDEs. However, the SDEs used by such methods (e.g., VP-SDE or VE-SDE) may be suboptimal in terms of transport cost, resulting in large NFEs. On the other hand, SB allows fast generation, as it learns the shortest path between domains. Furthermore, UNSB uses adversarial learning, so it may generate feasible samples even with NFE = 1. On the contrary, diffusion methods do not perform so well on complex tasks such as Horse2Zebra. 5 RESULTS FOR UNPAIRED IMAGE-TO-IMAGE TRANSLATION In this section, we provide experimental results of UNSB for large scale unpaired image-to-image translation tasks. UNSB settings. We use the Markovian discriminator in LAdv and the patch-wise contrastive matching loss (Park et al., 2020) in LReg. We set the number of timesteps as N = 5, so we discretize the unit interval into t0, t1, . . . , t5. We set λSB,ti = λReg,ti = 1 and τ = 0.01. We measure the sample quality of x1(xti) for each i = 0, . . . , N 1. So, NFE = i denotes sample evaluation or visualization of x1(xti 1). Other details are deferred to the Appendix. Evaluation. We use four benchmark datasets: Horse2Zebra, Map2Cityscape, Summer2Winter, and Map2Satellite. All images are resized into 256 256. We use the FID score (Heusel et al., 2017) and the KID score (Chen et al., 2020) to measure sample quality. Baselines. We note all SB methods SBCFM, DSB, and SB-FBSDE do not provide results on 256 resolution images in their paper due to scalability issues. While we were able to make SBCFM work on 256 resolution images, the results were poor, so we deferred its results to the Appendix. Instead, we used Neural Optimal Transport (NOT) (Korotin et al., 2023) as a representative for OT. For GAN methods, we used Cycle GAN (Zhu et al., 2017), MUNIT (Huang et al., 2018), Distance GAN (Benaim & Wolf, 2017), Gc GAN (Fu et al., 2019), and CUT (Park et al., 2020). For diffusion, we used SDEdit (Meng et al., 2022) and P2P (Hertz et al., 2022) with LDM (Rombach et al., 2022). Published as a conference paper at ICLR 2024 Method NFE Time Horse2Zebra Summer2Winter Label2Cityscape Map2Satellite FID KID FID KID FID KID FID KID NOT 1 0.006 104.3 5.012 185.5 8.732 221.3 19.76 224.9 16.59 Cycle GAN 1 0.004 77.2 1.957 84.9 1.022 76.3 3.532 54.6 3.430 MUNIT 1 0.011 133.8 3.790 115.4 4.901 91.4 6.401 181.7 12.03 Distance 1 0.009 72.0 1.856 97.2 2.843 81.8 4.410 98.1 5.789 Gc GAN 1 0.0027 86.7 2.051 97.5 2.755 105.2 6.824 79.4 5.153 CUT 1 0.0033 45.5 0.541 84.3 1.207 56.4 1.611 56.1 3.301 SDEdit 30 1.98 97.3 4.082 118.6 3.218 P2P 50 120 60.9 1.093 99.1 2.626 Ours-best 5 0.045 35.7 0.587 73.9 0.421 53.2 1.191 47.6 2.013 Table 2: NFE and generation time (sec.) per image and FID and KID 100 of generated samples. (a) Comparison on natural domain data. (b) Comparison on artificial domain data. Figure 5: Qualitative comparison of image-to-image translation results from our UNSB and baseline I2I methods. Compared to other one-step baseline methods, our model generates more realistic domain-changed outputs while preserving the structural information of the source images. Since LDM is trained on natural images, we can use LDM on Horse2Zebra and Summer2Winter, but not on Label2Cityscape and Map2Satellite. Hence, we omit diffusion results on those two tasks. Comparison results. We show quantitative results in Table 2, where we observe our model outperforms baseline methods in all datasets. In particular, out model largely outperformed early GAN-based translation methods. When compared to recent models such as CUT, our model still shows better scores. NOT suffers from degraded performance in all of the datasets. Qualitative comparison in Figure 5 provides insight into the superior performance of UNSB. Our model successfully translates source images to the target domain while preventing structural inconsis- Published as a conference paper at ICLR 2024 tency. For other GAN-based methods, the model outputs do not properly reflect the target domain information, and some models fail to preserve source image structure. For NOT, we observe that the model failed to overcome the curse of dimensionality. Specifically, NOT hallucinates structures not present in the source image (for instance, see the first row of Figure 5), leading to poor samples. For diffusion-based methods, we see they fail to fully reflect the target domain style. Figure 6: NFE analysis and failure case. Disc. Reg. NFE FID KID 5 230 18.8 Instance 1 104 5.0 Patch 1 66.3 1.6 Patch 5 58.9 1.5 Patch 5 35.7 0.587 Table 3: Ablation study. Figure 7: UNSB output variation. Figure 8: Input-output distance. NFE analysis. We investigate the relationship between NFE and the translated sample quality. In Figure 6 top, we observe that using NFE = 1 resulted in relatively poor performance, sometimes even worse than existing one-step GAN methods. However, we see that increasing the NFE consistently improved generation quality: best FIDs were achieved for NFE values between 3 and 5 for all datasets. This trend is also evident in our qualitative comparisons. Looking at Figure 1, we observe that as NFE increases, there are gradual improvements in image quality. However, UNSB at times suffers from the problem where artifacts occur with large NFE, as shown in Figure 6 bottom. We speculate this causes increasing FID for large NFEs on datasets such as Map2Satellite. Ablation study. Unlike GAN-based methods, UNSB is able to generate samples in multiple steps. Also unlike SB methods, UNSB is able to use advanced discriminators and regularization. To understand the influence of those factors on the final performance, we performed an ablation study on the Horse2Zebra dataset. We report the results in Table 3. In the case where we use neither advanced discriminator nor regularization (corresponding to the previous SB method SBCFM), we see poor results. Then, as we add multistep generation, advanced discriminator (Markovian discriminator, denoted Patch ), and regularization, the result gradually improves until we obtain the best performance given by UNSB. This implies that the three components of UNSB play orthogonal roles. Stochasticity analysis. SB is stochastic and should return diverse outputs. In Figure 7, we see meaningful variation in the generated images. Hence, UNSB indeed learns a stochastic map. Transport cost analysis. SB solves OT, so inputs and outputs must be close. In Figure 8, we compare input-output distance for UNSB pairs and SK pairs computed between actual dataset images. we observe that input-output distance for UNSB is smaller than those of SK. This implies UNSB successfully generalizes beyond observed points to generate pairs which are close under L2 norm and are realistic. 6 CONCLUSION In this work, we proposed Unpaired Neural Schrödinger Bridge (UNSB) which solves the Schrödinger Bridge problem (SBP) via adversarial learning. UNSB formulation of SB allowed us to combine SB with GAN training techniques for unpaired image-to-image translation. We demonstrated the scalability and effectiveness of UNSB through various data-to-data or image-to-image translation tasks. In particular, while all previous methods for SB or OT fail, UNSB achieved results that often surpass those of one-step models. Overall, our work opens up a previously unexplored research direction for applying diffusion models to unpaired image translation. Published as a conference paper at ICLR 2024 ETHICS AND REPRODUCIBILITY STATEMENTS Ethics statement. UNSB extends diffusion to translation between two arbitrary distributions, allowing us to explore a wider range of applications. In particular, UNSB may be used in areas with beneficial impacts, such as medical image restoration. However, UNSB may also be used to create malicious content such as fake news, and this must be prevented through proper regulation. Reproducibility statement. Pseudo-codes and hyper-parameters are described in the main paper and the Appendix. ACKNOWLEDGMENTS This research was supported by the National Research Foundation of Korea (NRF) (RS202300262527), Field-oriented Technology Development Project for Customs Administration funded by the Korean government (the Ministry of Science & ICT and the Korea Customs Service) through the National Research Foundation (NRF) of Korea under Grant NRF2021M3I1A1097910 & NRF2021M3I1A1097938, Korea Medical Device Development Fund grant funded by the Korea government (the Ministry of Science and ICT, the Ministry of Trade, Industry, and Energy, the Ministry of Health & Welfare, the Ministry of Food and Drug Safety) (Project Number: 1711137899, KMDF PR 20200901 0015), and Culture, Sports, and Tourism R&D Program through the Korea Creative Content Agency grant funded by the Ministry of Culture, Sports and Tourism in 2023. Mohamed Ishmael Belghazi, Aristide Baratin, Sai Rajeswar, Sherjil Ozair, Yoshua Bengio, Aaron Courville, and R Devon Hjelm. Mutual information neural estimation. In ICML, 2018. Sagie Benaim and Lior Wolf. One-sided unsupervised domain mapping. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https://proceedings.neurips.cc/paper/2017/file/ 59b90e1005a220e2ebc542eb9d950b1e-Paper.pdf. Valentin De Bortoli, James Thornton, Jeremy Heng, and Arnaud Doucet. Diffusion schrödinger bridge with applications to score-based generative modeling. In Neur IPS, 2021. Charlotte Bunne, Ya-Ping Hsieh, Marco Cuturi, and Andreas Krause. The schrödinger bridge between gaussian measures has a closed form. In AISTATS, 2023. Kenneth F Caluya and Abhishek Halder. Reflected schrödinger bridge: Density control with path constraints. In 2021 American Control Conference (ACC), pp. 1137 1142. IEEE, 2021. Runfa Chen, Wenbing Huang, Binghui Huang, Fuchun Sun, and Bin Fang. Reusing discriminators for encoding: Towards unsupervised image-to-image translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8168 8177, 2020. Tianrong Chen, Guan-Horng Liu, and Evangelos Theodorou. Likelihood training of schrödinger bridge using forward-backward sdes theory. In ICLR, 2022. Yongxin Chen, Tryphon T Georgiou, and Michele Pavon. Stochastic control liaisons: Richard sinkhorn meets gaspard monge on a schrodinger bridge. Siam Review, 63(2):249 313, 2021. Yunjey Choi, Youngjung Uh, Jaejun Yoo, and Jung-Woo Ha. Stargan v2: Diverse image synthesis for multiple domains. In CVPR, 2022. Hyungjin Chung, Jeongsol Kim, Michael Thompson Mccann, Marc Louis Klasky, and Jong Chul Ye. Diffusion posterior sampling for general noisy inverse problems. In ICLR, 2023. Mauricio Delbracio and Peyman Milanfar. Inversion by direct iteration: An alternative to denoising diffusion for image restoration. arxiv preprint ar Xiv:2303.11435, 2023. Published as a conference paper at ICLR 2024 Laurent Dinh, David Krueger, and Yoshua Bengio. Nice: Non-linear independent components estimation. arxiv preprint ar Xiv:1410.8516, 2015. Huan Fu, Mingming Gong, Chaohui Wang, Kayhan Batmanghelich, Kun Zhang, and Dacheng Tao. Geometry-consistent generative adversarial networks for one-sided unsupervised domain mapping. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Neur IPS, 2014. Nikita Gushchin, Alexander Kolesov, Alexander Korotin, Dmitry Vetrov, and Evgeny Burnaev. Entropic neural optimal transport via diffusion processes. In Neur IPS, 2023a. Nikita Gushchin, Alexander Kolesov, Alexander Korotin, Dmitry Vetrov, and Evgeny Burnaev. Entropic neural optimal transport via diffusion processes. In Neur IPS, 2023b. Nikita Gushchin, Alexander Kolesov, Petr Mokrov, Polina Karpikova, Andrey Spiridonov, Evgeny Burnaev, and Alexander Korotin. Building the bridge of schrödinger: A continuous entropic optimal transport benchmark. In Neur IPS Track on Datasets and Benchmarks, 2023c. Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Promptto-prompt image editing with cross attention control. ar Xiv preprint ar Xiv:2208.01626, 2022. Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017. Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In Neur IPS, 2020. Xun Huang, Ming-Yu Liu, Serge Belongie, and Jan Kautz. Multimodal unsupervised image-to-image translation. In Proceedings of the European Conference on Computer Vision (ECCV), September 2018. Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1125 1134, 2017. Chanyong Jung, Gihyun Kwon, and Jong Chul Ye. Exploring patch-wise semantic relation for contrastive learning in image-to-image translation tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 18260 18269, June 2022. Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of GANs for improved quality, stability, and variation. In ICLR, 2018. Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. In ICLR, 2014. Alexander Korotin, Daniil Selikhanovych, and Evgeny Burnaev. Neural optimal transport. In ICLR, 2023. Christian Léonard. A survey of the schrödinger problem and some of its connections with optimal transport. ar Xiv preprint ar Xiv:1308.0215, 2013. Jae Hyun Lim, Aaron Courville, Christopher Pal, and Chin-Wei Huang. Ar-dae: Towards unbiased neural entropy estimation. In ICML, 2020. Guan-Horng Liu, Tianrong Chen, Oswin So, and Evangelos A Theodorou. Deep generalized schrödinger bridge. ar Xiv preprint ar Xiv:2209.09893, 2022. Guan-Horng Liu, Arash Vahdat, De-An Huang, Evangelos A. Theodorou, Weili Nie, and Anima Anandkumar. I2sb: Image-to-image schrödinger bridge. arxiv preprint ar Xiv:2302.05872, 2023. Chenlin Meng, Yutong He, Yang Song, Jiaming Song, Jiajun Wu, Jun-Yan Zhu, and Stefano Ermon. Sdedit: Guided image synthesis and editing with stochastic differential equations. In ICLR, 2022. Published as a conference paper at ICLR 2024 Taesung Park, Alexei A. Efros, Richard Zhang, and Jun-Yan Zhu. Contrastive learning for unpaired image-to-image translation. In Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm (eds.), Computer Vision ECCV 2020, pp. 319 345, Cham, 2020. Springer International Publishing. ISBN 978-3-030-58545-7. Aram-Alexandre Pooladian, Heli Ben-Hamu, Carles Domingo-Enrich, Brandon Amos, Yaron Lipman, and Ricky T. Q. Chen. Multisample flow matching: Straightening flows with minibatch couplings. In ICML, 2023. Paolo Dai Pra. A stochastic control appraoch to reciprocal diffusion processes. Applied Mathematics and Optimization, 23(1):313 329, 1991. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. Highresolution image synthesis with latent diffusion models. In CVPR, 2022. Hamza Ruzayqat, Alexandros Beskos, Dan Crisan, Ajay Jasra, and Nikolas Kantas. Unbiased estimation using a class of diffusion processes. Journal of Computational Physics, 472:111643, 2023. Antoine Salmona, Agnès Desolneux, Julie Delon, and Valentin De Bortoli. Can push-forward generative models fit multimodal distributions? In Neur IPS, 2022. Erwin Schrödinger. Sur la théorie relativiste de l électron et l interprétation de la mécanique quantique. Annales de l institut Henri Poincaré, 2(4):269 310, 1932. Yuyang Shi, Valentin De Bortoli, George Deligiannidis, and Arnaud Doucet. Conditional simulation using diffusion schrödinger bridges. In Uncertainty in Artificial Intelligence, pp. 1792 1802. PMLR, 2022. Yuyang Shi, Valentin De Bortoli, Andrew Campbell, and Arnaud Doucet. Diffusion schrödinger bridge matching. In Neur IPS, 2023. Jascha Sohl-Dickstein, Eric A. Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learing using nonequilibrium thermodynamics. In ICML, 2015. Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. In ICLR, 2021a. Yang Song, Jascha Sohl-Dickstein, Diederik P. Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. In ICLR, 2021b. Xuan Su, Jiaming Song, Chenlin Meng, and Stefano Ermon. Dual diffusion implicit bridges for image-to-image translation. In ICLR, 2023. Ella Tamir, Martin Trapp, and Arno Solin. Transport with support: Data-conditional diffusion bridges. ar Xiv preprint ar Xiv:2301.13636, 2023. James Thornton, Michael Hutchinson, Emile Mathieu, Valentin De Bortoli, Yee Whye Teh, and Arnaud Doucet. Riemannian diffusion schrödinger bridge. ar Xiv preprint ar Xiv:2207.03024, 2022. Alexander Tong, Nikolay Malkin, Guillaume Huguet, Yanlei Zhang, Jarrid Rector-Brooks, Kilian Fatras, Guy Wolf, and Yoshua Bengio. Conditional flow matching: Simulation-free dynamic optimal transport. arxiv preprint ar Xiv:2302.00482, 2023. Francisco Vargas, Pierre Thodoroff, Austen Lamacraft, and Neil Lawrence. Solving schrödinger bridges via maximum likelihood. Entropy, 23(9):1134, 2021. Gefei Wang, Yuling Jiao, Qian Xu, Yang Wang, and Can Yang. Deep generative learning via schrödinger bridge. In International Conference on Machine Learning, pp. 10794 10804. PMLR, 2021a. Weilun Wang, Wengang Zhou, Jianmin Bao, Dong Chen, and Houqiang Li. Instance-wise hard negative example generation for contrastive learning in unpaired image-to-image translation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 14020 14029, October 2021b. Published as a conference paper at ICLR 2024 Zhisheng Xiao, Karsten Kreis, and Arash Vahdat. Tackling the generative learning trilemma with denoising diffusion gans. In ICLR, 2022. Qinsheng Zhang and Yongxin Chen. Path integral sampler: a stochastic control approach for sampling. ar Xiv preprint ar Xiv:2111.15141, 2021. Min Zhao, Fan Bao, Chongxuan Li, and Jun Zhu. Egsde: Unpaired image-to-image translation via energy-guided stochastic differential equations. In Neur IPS, 2022. Chuanxia Zheng, Tat-Jen Cham, and Jianfei Cai. The spatially-correlative loss for various image translation tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16407 16417, June 2021. Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Oct 2017. Published as a conference paper at ICLR 2024 Lemma 1 (Self-similarity). Let [ta, tb] [0, 1] and {xt} QSB. The SB restricted to [ta, tb], defined as the distribution of {xt}|[ta,tb] := {xt(s) : s [0, 1]} where t(s) := ta + (tb ta)s solves min Q P(Ω) DKL(Q Wτ(tb ta)) s.t. Q0 = QSB ta , Q1 = QSB tb . (18) Proof. We claim u SB t restricted to the interval [ta, tb] is also a SB from QSB ta to QSB tb . Suppose the claim is false, so there is another drift ˆut on [ta, tb] such that ta ˆut 2 dt < E Z tb ta u SB t 2 dt and dxt = ˆut dt + τ dwt, xta QSB ta , xtb QSB tb . We can extend ˆut to the entire interval [0, 1] by defining u SB t if 0 t < ta, ˆut if ta t < tb, u SB t if tb t < 1. We then have 0 ˆut 2 dt < E Z 1 0 u SB t 2 dt and dxt = ˆut dt + τ dwt, x0 π0, x1 π1, which contradicts our assumption that u SB t solves Eq. (3). We note that by change of time variable, dxt = u SB t dt + τ dwt, ta t tb is equivalent to the SDE dxs = (tb ta)u SB t(s) ds + p τ(tb ta) dws, 0 s 1 By comparing the stochastic control formulation Eq. (3) with the original SBP Eq. (1), we see that this means the reference Wiener measure for the SBP restricted to [ta, tb] has variance τ(tb ta). Lemma 2 (Static formulation of restricted SBs). Let t [ta, tb] [0, 1] and {xt} QSB. Then p(xt|xta, xtb) = N(xt|s(t)xtb + (1 s(t))xta, s(t)(1 s(t))τ(tb ta)I) (19) where s(t) := (t ta)/(tb ta) is the inverse function of t(s). Moreover, QSB tatb = arg min γ Π(Qta,Qtb) E(xta,xtb) γ[ xta xtb 2] 2τ(tb ta)H(γ). (20) Proof. Translate the SBP Eq. (18) into the static formulation, taking into account the reduced variance τ(tb ta) of the reference Weiner measure Wτ(tb ta). Proof of Theorem 1. Let ta = ti and tb = 1 in Lemma 2. If the class of distributions expressed by qϕi(x1|xti) is sufficiently large, for any γ Π(Qti, Q1), dγ(xti, x1) = qϕi(xti, x1) (21) for some ϕi which satisfies DKL(qϕi(x1) p(x1)) = 0. (22) Thus, Eq. (20) and Eq. (9) under the constraint Eq. (10) are equivalent optimization problems, which yield the same solutions, namely p(xti, x1). Then, with optimal ϕi, qϕi(x1|xti)p(xti) = qϕi(xti, x1) = p(xti, x1). (23) Together with the observation that Eq. (11) is identical to Eq. (19), Eq. (23) implies Eq. (13). Published as a conference paper at ICLR 2024 B OMITTED EXPERIMENT DETAILS Training. All experiments are conducted on a single RTX3090 GPU. On each dataset, we train our UNSB network for 400 epochs with batch size 1 and Adam optimizer with β1 = 0.5, β2 = 0.999, and initial learning rate 0.0002. Learning rate is decayed linearly to zero until the end of training. All images are resized into 256 256 and normalized into range [ 1, 1]. For SB training and simulation, we discretize the unit interval [0, 1] into 5 intervals with uniform spacing. We used λSB = λReg = 1 and τ = 0.01. To estimate the entropy loss, we used mutual information neural estimation method Belghazi et al. (2018). To incorporate timestep embedding and stochastic conditioning into out UNSB network, we used positional embedding and Ada IN layers, respectively, following the implementation of DDGAN Xiao et al. (2022). For I2I tasks, we used CUT loss as regularization. On Summer2Winter translation task, we used a pre-trained VGG16 network as our feature selection source, following the strategy in the previous work by Zheng et al. (2021). Entropy approximation. We first describe how to estimate the entropy of a general random variable. We observe that for a random variable X, I(X, X) = H(X) H(X|X) | {z } =0 = H(X) (24) so mutual information may be used to estimate the entropy of a random variable. Works such as Belghazi et al. (2018) and Lim et al. (2020) provide methods to estimate mutual information of two random variables by optimizing a neural network. For instance, Belghazi et al. (2018) tells us that IΘ(X, Z) := sup θ Θ EPXZ[Tθ] log EPX PZ[e Tθ] (25) where Tθ is a neural network parametrized by θ Θ is capable of approximating I(X, Z) up to arbitrary accuracy. Hence, we can estimate H(X) = I(X, X) by setting X = Z and optimizing Eq. (25). In the UNSB objective Eq. (9), we use Eq. (25) to estimate the entropy of X = Z = (x1, xti) where (xti, x1) qϕi(xti, x1) for qϕi defined in Eq. (8). In practice, we use the same neural net architecture for Tθ and the discriminator. We use the same network for Tθ for each time-step ti by incorporating time embedding into Tθ. Tθ is updated once with gradient ascent every qϕi update, analogous to adversarial training. Evaluation. Metric calculation codes FID : https://github.com/mseitzer/pytorch-fid KID : https://github.com/alpc91/NICE-GAN-pytorch Evaluation protocol: we follow standard procedure, as described in Park et al. (2020). KID measurement Distance GAN, MUNIT, NOT, SBCFM : trained from scratch using official code. Baselines excluding above four methods : KID measured using samples generated by models from official repository. FID measurement Horse2Zebra and Label2Cityscape : numbers taken from Table 1 in Park et al. (2020). Summer2Winter and Map2Satellite : all baselines trained using official code. Published as a conference paper at ICLR 2024 Method FID KID SBCFM 229.5 18.8 Ours 35.7 0.587 Table 4: Quantitative comparison result with another SB-based method on Horse2Zebra. The baseline model performance is severely degraded, while our generated results show superior perceptual quality. Figure 9: Visualization of our generated result and the result from conditional flow matching (CFM). Figure 10: UNSB output pixel-wise standard deviation given input. Method Time NFE FID KID EGSDE 60 500 45.2 3.62 Star GAN v2 0.006 1 48.55 3.20 NOT 0.006 1 51.5 3.17 Ours 0.045 5 37.87 1.54 Table 5: Male2Female results. Figure 11: Male2Female samples. C ADDITIONAL EXPERIMENTS C.1 OTHER SB METHODS In this part, we compare our proposed UNSB with other Schrödinger bridge based image translation methods. The recent work of I2SB and In DI require supervised training setting, therefore we cannot compare our proposed method with the baselines. For most of other SB-based methods, they focus on relatively easy problem, where one side of distribution is simple. Since we firstly proposed applying SB-based approach to high-resolution I2I problem, there is no direct baseline to compare. However, we compared our model with SB-based approach of Schrödinger Bridge Conditional Flow Matching (SBCFM) (Shi et al., 2022) by adapting the SBCFM to high-resolution case. In Table 4 and Fig. 9, we can see that the baseline SBCFM model could not generate the proper outputs where the outputs are totally unrelated to the input image. The results show that the model could not overcome the curse of dimensionality problem. Published as a conference paper at ICLR 2024 C.2 MORE STOCHASTICITY ANALYSIS In Figure 10, we visualize output pixel-wise standard deviation given input for more samples. We note that all pixel values are normalized into [0, 1]. A SB between two domains is stochastic, so it does one-to-many translation. Thus, a necessary condition for a model to be a SB is stochasticity. Figure 10 shows that UNSB satisfies this necessary condition. Greater std for foreground pixels tells us UNSB does one-to-many generation, and generated images lie in the target domain. For instance, a model which simply adds Gaussian noise also does one-tomany generation, but is not meaningful. UNSB output shows high variation for locations which are relevant to the target domain (Zebra). C.3 MALE2FEMALE TRANSLATION We provide results on the Male2Female translation task with the Celeb A-HQ-256 dataset (Karras et al., 2018). Baseline methods are EGSDE (Zhao et al., 2022), Star GAN v2 (Choi et al., 2022), and NOT. UNSB is trained according to Appendix B. We used official pre-trained models for EGSDE and Star GAN v2 and trained NOT from scratch using the official code. We use the evaluation protocol in CUT (Park et al., 2020) to evaluate all models, i.e., we use test set images to measure FID. We note that (Zhao et al., 2022) and (Choi et al., 2022) use a different evaluation protocol in their papers, (they use train set images to measure FID) so the numbers in our paper and their papers may differ. Results are shown in Table 5 and Figure 11. We note that UNSB produces outputs closer to the target image domain, as evidenced by FID and KID. Also, EGSDE, Star GAN v2, and NOT often fail to preserve input structure information. For instance, EGSDE fails to preserve eye direction or hair style. Star GAN v2 hallucinates earrings and fails to preserve hair style. NOT changes mouth shape (teeth are showing for outputs). C.4 ENTROPIC OT BENCHMARK Method ENOT Ours p(x) c FID 40.5 63.0 104 Table 6: EOT benchmark. We evaluate UNSB on the entropic OT benchmark provided by Gushchin et al. (2023c) for 64 64 resolution images. We use ϵ = 0.1 (using the notation in (Gushchin et al., 2023c)) and follow their training and evaluation protocol. Specifically, we trained UNSB to translate noisy Celeb A images (i.e., y-samples) to clean Celeb A images (i.e., x-samples). For comparison, we provide c FID numbers for SB p(x|y) learned by ENOT (Gushchin et al., 2023b) and unconditional p(x|y) = p(x) where we use 5k ground-truth samples from p(x). c FID for ENOT is taken from Table 7 in Gushchin et al. (2023c). In Table 6, we see that UNSB performs much better than the unconditional map p(x|y) = p(x) but not quite as well as ENOT. We emphasize that this is only a preliminary result with ad-hoc hyper-parameter choices due to time constraints, and we expect better results with further tuning. Some points of improvement are: Using more samples to measure c FID. We only used 100 x-samples for each y-sample. The benchmark provides 5k x-samples for each y-sample, so using more x-samples per y-sample could improve the c FID. Better architecture design. UNSB architecture is specialized to 256 256 resolution images, whereas the benchmark consists of 64 64 resolution images. Currently, we set UNSB networks for this task as resize(x, 64) G(x) resize(x, 256), where resize(x, s) is a function which resizes images to s s resolution, and G(x) is the UNSB network used in our main experiments. Better regularization. CUT regularization is specialized to translation between clean images, whereas the benchmark consists of translation between noisy and clean images. Longer training. We only trained UNSB for 150k iterations. On, for instance, Male2Female, we needed to train about 1.7m iterations to get good results. Published as a conference paper at ICLR 2024 Figure 13: Additional samples on Horse2Zebra and Summer2Winter. C.5 ZEBRA2HORSE TRANSLATION Figure 12: Zebra2Horse samples. Method Time NFE FID KID Cycle GAN 0.004 1 135.75 2.998 Ours 0.045 5 134.09 2.581 Table 7: Zebra2Horse results. Disc. Reg. NFE FID KID 5 254.2 23.91 Instance 1 185.5 8.732 Patch 1 155.5 7.697 Patch 1 75.6 0.491 Patch 5 73.9 0.421 Table 8: Summer2Winter ablation. We show that UNSB is capable of reverse translation on tasks in our paper as well. Specifically, we compare Cycle GAN and UNSB on Zebra2Horse translation task. In Table 7, we see that UNSB beats Cycle GAN in terms of both FID and KID. In Figure 12, we observe UNSB successfully preserves input structure as well. C.6 MORE ABLATION STUDY In Table 8, we provide another ablation study result on the Summer2Winter translation task, where we observe similar trends as before. Adding Markovian discrminator, regularization, and multi-step sampling monotonically increases the performance. This tells us the components of UNSB play orthogonal roles in unpaired image-to-image translation. C.7 ADDITIONAL SAMPLES 256 256 images. In Fig. 13, 14, we show additional generated samples from our proposed UNSB and baseline models for our tasks in our main paper. 512 512 images. In Fig. 15, we show samples from UNSB for Cat2Dog translation with AFHQ 512 512 resolution images, NFE = 5. This is a preliminary result with UNSB at training epoch 100. Published as a conference paper at ICLR 2024 Figure 14: Additional samples on Label2Cityscapes and Map2Satellite dataset. Figure 15: Cat2Dog on AFHQ 512 512 with UNSB. Preliminary result at training epoch 100. Published as a conference paper at ICLR 2024 D DISCUSSION ON SCENE-LEVEL VS. OBJECT-LEVEL TRANSFER We note that unpaired image-to-image translation tasks can roughly be categorized into scenelevel transfer and object-level transfer. In scene-level transfer (e.g., Horse2Zebra, Summer2Winter, etc.), mostly color and texture need to be altered during translation. In object-level transfer (e.g., Male2Female, Cat2Dog, etc.), object identities need to be altered during translation. We discuss differences between UNSB and diffusion-based translation methods on scene-level and object-level transfer tasks, respectively. Scene-level transfer. While diffusion-based methods such as SDEdit are capable of translating between high-resolution images, we note that they do not perform so well on scene-level translation tasks considered in our paper. For instance, see Table 2 for SDEdit and P2P results on Horse2Zebra, where they show 97.3 FID and 60.9 FID, respectively. UNSB achieves 35.7 FID. Figure 5a provides some clue as to why diffusion-based methods do not perform so well. P2P fails to preserve structural integrity (in the first row, the direction of the right horse is reversed) and SDEdit fails to properly translate images to zebra domain. Similar phenomenon occurs for Summer2Winter task as well. On the other hand, UNSB does not suffer from this problem as it uses a Markovian discriminator along with CUT regularization. Object-level transfer. Diffusion-based methods perform reasonably on object-level translation tasks such as Male2Female or Wild2Cat. We speculate this is because object-level translation tasks use images with cropped and centered subjects, so it is relatively easy to preserve structural integrity. In fact, as shown in Figure 16, semantically meaningful pairs on such dataset pairs can be obtained even with Sinkhorn-Knopp (a computational optimal transport method), which does not rely on deep learning. However, Sinkhorn-Knopp is unable to find meaningful pairs on Horse2Zebra, where subjects are not cropped and centered. (a) Wild2Cat (b) Horse2Zebra Figure 16: Unpaired image-to-image translation results with Sinkhorn-Knopp.