# _decentralized_attribution_of_generative_models__4268043e.pdf Published as a conference paper at ICLR 2021 DECENTRALIZED ATTRIBUTION OF GENERATIVE MODELS Changhoon Kim1, , Yi Ren2, , Yezhou Yang1 School of Computing, Informatics, and Decision Systems Engineering1 School for Engineering of Matter, Transport, and Energy2 Arizona State University {kch, yiren, yz.yang}@asu.edu Growing applications of generative models have led to new threats such as malicious personation and digital copyright infringement. One solution to these threats is model attribution, i.e., the identification of user-end models where the contents under question are generated from. Existing studies showed empirical feasibility of attribution through a centralized classifier trained on all user-end models. However, this approach is not scalable in reality as the number of models ever grows. Neither does it provide an attributability guarantee. To this end, this paper studies decentralized attribution, which relies on binary classifiers associated with each user-end model. Each binary classifier is parameterized by a user-specific key and distinguishes its associated model distribution from the authentic data distribution. We develop sufficient conditions of the keys that guarantee an attributability lower bound. Our method is validated on MNIST, Celeb A, and FFHQ datasets. We also examine the trade-off between generation quality and robustness of attribution against adversarial post-processes.1 1 INTRODUCTION Figure 1: FFHQ dataset projected to the space spanned by two keys φ1 and φ2. We develop sufficient conditions for model attribution: Perturbing the authentic dataset along different keys with mutual angles larger than a data-dependent threshold guarantees attributability of the perturbed distributions. (a) A threshold of 90 deg suffices for benchmark datasets (MNIST, Celeb A, FFHQ). (b) Smaller angles may not guarantee attributability. Recent advances in generative models (Goodfellow et al., 2014) have enabled the creation of synthetic contents that are indistinguishable even by naked eyes (Pathak et al., 2016; Zhu et al., 2017; Zhang et al., 2017; Karras et al., 2017; Wang et al., 2018; Brock et al., 2018; Miyato et al., 2018; Choi et al., 2018; Karras et al., 2019a;b; Choi et al., 2019). Such successes raised serious concerns regarding emerging threats due to the applications of generative models (Kelly, 2019; Breland, 2019). This paper is concerned about two particular types of threats, namely, malicious personation (Satter, 2019) , and digital copyright infringement. In the former, the attacker uses generative models to create and disseminate inappropriate or illegal contents; in the latter, the attacker steals the ownership of a copyrighted content (e.g., an art piece created through the assistance of a generative model) by making modifications to it. We study model attribution, a solution that may address both threats. Model attribution is defined as the identification of user-end models where the contents under question are generated from. Existing Equal contribution. 1https://github.com/ASU-Active-Perception-Group/decentralized_ attribution_of_generative_models Published as a conference paper at ICLR 2021 studies demonstrated empirical feasibility of attribution through a centralized classifier trained on all existing user-end models (Yu et al., 2018). However, this approach is not scalable in reality where the number of models ever grows. Neither does it provide an attributability guarantee. To this end, we propose in this paper a decentralized attribution scheme: Instead of a centralized classifier, we use a set of binary linear classifiers associated with each user-end model. Each classifier is parameterized by a user-specific key and distinguishes its associated model distribution from the authentic data distribution. For correct attribution, we expect one-hot classification outcomes for generated contents, and a zero vector for authentic data. To achieve correct attribution, we study the sufficient conditions of the user-specific keys that guarantee an attributability lower bound. The resultant conditions are used to develop an algorithm for computing the keys. Lastly, we assume that attackers can post-process generated contents to potentially deny the attribution, and study the tradeoff between generation quality and robustness of attribution against post-processes. Problem formulation We assume that for a given dataset D Rdx, the registry generates userspecific keys, Φ := {φ1, φ2, ...} where φi Rdx and ||φi|| = 1. || || is the l2 norm. A user-end generative model is denoted by Gφ( ; θ) : Rdz Rdx where z and x are the latent and output variables, respectively, and θ are the model parameters. When necessary, we will suppress θ and φ to reduce the notational burden. The dissemination of the user-end models is accompanied by a public service that tells whether a query content belongs to Gφ (labeled as 1) or not (labeled as 1). We model the underlying binary linear classifier as fφ(x) = sign(φT x). Note that linear models are necessary for the development of sufficient conditions of attribution presented in this paper, although sufficient conditions for nonlinear classifiers are worth exploring in the future. The following quantities are central to our investigation: (1) Distinguishability of Gφ measures the accuracy of fφ(x) at classifying Gφ against D: 2Ex PGφ,x0 PD [1(fφ(x) = 1) + 1(fφ(x0) = 1)] . (1) Here PD is the authentic data distribution, and PGφ the user-end distribution dependent on φ. G is (1 δ)-distinguishable for some δ (0, 1] when D(G) 1 δ. (2) Attributability measures the averaged multi-class classification accuracy of each model distribution over the collection G := {Gφ1, ..., GφN }: i=1 Ex Gφi1(φT j x < 0, j = i, φT i x > 0). (2) G is (1 δ)-attributable when A(G) 1 δ. (3) Lastly, We denote by G( ; θ0) (or shortened as G0) the root model trained on D, and assume PG0 = PD. We will measure the (lack of) generation quality of Gφ by the FID score (Heusel et al., 2017) and the l2 norm of the mean output perturbation: x(φ) = Ez Pz[Gφ(z; θ) G(z; θ0)], (3) where Pz is the latent distribution. This paper investigates the following question: What are the sufficient conditions of keys so that the user-end generative models can achieve distinguishability individually and attributability collectively, while maintaining their generation quality? Contributions We claim the following contributions: 1. We develop sufficient conditions of keys for distinguishability and attributability, which connect these metrics with the geometry of the data distribution, the angles between keys, and the generation quality. 2. The sufficient conditions lead to simple design rules for the keys: keys should be (1) data compliant, i.e., φT x < 0 for x PD, and (2) orthogonal to each other. We validate these rules using DCGAN (Radford et al., 2015) and Style GAN (Karras et al., 2019a) on benchmark datasets including MNIST (Le Cun & Cortes, 2010), Celeb A (Liu et al., 2015), and FFHQ (Karras et al., 2019a). See Fig. 1 for a visualization of the attributable distributions perturbed from the authentic FFHQ dataset. 3. We empirically test the tradeoff between generation quality and robust attributability under random post-processes including image blurring, cropping, noising, JPEG conversion, and a combination of all. Published as a conference paper at ICLR 2021 2 SUFFICIENT CONDITIONS FOR ATTRIBUTABILITY From the definitions (Eq. (1) and Eq. (2)), achieving distinguishability is necessary for attributability. In the following, we first develop the sufficient conditions for distinguishability through Proposition 1 and Theorem 1, and then those for attributability through Theorem 2. Distinguishability through watermarking First, consider constructing a user-end model Gφ by simply adding a perturbation x to the outputs of the root model G0. Assuming that φ is datacompliant, this model can achieve distinguishability by solving the following problem with respect to x: min || x|| ε Ex PD max{1 φT (x + x), 0} , (4) where ε > 0 represents a generation quality constraint. The following proposition reveals the connection between distinguishability, data geometry, and generation quality (proof in Appendix A): Proposition 1. Let dmax(φ) := maxx PD |φT x|. If ε 1 + dmax(φ), then x = (1 + dmax(φ))φ solves Eq. (4), and fφ(x + x ) > 0, x PD. Watermarking through retraining user-end models The perturbation x can potentially be reverse engineered and removed when generative models are white-box to users (e.g., when models are downloaded by users). Therefore, we propose to instead retrain the user-end models Gφ using the perturbed dataset Dγ,φ := {G0(z)+γφ | z Pz} with γ > 0, so that the perturbation is realized through the model architecture and weights. Specifically, the retraining fine-tunes G0 so that Gφ(z) matches with G0(z) + γφ for z Pz. Since this matching will not be perfect, we use the following model to characterize the resultant Gφ: Gφ(z) = G0(z) + γφ + ϵ, (5) where the error ϵ N(µ, Σ). In Sec. 3 we provide statistics of µ and Σ on the benchmark datasets, to show that the retraining captures the perturbations well (µ close to 0 and small variances in Σ). Updating Proposition 1 due to the existence of ϵ leads to Theorem 1, where we show that γ needs to be no smaller than dmax(φ) in order for Gφ to achieve distinguishability (proof in Appendix B): Theorem 1. Let dmax(φ) = maxx D |φT x|, σ2(φ) = φT Σφ, δ [0, 1], and φ be a data-compliant key. D(Gφ) 1 δ/2 if γ dmax(φ) + σ(φ) Remarks The computation of σ(φ) requires Gφ, which in turn requires γ. Therefore, an iterative search is needed to determine γ that is small enough to limit the loss of generation quality, yet large enough for distinguishability (see Alg. 1). Attributability We can now derive the sufficient conditions for attributability of the generative models from a set of N keys (proof in Appendix C): Theorem 2. Let dmin = minx D |φT x|, dmax = maxx D |φT x|, σ2(φ) = φT Σφ, δ [0, 1]. Let a(φ, φ ) := 1 + dmax(φ ) + dmin(φ ) 2φ T µ δ2 + dmax(φ ) φ T µ , (7) for keys φ and φ . Then A(G) 1 Nδ, if D(G) 1 δ for all Gφ G, and φT φ a(φ, φ ) (8) for any pair of data-compliant keys φ and φ . Remarks When σ(φ ) is negligible for all φ and µ = 0, a(φ, φ ) is approximately dmin(φ )/dmax(φ ) > 0, in which case φT φ 0 is sufficient for attributability. In Sec. 3 we empirically show that this approximation is plausible for the benchmark datasets. Published as a conference paper at ICLR 2021 Figure 2: (a) Validation of Theorem 1: All points should be close to the diagonal line or to its right. (b) Support for orthogonal keys: Min. RHS value of Eq.(7) for all keys are either positive (MNIST) or close to zero (Celeb A). (c,d) Statistics of µ and Σ for two sample user-end models for MNIST and Celeb A. Small µ and small diag(Σ) suggest good match of Gφ to the perturbed data distributions. (e-h) Distinguishability, attributability, perturbation length, and orthogonality of 100 Style GAN user-end models on FFHQ and 100 DCGAN user-end models on MNIST or Celeb A, respectively. 3 EXPERIMENTS AND ANALYSIS In this section we test Theorem 1, provide empirical support for the orthogonality of keys, and present experimental results on model attribution using MNIST, Celeb A, and FFHQ. Note that tests on the theorems require estimation of Σ, which is costly for models with high-dimensional outputs, and therefore are only performed on MNIST and Celeb A. Key generation We generate keys by iteratively solving the following convex problem: φi = arg min φ Ex PD,G0 max{1 + φT x, 0} + j=1 max{φT j φ, 0}. (9) The orthogonality penalty is omitted for the first key. The solutions are normalized to unit l2 norm before being inserted into the next problem. We note that PD and PG0 do not perfectly match in practice, and therefore we draw with equal chance from both distributions during the computation. G0s are trained using the standard DCGAN architecture for MNIST and Celeb A, and Style GAN for FFHQ. Training details are deferred to Appendix D. User-end generative models The training of Gφ follows Alg. 1, where γ is iteratively tuned to balance generation quality and distinguishability. For each γ, we collect a perturbed dataset Dγ,φ and solve the following training problem: min θ E(z,x) Dγ,φ ||Gφ(z; θ) x||2 , (10) starting from θ = θ0. If the resultant model does not meet the distinguishability requirement due to the discrepancy between Dγ,φ and Gφ, the perturbation is updated as γ = αγ. In experiments, we use a standard normal distribution for Pz, and set δ = 10 2 and α = 1.1. Validation of Theorem 1 Here we validate the sufficient condition for distinguishability. Fig. 2a compares the LHS and RHS values of Eq. (6) for 100 distinguishable user-end models. The empirical distinguishability of these models are reported in Fig. 2e. Calculation of the RHS of Eq. (6) requires estimations of µ and Σ. To do this, we sample ϵ(z) = Gφ(z; θ) G(z; θ0) γφ (11) Published as a conference paper at ICLR 2021 using 5000 samples of z Pz, where Gφ and γ are derived from Alg. 1. Σ and µ are then estimated for each φ. Fig. 2c and d present histograms of the elements in µ and Σ for two user-end models of the benchmark datasets. Results in Fig. 2a show that the sufficient condition for distinguishability (Eq. (6)) is satisfied for most of the sampled models through the training specified in Alg. 1. Lastly, we notice that the LHS values for MNIST are farther away from the equality line than those for Celeb A. This is because the MNIST data distribution resides at corners of the unit box. Therefore perturbations of the distribution are more likely to exceed the bounds for pixel values. Clamping of these invalid pixel values reduces the effective perturbation length. Therefore to achieve distinguishability, Alg. 1 seeks γs larger than needed. This issue is less observed in Celeb A, where data points are rarely close to the boundaries. Fig. 2g present the values of γs of all user-end models. Algorithm 1: Training of Gφ input : φ, G0 output: Gφ, γ 1 set γ = dmax(φ) ; 2 collect Dγ,φ ; 3 train Gφ by solving Eq. (10) using Dγ,φ ; 4 compute empirical D(Gφ) ; 5 if D(Gφ) < 1 δ then 6 set γ = αγ ; 7 goto step 2 ; Validation of Theorem 2 Recall that from Theorem 2, we recognized that orthogonal keys are sufficient. To support this design rule, Fig. 2b presents the minimum RHS values of Eq. (8) for 100 user-end models. Specifically, for each φi, we compute a(φi, φj) (Eq. (7)) using φj for j = 1, ..., i 1 and report minj a(φi, φj), which sets an upper bound on the angle between φi and all existing φs. The resultant minj a(φi, φj) are all positive for MNIST and close to zero for Celeb A. From this result, an angle of 94 deg, instead of 90 deg, should be enforced between any pairs of keys for Celeb A. However, since the conditions are sufficient, orthogonal keys still empirically achieve high attributability (Fig. 2f), although improvements can be made by further increasing the angle between keys. Also notice that the current computation of keys (Eq. (9)) does not enforce a hard constraint on orthogonality, leading to slightly acute angles (87.7 deg) between keys for Celeb A (Fig. 2h). On the other hand, the positive values in Fig. 2b for MNIST suggests that further reducing the angles between keys is acceptable if one needs to increase the total capacity of attributable models. However, doing so would require the derivation of new keys to rely on knowledge about all existing user-end models (in order to compute Eq. (7)). Table 1: Empirical average of distinguishability ( D),attributability (A(G)), || x||, and FID scores. DCGANM (DCGANC) for MNIST (Celeb A). Std in parenthesis. FID0: FID for G0. means lower is better and means higher is better. GANs Angle D A(G) || x|| FID0 FID DCGANM Orthogonal 0.99 0.99 3.97 (0.29) 7.82 (0.15) 10.43 (0.77) 45 degree 0.99 0.13 3.85 (0.12) - 11.02 (0.86) DCGANC Orthogonal 0.99 0.93 4.04 (0.32) 35.95 (0.12) 58.69 (5.21) 45 degree 0.99 0.15 4.57 (0.35) - 59.81 (5.34) Style GAN Orthogonal 0.99 0.99 36.04 (23.35) 12.43(0.16) 35.23(14.67) 45 degree 0.97 0.18 67.19 (35.53) - 47.86(10.66) Empirical results on benchmark datasets Tab. 1 reports the metrics of interest measured on the 100 user-end models for each of MNIST and Celeb A, and 20 models for FFHQ. All models are trained to be distinguishable. And by utilizing Theorem 2, they also achieve high attributability. As a comparison, we demonstrate results where keys are 45 deg apart (φT φ = 0.71) using a separate set of 20 user-end models for each of MNIST and Celeb A, and 5 models for FFHQ, in which case distinguishability no longer guarantees attributability. Regarding generation quality, Gφs receive worse FID scores than G0 due to the perturbations. We visualize samples from user-end models and the corresponding keys in Fig. 3. Note that for human faces, FFHQ in particular, the perturbations create light shades around eyes and lips, which is an unexpected but reasonable result. Attribution robustness vs. generation quality We now consider the scenario where outputs of the generative models are post-processed (e.g., by adversaries) before being attributed. When the post-processes are known, we can take counter measures through robust training, which intuitively Published as a conference paper at ICLR 2021 Figure 3: Visualization of sample keys (1st row) and the corresponding user-end generated contents. will lead to additional loss of generation quality. To assess this tradeoff between robustness and generation quality, we train Gφ against post-processes T : Rdx Rdx from a distribution PT . Due to the potential nonlinearity of T and the lack of theoretical guarantee in this scenario, we resort to the following robust training problem for deriving the user-end models: min θi Ez Pz,T PT max{1 fφi(T(Gφi(z; θi))), 0} + C||G0(z) Gφi(z; θi)||2 , (12) where C is the hyper-parameter for generation quality. Detailed analysis and comparison for selecting C are provided in Appendix E. We consider five types of post-processes: blurring, cropping, noise, JPEG conversion and the combination of these four. Examples of the post-processed images are shown in Fig. 5. Blurring uses Gaussian kernel widths uniformly drawn from 1 3{1, 3, 5, 7, 9}. Cropping crops images with uniformly drawn ratios between 80% and 100%, and scales the cropped images back to the original size using bilinear interpolation. Noise adds white noise with standard deviation uniformly drawn from [0, 0.3]. JPEG applies JPEG compression. Combination performs each attack with a 50% chance in the order of Blurring, Cropping, Noise and JPEG. For differentiability, we use existing implementations of differentiable blurring (Riba et al. (2020)) and JPEG conversion (Zhu et al. (2018)). For robust training, we apply the post-process to mini-batches with 50% probability. We performed comprehensive tests using DCGAN (on MNIST and Celeb A), PGAN (on Celeb A), and Cycle GAN (on Cityscapes). Tab. 2 summarizes the average distinguishability, the attributability, the perturbation length || x||, and the FID score with and without robust training of Gφ. Results are based on 20 models for each architecture-dataset pair, where keys are kept orthogonal and data compliant. From the results, defense against these post-processes can be achieved, except for Combination. Importantly, there is a clear tradeoff between robustness and generation quality. This can be seen from Fig. 5, which compares samples with7 and without robust training from the tested models and datasets. Lastly, it is worth noting that the training formulation in Eq. (12) can also be applied to the training of non-robust user-end models in place of Eq. (10). However, the resultant model from Eq. (12) cannot be characterized by Eq. (5) with small µ and Σ, i.e., due to the nonlinearity of the training process of Eq. (12, the user-end model distribution is deformed while it is perturbed. This resulted in unsuccessful validation of the theorems, which led to the adoption of Eq. (10) for theorem-consistent training. Therefore, while the empirical results show feasibility of achieving robust attributability using Eq. (12, counterparts to Theorems 1 and 2 in this nonlinear setting are yet to be developed. Published as a conference paper at ICLR 2021 Table 2: DCGANM: MNIST. DCGANC: Celeb A. Dis.: Distinguishability before (Bfr) and after (Aft) robust training. Att.: Attributability. || x|| and FID are after robust training. Std in parenthesis. means lower is better and means higher is better. || x|| and FID before robust training: DCGANM:|| x|| = 5.05, FID = 5.36. DCGANC:|| x|| = 5.63, FID = 53.91. PGAN: || x|| = 9.29, FID = 21.62. Cycle GAN: || x|| = 55.85. FID does not apply to Cycle GAN. Metric Model Blurring Cropping Noise JPEG Combi. - - Bfr Aft Bfr Aft Bfr Aft Bfr Aft Bfr Aft DCGANM 0.49 0.96 0.52 0.99 0.85 0.99 0.54 0.99 0.50 0.66 DCGANC 0.49 0.99 0.49 0.99 0.95 0.98 0.51 0.99 0.50 0.85 PGAN 0.50 0.98 0.51 0.99 0.97 0.99 0.96 0.99 0.50 0.76 Cycle GAN 0.49 0.92 0.49 0.87 0.98 0.99 0.55 0.99 0.49 0.67 DCGANM 0.02 0.94 0.03 0.88 0.77 0.95 0.16 0.98 0.00 0.26 DCGANC 0.00 0.98 0.00 0.99 0.89 0.93 0.07 0.98 0.00 0.70 PGAN 0.13 0.99 0.07 0.99 0.97 0.99 0.99 0.99 0.06 0.98 Cycle GAN 0.08 0.98 0.05 0.93 0.97 0.98 0.47 0.99 0.05 0.73 DCGANM 15.96(2.18) 9.17(0.65) 5.93(0.34) 6.48(0.94) 17.08(1.86) DCGANC 11.83(0.65) 9.30(0.31) 4.75(0.17) 6.01(0.29) 13.69(0.59) PGAN 18.49(2.04) 21.27(0.81) 10.20(0.81) 10.08(1.03) 24.82(2.33) Cycle GAN 68.03(3.62) 80.03(3.59) 55.47(1.60) 57.42(2.00) 83.94(4.66) FID DCGANM 41.11(20.43) 21.58(2.44) 5.79(0.19) 6.50(1.70) 68.16(24.67) DCGANC 73.62(6.70) 98.86(9.51) 59.51(1.60) 60.35(2.57) 87.29(9.29) PGAN 28.15(3.43) 47.94(5.71) 25.43(2.19) 22.86(2.06) 45.16(7.87) 4 DISCUSSION Figure 4: Capacity of keys as a sphere packing problem: The feasible space (arc) is determined by the data compliance and generation quality constraints, and the size of spheres by the minimal angle between keys. Capacity of keys For real-world applications, we hope to maintain attributability for a large set of keys. Our study so far suggests that the capacity of keys is constrained by the data compliance and orthogonality requirements. While the empirical study showed the feasibility of computing keys through Eq. (9), finding the maximum number of feasible keys is a problem about optimal sphere packing on a segment of the unit sphere (Fig. 4). To explain, the unit sphere represents the identifiability requirement ||φ|| = 1. The feasible segment of the unit sphere is determined by the data compliance and generation quality constraints. And the spheres to be packed have radii following the sufficient condition in Theorem 2. Such optimal packing problems are known open challenges (Cohn et al. (2017); Cohn (2016)). For real-world applications where a capacity of attributable models is needed (which is the case for both malicious personation and copyright infringement settings), it is necessary to find approximated solutions to this problem. Generation quality control From Proposition 1 and Theorem 1, the inevitable loss of generation quality is directly related to the length of perturbation (γ), which is related to dmax. Fig. 6 compares outputs from user-end models with different dmaxs. While it is possible to filter φs based on their corresponding dmaxs for generation quality control, here we discuss a potential direction for prescribing a subspace of φs within which quality can be controlled. To start, we denote by J(x) the Jacobian of G0 with respect to its generator parameters θ0. Our discussion is related to the matrix M = Ex PG0[J(x)]Ex PG0[J(x)T ]. A spectral analysis of M reveals that the eigenvectors of M with large eigenvalues are more structured than those with small ones (Fig. 7(a)). This finding is consistent with the definition of M: The largest eigenvectors of M represent the principal axes of all mean sensitivity vectors, where the mean is taken over the latent space. For MNIST, these eigenvectors overlap with the digits; for Celeb A, they are structured color patterns. On the other hand, the smallest eigenvectors represent directions rarely covered by the sensitivity vectors, thus Published as a conference paper at ICLR 2021 Figure 5: Samples from user-end models with robust and non-robust training. For each subfigure - top: DCGAN on MNIST and Celeb A; bottom: PGAN (Celeb A) and Cycle GAN (Cityscapes). For each dataset - top: samples from G0 (after worst-case post-process in (b-f)); mid: samples from Gφ (after robust training in (b-f)); btm (a): difference between non-robust Gφ and G0; btm (b-h) difference between robust and non-robust Gφ. Figure 6: MNIST, Celeb A, and FFHQ examples from Gφs with (a-c) small dmax and (d-f) large dmax. All models are distinguishable and attributable. (Zooming in on pdf file is recommended.) resembling random noise. Based on this finding, we test the hypothesis that keys more aligned with the eigenspace of the small eigenvalues will have smaller dmax. We test this hypothesis by computing the Pearson correlations between dmax and φT Mφ using 100 models for each of MNIST and Celeb A. The resultant correlations are 0.33 and 0.53, respectively. In addition, we compare outputs from models using the largest and the smallest eigenvectors of M as the keys in Fig. 7b. While a concrete human study is needed, the visual results suggest that using eigenvectors of M is a promising approach to segmenting the space of keys according to their induced generation quality. 5 RELATED WORK Detection and attribution of model-generated contents This paper focused on the attribution of contents from generative models rather than the detection of hand-crafted manipulations (Agarwal & Farid (2017); Popescu & Farid (2005); O brien & Farid (2012); Rao & Ni (2016); Huh et al. (2018)). Detection methods rely on fingerprints intrinsic to generative models (Odena et al. (2016); Zhang Published as a conference paper at ICLR 2021 Figure 7: (a) Eigenvectors for the two largest and two smallest eigenvalues of M for DCGANs on MNIST (top) and Celeb A (bottom). (b) Left column: Samples from G0; Rest: G0 Gφ where φ are the eigenvectors in (a). et al. (2019b); Marra et al. (2019); Wang et al. (2019)), yet similar fingerprints present for models trained on similar datasets (Marra et al. (2019)). Thus fingerprints cannot be used for attribution. Skripniuk et al. (2020) studied decentralized attribution. Instead of using linear classifiers for attribution, they train a watermark encoder-decoder network that embeds (and reads) watermarks into (and from) the content, and compare the decoded watermark with user-specific ones. Their method does not provide sufficient conditions of the watermarks for attributability. IP protection of digital contents and models Watermarks have conventionally been used for IP protection (Tirkel et al., 1993; Van Schyndel et al., 1994; Bi et al., 2007; Hsieh et al., 2001; Pereira & Pun, 2000; Zhu et al., 2018; Zhang et al., 2019a) without considering the attribution guarantee. Another approach to content IP protection is blockchain (Hasan & Salah, 2019). However, this approach requires meta data to be transferred along with the contents, which may not be realistic in adversarial settings. E.g., one can simply take a picture of a synthetic image to remove any meta data attached to the image file. Aside from the protection of contents, mechanisms for protecting IP of models have also been studied (Uchida et al., 2017; Nagai et al., 2018; Le Merrer et al., 2019; Adi et al., 2018; Zhang et al., 2018; Fan et al., 2019; Szyller et al., 2019; Zhang et al., 2020). Model watermarking is usually done by adding watermarks into model weights (Uchida et al., 2017; Nagai et al., 2018), by embedding unique input-output mapping into the model (Le Merrer et al., 2019; Adi et al., 2018; Zhang et al., 2018), or by introducing a passport mechanism so that model accuracy drops if the right passport is not inserted (Fan et al., 2019). While closely related, existing work on model IP protection focused on the distinguishability of individual models, rather than the attributability of a model set. 6 CONCLUSION Motivated by emerging challenges with generative models, e.g., deepfake, this paper investigated the feasibility of decentralized attribution of such models. The study is based on a protocol where the registry generates user-specific keys that guides the watermarking of user-end models to be distinguishable from the authentic data. The outputs of user-end models will then be attributed by the registry through the binary classifiers parameterized by the keys. We developed sufficient conditions of the keys so that distinguishable user-end models achieve guaranteed attributability. These conditions led to simple rules for designing the keys. With concerns about adversarial postprocesses, we further showed that robust attribution can be achieved using the same design rules, and with additional loss of generation quality. Lastly, we introduced two open challenges towards real-world applications of the proposed attribution scheme: the prescription of the key space with controlled generation quality, and the approximation of the capacity of keys. 7 ACKNOWLEDGEMENTS Support from NSF Robust Intelligence Program (1750082), ONR (N00014-18-1-2761), and Amazon AWS MLRA is gratefully acknowledged. We would like to express our gratitude to Ni Trieu (ASU) for providing us invaluable advice, and Zhe Wang, Joshua Feinglass, Sheng Cheng, Yongbaek Cho and Huiliang Shao for helpful comments. Published as a conference paper at ICLR 2021 Yossi Adi, Carsten Baum, Moustapha Cisse, Benny Pinkas, and Joseph Keshet. Turning your weakness into a strength: Watermarking deep neural networks by backdooring. In 27th {USENIX} Security Symposium ({USENIX} Security 18), pp. 1615 1631, 2018. Shruti Agarwal and Hany Farid. Photo forensics from jpeg dimples. In 2017 IEEE Workshop on Information Forensics and Security (WIFS), pp. 1 6. IEEE, 2017. Ning Bi, Qiyu Sun, Daren Huang, Zhihua Yang, and Jiwu Huang. Robust image watermarking based on multiband wavelets and empirical mode decomposition. IEEE Transactions on Image Processing, 16(8):1956 1966, 2007. Ali Breland. The bizarre and terrifying case of the deepfake video that helped bring an african nation to the brink. motherjones, Mar 2019. URL https://www.motherjones.com/ politics/2019/03/deepfake-gabon-ali-bongo/. Andrew Brock, Jeff Donahue, and Karen Simonyan. Large scale gan training for high fidelity natural image synthesis. ar Xiv preprint ar Xiv:1809.11096, 2018. Yunjey Choi, Minje Choi, Munyoung Kim, Jung-Woo Ha, Sunghun Kim, and Jaegul Choo. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8789 8797, 2018. Yunjey Choi, Youngjung Uh, Jaejun Yoo, and Jung-Woo Ha. Stargan v2: Diverse image synthesis for multiple domains. ar Xiv preprint ar Xiv:1912.01865, 2019. Henry Cohn. A conceptual breakthrough in sphere packing. ar Xiv preprint ar Xiv:1611.01685, 2016. Henry Cohn, Abhinav Kumar, Stephen D Miller, Danylo Radchenko, and Maryna Viazovska. The sphere packing problem in dimension 24. Annals of Mathematics, pp. 1017 1033, 2017. Lixin Fan, Kam Woh Ng, and Chee Seng Chan. Rethinking deep neural network ownership verification: Embedding passports to defeat ambiguity attacks. In Advances in Neural Information Processing Systems, pp. 4716 4725, 2019. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in neural information processing systems, pp. 2672 2680, 2014. Haya R Hasan and Khaled Salah. Combating deepfake videos using blockchain and smart contracts. Ieee Access, 7:41596 41606, 2019. Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems, pp. 6626 6637, 2017. Ming-Shing Hsieh, Din-Chang Tseng, and Yong-Huai Huang. Hiding digital watermarks using multiresolution wavelet transform. IEEE Transactions on industrial electronics, 48(5):875 882, 2001. Minyoung Huh, Andrew Liu, Andrew Owens, and Alexei A Efros. Fighting fake news: Image splice detection via learned self-consistency. In Proceedings of the European Conference on Computer Vision (ECCV), pp. 101 117, 2018. Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of gans for improved quality, stability, and variation. ar Xiv preprint ar Xiv:1710.10196, 2017. Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4401 4410, 2019a. Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. Analyzing and improving the image quality of stylegan. ar Xiv preprint ar Xiv:1912.04958, 2019b. Published as a conference paper at ICLR 2021 Makena Kelly. Congress grapples with how to regulate deepfakes. Congress grapples with how to regulate deepfakes, Jun 2019. URL https://www.theverge.com/2019/6/13/18677847/ deep-fakes-regulation-facebook-adam-schiff-congress-artificial-intelligence. Erwan Le Merrer, Patrick Perez, and Gilles Tr edan. Adversarial frontier stitching for remote neural network watermarking. Neural Computing and Applications, pp. 1 12, 2019. Yann Le Cun and Corinna Cortes. MNIST handwritten digit database. 2010. URL http://yann. lecun.com/exdb/mnist/. Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, pp. 3730 3738, 2015. Francesco Marra, Diego Gragnaniello, Luisa Verdoliva, and Giovanni Poggi. Do gans leave artificial fingerprints? In 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), pp. 506 511. IEEE, 2019. Takeru Miyato, Toshiki Kataoka, Masanori Koyama, and Yuichi Yoshida. Spectral normalization for generative adversarial networks. ar Xiv preprint ar Xiv:1802.05957, 2018. Yuki Nagai, Yusuke Uchida, Shigeyuki Sakazawa, and Shin ichi Satoh. Digital watermarking for deep neural networks. International Journal of Multimedia Information Retrieval, 7(1):3 16, 2018. James F O brien and Hany Farid. Exposing photo manipulation with inconsistent reflections. ACM Trans. Graph., 31(1):4 1, 2012. Augustus Odena, Vincent Dumoulin, and Chris Olah. Deconvolution and checkerboard artifacts. Distill, 2016. doi: 10.23915/distill.00003. URL http://distill.pub/2016/ deconv-checkerboard. Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, and Alexei A Efros. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2536 2544, 2016. Shelby Pereira and Thierry Pun. Robust template matching for affine resistant image watermarks. IEEE transactions on image Processing, 9(6):1123 1129, 2000. Alin C Popescu and Hany Farid. Exposing digital forgeries by detecting traces of resampling. IEEE Transactions on signal processing, 53(2):758 767, 2005. Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks, 2015. URL http://arxiv.org/abs/ 1511.06434. cite arxiv:1511.06434Comment: Under review as a conference paper at ICLR 2016. Yuan Rao and Jiangqun Ni. A deep learning approach to detection of splicing and copy-move forgeries in images. In 2016 IEEE International Workshop on Information Forensics and Security (WIFS), pp. 1 6. IEEE, 2016. E. Riba, D. Mishkin, D. Ponsa, E. Rublee, and G. Bradski. Kornia: an open source differentiable computer vision library for pytorch, 2020. URL https://arxiv.org/pdf/1910. 02190.pdf. Raphael Satter. Experts: Spy used ai-generated face to connect with targets. Experts: Spy used AI-generated face to connect with targets, Jun 2019. URL https://apnews.com/ bc2f19097a4c4fffaa00de6770b8a60d. Vladislav Skripniuk, Ning Yu, Sahar Abdelnabi, and Mario Fritz. Black-box watermarking for generative adversarial networks. ar Xiv preprint ar Xiv:2007.08457, 2020. Sebastian Szyller, Buse Gul Atli, Samuel Marchal, and N Asokan. Dawn: Dynamic adversarial watermarking of neural networks. ar Xiv preprint ar Xiv:1906.00830, 2019. Published as a conference paper at ICLR 2021 Anatol Z Tirkel, GA Rankin, RM Van Schyndel, WJ Ho, NRA Mee, and Charles F Osborne. Electronic watermark. Digital Image Computing, Technology and Applications (DICTA 93), pp. 666 673, 1993. Yusuke Uchida, Yuki Nagai, Shigeyuki Sakazawa, and Shin ichi Satoh. Embedding watermarks into deep neural networks. In Proceedings of the 2017 ACM on International Conference on Multimedia Retrieval, pp. 269 277, 2017. Ron G Van Schyndel, Andrew Z Tirkel, and Charles F Osborne. A digital watermark. In Proceedings of 1st International Conference on Image Processing, volume 2, pp. 86 90. IEEE, 1994. Sheng-Yu Wang, Oliver Wang, Andrew Owens, Richard Zhang, and Alexei A Efros. Detecting photoshopped faces by scripting photoshop. In Proceedings of the IEEE International Conference on Computer Vision, pp. 10072 10081, 2019. Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Andrew Tao, Jan Kautz, and Bryan Catanzaro. Highresolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8798 8807, 2018. Ning Yu, Larry Davis, and Mario Fritz. Attributing fake images to gans: Analyzing fingerprints in generated images. ar Xiv preprint ar Xiv:1811.08180, 2018. Han Zhang, Tao Xu, Hongsheng Li, Shaoting Zhang, Xiaogang Wang, Xiaolei Huang, and Dimitris N. Metaxas. Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017. Jialong Zhang, Zhongshu Gu, Jiyong Jang, Hui Wu, Marc Ph Stoecklin, Heqing Huang, and Ian Molloy. Protecting intellectual property of deep neural networks with watermarking. In Proceedings of the 2018 on Asia Conference on Computer and Communications Security, pp. 159 172, 2018. Jie Zhang, Dongdong Chen, Jing Liao, Han Fang, Weiming Zhang, Wenbo Zhou, Hao Cui, and Nenghai Yu. Model watermarking for image processing networks. In AAAI, pp. 12805 12812, 2020. Kevin Alex Zhang, Alfredo Cuesta-Infante, Lei Xu, and Kalyan Veeramachaneni. Steganogan: High capacity image steganography with gans. ar Xiv preprint ar Xiv:1901.03892, 2019a. Xu Zhang, Svebor Karaman, and Shih-Fu Chang. Detecting and simulating artifacts in gan fake images. ar Xiv preprint ar Xiv:1907.06515, 2019b. Jiren Zhu, Russell Kaplan, Justin Johnson, and Li Fei-Fei. Hidden: Hiding data with deep networks. In Proceedings of the European Conference on Computer Vision (ECCV), pp. 657 672, 2018. Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision, pp. 2223 2232, 2017. Published as a conference paper at ICLR 2021 A PROOF OF PROPOSITION 1 Proposition 1. Let dmax(φ) := maxx PD |φT x|. If ε 1 + dmax(φ), then x = (1 + dmax(φ))φ solves Eq. (4), and fφ(x + x) > 0 x G0. Proof. Let φ be a data-compliant key and let x be sampled from PD. First, from the KKT conditions for Eq. (4) we can show that the solution x is proportional to φ: x = φ/µ , (13) where µ 0 is the Lagrange multiplier. To minimize the objective, we seek µ such that 1 (x + x )T φ = 1 x T φ 1/µ 0, (14) for all x. Since x T φ < 0 (data compliance), this requires 1/µ = 1 + dmax(φ). Therefore, when ε 1 + dmax(φ), x = (1 + dmax(φ))φ solves Eq. (4). And fφ(x + x ) = φT (x + (1 + dmax(φ))φ) = φT x + 1 + dmax(φ) > 0. B PROOF OF THEOREM 1 Theorem 1. Let dmax(φ) = maxx D |φT x|, σ2(φ) = φT Σφ, δ [0, 1], and φ be a data-compliant key. D(Gφ) 1 δ/2 if + dmax(φ) φT µ. (15) Proof. We first note that due to data compliance of keys, Ex PD 1(φT x < 0) = 1. Therefore D(Gφ) 1 δ/2 iff Ex PGφ 1(φT x > 0) 1 δ, i.e., Pr(φT x > 0) 1 δd for x PGφ. We now seek a lower bound for Pr(φT x > 0). To do so, let x and x0 be sampled from PGφ and PG0, respectively. Then we have φT x = φT (x0 + γφ + ϵ) = φT x0 + γ + φT ϵ, (16) and Pr(φT x > 0) = Pr φT ϵ > φT x0 γ . (17) Since dmax(φ) φT x0, we have Pr(φT x > 0) Pr φT ϵ > dmax(φ) γ = Pr φT (ϵ µ) γ dmax(φ) + φT µ . (18) The latter sign switching in equation 18 is granted by the symmetry of the distribution of φT (ϵ µ), which follows N(0, φT Σφ). A sufficient condition for Pr(φT x > 0) 1 δ is then Pr φT (ϵ µ) γ dmax(φ) + φT µ 1 δ. (19) Recall the following tail bound of x N(0, σ2) for y 0: Pr(x σy) 1 exp( y2/2). (20) Compare equation 20 with equation 19, the sufficient condition becomes + dmax(φ) φT µ. (21) Published as a conference paper at ICLR 2021 C PROOF OF THEOREM 2 Theorem 2. Let dmin = minx D |φT x|, dmax = maxx D |φT x|, σ2(φ) = ||φ||2 Σ, δ [0, 1]. A(G) 1 Nδ if D(G) 1 δ for all Gφ G and for any pair of data-compliant keys φ and φ : φT φ 1 + dmax(φ ) + dmin(φ ) 2φ T µ δ2 + dmax(φ ) φ T µ . (22) Proof. Let φ and φ be any pair of keys. Let x and x0 be sampled from PGφ and PG0, respectively. We first derive the sufficient conditions for Pr(φ T x < 0) 1 δ. Since x = x0 + γφ + ϵ for x Gφ, we have φ T x = φ T (x0 + γφ + ϵ) = φ T x0 + γφT φ + φ T ϵ. (23) Then Pr(φ T x < 0) = Pr φ T ϵ < φ T x0 γφT φ Pr φ T (ϵ µ) < dmin(φ ) γφT φ φ T µ , (24) where dmin(φ ) := minx D |φ T x| and φ T (ϵ µ) N(0, σ2(φ )). Using the same tail bound of normal distribution and Theorem 1, we have Pr(φT x < 0) 1 δ if γφT φ σ(φ ) dmin(φ ) + φ T µ φT φ 1 + dmax(φ ) + dmin(φ ) 2φ T µ δ2 + dmax(φ ) φ T µ Note that Pr(A = 1, B = 1) = 1 Pr(A = 0) Pr(B = 0) + Pr(A = 0, B = 0) 1 Pr(A = 0) Pr(B = 0) for binary random variables A and B. With this, it is straight forward to show that when Pr(φ T x < 0) 1 δ for all φ = φ, and Pr(φT x > 0) 1 δ for all φ, then Pr(φT x > 0, φ T x < 0 φ = φ) 1 Nδ and A(G) 1 Nδ. D TRAINING DETAILS We trained user-end models based on the objective function (Eq.(10) in the main text). For datasets where the root models follow DCGAN and PGAN, the user-end models follow the same architecture. For the FFHQ dataset where Style GAN is used, we introduce an additional shallow convolutional network as a residual part, which is added to the original Style GAN output to match with the perturbed datasets Dγ,φ. In this case, the training using Eq.(10) is limited to the additional shallow network, while the Style GAN weights are frozen. More specifically, denoting the combination of convolution, Re LU, and max-pooling by Conv-Re LU-Max, the shallow network consists of three Conv-Re LU-Max blocks and one fully connected layer. All of the convolution layers have 4 x 4 kernels, stride 2, and padding 1. And all of the max-pooling layers have 3 x 3 kernels and stride 2. D.2 PARAMETERS We adopt the Adam optimizer for training. Training hyper-parameters are summarized in Table 3. Published as a conference paper at ICLR 2021 Table 3: Hyper-parameters to train keys (φ) and generators (Gφ). GANs Dataset Batch Size Learning Rate β1 β2 Epochs DCGAN MNIST 16 0.001 0.9 0.99 10 DCGAN Celeb A 64 0.001 0.9 0.99 2 Style GAN FFHQ 8 0.001 0.9 0.99 5 D.3 TRAINING TIME All experiments are conducted on V100 Tesla GPUs. Table 4 summarizes the number of GPUs used and the training time for the non-robust models (Eq.(10) in the main text) and robust models (Eq.(12) in the main text). Recall that we chose Eq.(10) for training the non-robust user-end models for consistency with the theorems, although Eq.(12) can be used to achieve attributability in practice, as is shown in the robust attribution study. Therefore, the non-robust training takes longer to due the iteration of γ in Alg. 1. Table 4: Training time (in minute) of one key (Eq.(9) in main text) and one generator (Eq.(10) in main text). DCGANM: DCGAN for MNIST, DCGANC: DCGAN for Celeb A. GANs GPUs Key Non-robust Blurring Cropping Noise JPEG Combination DCGANM 1 1.77 14 4.12 3.96 4.19 5.71 5.12 DCGANC 1 5.31 15 10.33 9.56 10.35 10.25 10.76 PGAN 2 50.89 141.07 140.05 131.90 133.46 132.46 135.07 Cycle GAN 1 20.88 16.04 16.26 15.43 15.71 15.98 16.41 Style GAN 1 54.23 3.12 - - - - - E ABLATION STUDY Here we conduct an ablation study on the hyper-parameter C for the robust training formulation (Eq.(12)). Training with larger C focuses more on generation quality, thus sacrificing distinguishability and attributability. These effects are reported in Table 5 and Table 6. Due to limited time, the results here are averaged over five models for each C and data-model pairs. Published as a conference paper at ICLR 2021 Table 5: Distinguishability (top), attributability (btm) before (Bfr) and after (Aft) robust training. DCGANM: DCGAN for MNIST, DCGANC: DCGAN for Celeb A. Model C Blurring Cropping Noise JPEG Combination - - Bfr Aft Bfr Aft Bfr Aft Bfr Aft Bfr Aft DCGANM 10 0.49 0.97 0.51 0.99 0.84 0.99 0.53 0.99 0.50 0.63 DCGANM 100 0.49 0.61 0.51 0.98 0.76 0.98 0.53 0.99 0.50 0.52 DCGANM 1K 0.49 0.50 0.51 0.81 0.69 0.91 0.53 0.97 0.50 0.51 DCGANC 10 0.49 0.99 0.49 0.99 0.96 0.99 0.50 0.99 0.49 0.85 DCGANC 100 0.50 0.96 0.49 0.99 0.92 0.93 0.50 0.99 0.49 0.61 DCGANC 1K 0.50 0.62 0.49 0.97 0.88 0.91 0.50 0.99 0.49 0.51 PGAN 100 0.50 0.98 0.50 0.99 0.96 0.99 0.96 0.99 0.50 0.81 PGAN 1K 0.50 0.89 0.49 0.95 0.94 0.95 0.88 0.99 0.50 0.60 PGAN 10K 0.50 0.61 0.50 0.76 0.89 0.90 0.76 0.98 0.50 0.51 Cycle GAN 1K 0.49 0.92 0.50 0.87 0.98 0.99 0.55 0.99 0.49 0.62 Cycle GAN 10K 0.49 0.70 0.50 0.66 0.94 0.96 0.52 0.98 0.50 0.51 DCGANM 10 0.02 0.94 0.03 0.88 0.77 0.95 0.16 0.98 0.00 0.26 DCGANM 100 0.00 0.87 0.00 0.85 0.73 0.90 0.10 0.95 0.00 0.13 DCGANM 1K 0.00 0.75 0.00 0.80 0.63 0.80 0.10 0.91 0.00 0.05 DCGANC 10 0.00 0.98 0.00 0.99 0.89 0.93 0.07 0.98 0.00 0.70 DCGANC 100 0.00 0.95 0.00 0.93 0.82 0.85 0.02 0.93 0.00 0.61 DCGANC 1K 0.00 0.90 0.00 0.89 0.77 0.81 0.00 0.88 0.00 0.43 PGAN 100 0.26 1.00 0.21 1.00 0.99 0.99 0.99 0.99 0.00 0.99 PGAN 1K 0.21 0.99 0.00 0.99 0.97 0.98 0.98 0.99 0.00 0.54 PGAN 10K 0.00 0.51 0.00 0.90 0.90 0.92 0.83 0.99 0.00 0.22 Cycle GAN 1K 0.00 0.99 0.00 0.97 0.97 0.99 0.45 0.99 0.00 0.77 Cycle GAN 10K 0.00 0.87 0.00 0.77 0.95 0.96 0.30 0.99 0.00 0.31 Table 6: || x|| (top) and FID score (btm). Standard deviations in parenthesis. DCGANM: DCGAN for MNIST, DCGANC: DCGAN for Celeb A, Combi.: Combination attack. Lower is better. Model C Baseline Blurring Cropping Noise JPEG Combi. DCGANM 10 5.05(0.09) 15.96(2.18) 9.17(0.65) 5.93(0.34) 6.48(0.94) 17.08(1.86) DCGANM 100 4.09(0.53) 12.95(4.47) 7.62(1.55) 4.57(0.78) 4.70(1.02) 12.70(3.37) DCGANM 1K 3.88(0.60) 7.17(2.10) 7.43(1.37) 4.22(0.77) 5.12(1.94) 7.56(1.41) DCGANC 10 5.63(0.11) 11.83(0.65) 9.30(0.31) 4.75(0.17) 6.01(0.29) 13.69(0.59) DCGANC 100 3.08(0.27) 10.00(1.61) 7.80(0.58) 3.20(0.45) 4.26(0.59) 11.65(1.48) DCGANC 1K 2.55(0.36) 7.68(1.53) 7.13(0.47) 2.65(0.24) 3.39(0.58) 9.23(1.22) PGAN 100 9.29(0.95) 18.49(2.04) 21.27(0.81) 10.20(0.81) 10.08(1.03) 24.82(2.33) PGAN 1K 6.52(1.85) 14.79(4.15) 18.88(1.96) 6.40(1.48) 7.09(1.62) 22.09(2.12) PGAN 10K 5.04(1.63) 10.19(2.87) 18.23(0.94) 5.13(1.14) 5.67(1.62) 17.26(1.39) Cycle GAN 1K 55.85(3.67) 68.03(3.62) 80.03(3.59) 55.47(1.60) 57.42(2.00) 83.94(4.66) Cycle GAN 10K 49.66(5.01) 58.64(3.70) 66.05(3.47) 53.14(0.44) 54.52(2.30) 66.24(5.29) DCGANM 10 5.36(0.12) 41.11(20.43) 21.58(2.44) 5.79(0.19) 6.50(1.70) 68.16(24.67) DCGANM 100 5.32(0.11) 23.83(14.29) 18.39(3.70) 5.41(0.18) 5.46(0.11) 36.05(16.20) DCGANM 1K 5.23(0.12) 10.85(4.28) 18.08(1.77) 5.37(0.14) 5.30(0.96) 21.86(4.16) DCGANC 10 53.91(2.20) 73.62(6.70) 98.86(9.51) 59.51(1.60) 60.35(2.57) 87.29(9.29) DCGANC 100 45.02(3.37) 73.12(11.03) 85.50(12.25) 47.60(2.57) 50.48(4.58) 78.11(12.95) DCGANC 1K 40.85(3.41) 55.63(7.97) 72.11(13.81) 40.87(3.03) 45.46(5.03) 57.13(7.20) PGAN 100 21.62(1.73) 28.15(3.43) 47.94(5.71) 25.43(2.19) 22.86(2.06) 45.16(7.87) PGAN 1K 19.05(3.14) 25.19(5.26) 43.48(12.24) 19.20(2.96) 19.05(2.82) 35.07(8.72) PGAN 10K 16.75(1.87) 18.96(2.65) 37.01(8.74) 16.94(1.89) 17.39(2.33) 26.63(4.44)