# instructg2i_synthesizing_images_from_multimodal_attributed_graphs__45bcb22f.pdf INSTRUCTG2I: Synthesizing Images from Multimodal Attributed Graphs Bowen Jin, Ziqi Pang, Bingjun Guo, Yu-Xiong Wang, Jiaxuan You, Jiawei Han Department of Computer Science University of Illinois at Urbana-Champaign bowenj4@illinois.edu https://instructg2i.github.io/ In this paper, we approach an overlooked yet critical task Graph2Image: generating images from multimodal attributed graphs (MMAGs). This task poses significant challenges due to the explosion in graph size, dependencies among graph entities, and the need for controllability in graph conditions. To address these challenges, we propose a graph context-conditioned diffusion model called INSTRUCTG2I. INSTRUCTG2I first exploits the graph structure and multimodal information to conduct informative neighbor sampling by combining personalized page rank and re-ranking based on vision-language features. Then, a Graph QFormer encoder adaptively encodes the graph nodes into an auxiliary set of graph prompts to guide the denoising process of diffusion. Finally, we propose graph classifier-free guidance, enabling controllable generation by varying the strength of graph guidance and multiple connected edges to a node. Extensive experiments conducted on three datasets from different domains demonstrate the effectiveness and controllability of our approach. The code is available at https://github.com/Peter Griffin Jin/Instruct G2I. 1 Introduction This paper investigates an overlooked yet critical source of information for image generation: the pervasive graph-structured relationships of real-world entities. In contrast to the commonly adopted language conditioning in models represented by Stable Diffusion [32], graph connections have combinatorial complexity and cannot be trivially captured as a sequence. Such graph-structured relationships among the entities are expressed through Multimodal Attributed Graphs (MMAGs), where nodes are enriched with image and text information. As a concrete example (Figure 1(a)), the graph of artworks is constructed by nodes containing images (pictures) and texts (titles), as well as edges corresponding to shared genre and authorship. Such a graph uniquely depicts a piece of artwork by its thousands of peers in the graph, beyond the mere description of language. To this end, we formulate and propose the Graph2Image challenge, requiring the generative models to synthesize image conditioning on both text descriptions and graph connections of a node. This task featuring the image generation on MMAGs is well-grounded in real-world applications. For instance, generating an image for a virtual artwork node in the art MMAG is akin to creating virtual artwork according to the nuanced styles of artists and genres [5] (as in Figure 1(a)). Similarly, generating an image for a product node connected to other products through co-purchase links in an e-commerce MMAG equates to recommending future products for users [24]. Without surprise, our exploiting the graph-structured information indeed improves the consistency of generated images compared to models only using texts or images as conditioning (Figure 1(b)). 38th Conference on Neural Information Processing Systems (Neur IPS 2024). Composition 8 The Artist s Garden The Blue Rider Murnau with Church Wassily Kandinsky Claude Monet Impressionism Virtual Painting House in Snow 50% Monet 50% Kandinsky (a)Multimodal Attribute Graph & Graph2Image (Virtual Artist Application) Stable Diffusion Instruct Pix2Pix Control Net Instruct G2I (Ours) w/o Graph Conditioning w/ Graph Conditioning (b) Image Synthesis Comparison Text Condition Graph Condition Monet Kandinsky (c) Controllable Generation Figure 1: We propose a new task Graph2Image featuring image synthesis by conditioning on graph information and introduce a novel graph-conditioned diffusion model called INSTRUCTG2I to tackle this problem. (a) Graph2Image is supported by prevalent multimodal attributed graphs and is grounded in real-world applications, e.g., virtual artistry. (b) INSTRUCTG2I outperforms baseline image generation techniques, demonstrating the usefulness of graph information. (c) To accommodate realistic user queries, INSTRUCTG2I exhibits smooth controllability in utilizing text/graph information and managing the strength of multiple graph edges. Despite the usefulness of graph information, existing methods conditioning on either text [32] or images [2, 41] are incapable of direct integration with MMAGs. Therefore, we propose a graph context-aware diffusion model INSTRUCTG2I inherited from Stable Diffusion that mitigates gaps. A most prominent challenge directly originates from the combinatorial complexity of graphs, which we term as Graph Size Explosion: inputting the entire local subgraph structure to a model, including all the images and texts, is impractical due to the exponential increase in size, especially with additional hops. Therefore, INSTRUCTG2I learns to compress the massive amounts of contexts from the graph into a set of graph conditioning tokens with fixed capacity, which functions alongside the common text conditioning tokens in Stable Diffusion. Such a compression process is enhanced with a semantic personalized pagerank-based graph sampling approach to actively select the most informative neighboring nodes based on both structural and semantic perspectives. Besides the number of contexts, the graph structures in MMAGs additionally specify the proximity of entities, which is not captured in conventional text or image conditioning. This challenge of Graph Entity Dependency reflects the implicit preference of image generation: synthesizing a shirt image linked to light-colored clothing is likely to have a pastel tone (image-image dependency), and generating a picture titled a running horse should reference interconnected animal images rather than scenic ones (text-image dependency). To enable the nuanced proximity understanding on graphs, we further improve our graph conditioning tokens via a Graph-QFormer architecture learning to encode the graph information guided by texts. Finally, we propose that our graph conditioning is a natural interface for controllable generation, reflecting the strength of edges in MMAGs. Take the virtual art generation (Figure 1(c)) for example: INSTRUCTG2I can flexibly offer different strengths of graph guidance and can smoothly transition between the style of Monet and Kandinsky, defined by its strength of connection with either of the two artists. Such an advantage is grounded for real-world application and is a plug-and-play test-time algorithm inspired by classifier-free guidance [18]. In sum, our contributions include: Formulation and Benchmark. We are the first to identify the usefulness of multimodal attributed graphs (MMAGs) in image synthesis and formulate the Graph2Image problem. Our formulation is supported by three benchmarks grounded in the real-world applications of art and e-commerce. Algorithm. Methodologically, we propose INSTRUCTG2I, a context-aware diffusion model that effectively encodes graph conditional information as graph prompts for controllable image generation (as shown in Figure 1(b,c)). Experiments and Evaluation. Empirically, we conduct experiments on graphs from three different domains, demonstrating that INSTRUCTG2I consistently outperforms competitive baselines (as shown in Figure 1(b)). 2 Problem Formulation 2.1 Multimodal Attributed Graphs Definition 1 (Multimodal Attributed Graphs (MMAGs)) A multimodal attributed graph can be defined as G p V, E, P, Dq, where V, E, P and D represent the sets of nodes, edges, images, and documents, respectively. Each node vi P V is associated with some textual information dvi P D and some image information pvi P P. For example, in an e-commerce product graph, nodes (v P V) represent products, edges (e P E) denote co-viewed semantic relationships, images (p P P) are product images, and documents (d P D) are product titles. Similarly, in an art graph (shown in Figure 1), nodes represent artworks, edges signify shared artists or genres, images are artwork pictures, and documents are artwork titles. In this work, we focus on graphs where edges provide semantic correlations between images (nodes). For instance, in an e-commerce product graph, connected products (those frequently co-viewed by many users) are highly related. Similarly, in an art graph, linked artworks (those created by the same author or within the same genre) are likely to have similar styles. 2.2 Problem Definition In this work, we explore the problem of node image generation on MMAGs. Given a node vi in an MMAG G, our objective is to generate pvi based on dvi and G. This problem has multiple real-world applications. For example, in e-commerce, it translates to generating the image (pvi) for a product (vi) based on a user query (dvi) and user purchase history (G), which is a generative retrieval task. In the art domain, it involves generating the picture (pvi) for an artwork (vi) based on its title (di) and its associated artist style or genre (G), which is a virtual artwork creation task. Definition 2 (Node Image Generation on MMAGs) In a multimodal attributed graph G p V, E, P, Dq, given a node vi P V within the graph G with a textual description dvi, the goal is to synthesize pvi, the corresponding image at vi, with a learned model ˆpvi fpvi, dvi, Gq. Our evaluation emphasizes instance-level similarity, assessing how closely ˆpvi matches pvi. We conduct evaluations on artwork graphs, e-commerce graphs, and literature graphs. More details can be found in Section 4.1. 3 Methodology In this section, we present our INSTRUCTG2I framework, overviewed in Figure 2. We begin by introducing graph conditions into stable diffusion models in Section 3.1. Next, we discuss semantic personalized Page Rank-based sampling to select informative graph conditions in Section 3.2. Furthermore, we propose Graph-QFormer to extract dependency-aware representations for graph conditions in Section 3.3. Finally, we introduce controllable generation to balance the condition scale between text and graph guidance, as well as manage multiple graph guidances in Section 3.4. 3.1 Graph Context-aware Stable Diffusion Stable Diffusion (SD). INSTRUCTG2I is built upon Stable Diffusion (SD). SD conducts diffusion in the latent space, where an input image x is first encoded from pixel space into a latent representation z Encpxq. A decoder then transfers the latent representation z1 back to the pixel space, yielding x1 Decpz1q. The diffusion model generates the latent representation z1 conditioned on a text prompt c T . The training objective of SD is defined as follows: L Ez Encpxq,c T ,ϵ Np0,1q,t }ϵ ϵθpzt, t, hpc T qq}2 . (1) Target Text Prompt e.g. House in Snow Text Encoder (Stable Diffusion) Graph Q-former Image Encoder x N Self-Attention Cross-Attention Graph Prompt Tokens Text Prompt Tokens Graph2Image Diffusion (Denoising 𝑇Timestamps) Conditioning (c) Instruct G2I Model (a) Training Setup Text Prompt e.g. House in Snow Structural Relevance Personalized Page Rank (PPR) Semantic Relevance Similarity-based Re-ranking Vision Language Feature Similarity Target Text Prompt (b) Semantic PPR-based Neighbor Sampling Figure 2: The overall framework of INSTRUCTG2I. (a) Given a target node with a text prompt (e.g., House in Snow) in a Multimodal Attributed Graph (MMAG) for which we want to generate an image, (b) we first perform semantic PPR-based neighbor sampling, which involves structureaware personalized Page Rank and semantic-aware similarity-based reranking to sample informative neighboring nodes in the graph. (c) These neighboring nodes are then inputted into a Graph-QFormer, encoded by multiple self-attention and cross-attention layers, represented as graph tokens and used to guide the denoising process of the diffusion model, together with text prompt tokens. At each timestep t, the denoising network ϵθp q predicts the noise by conditioning on the current latent representation zt, timestep t and text prompt vectors hpc T q. To compute hpc T q P Rdˆlc T , where lc T is the length of c T and d is the hidden dimension, the text prompt c T is processed by the CLIP text encoder [31]: hpc T q CLIPpc T q. Introducing Graph Conditions into SD. In the context of MMAGs, synthesizing the image for a node vi involves not only the text dvi, but also the semantic information from the node s proximity on the graph. Therefore, we introduce an auxiliary set of graph conditioning tokens h Gpc Gq to the SD models (as shown in Figure 2(c)), working in parallel with the existing text conditions h T pc T q. hpc T , c Gq rh T pc T q, h Gpc Gqs P Rdˆplc T lc Gq, (2) where lc G is the length of the graph condition. The training objective then becomes: L Ez Encpxq,c T ,c G,ϵ Np0,1q,t }ϵ ϵθpzt, t, hpc T , c Gqq}2 . (3) For h T pc T q, we can directly use the CLIP text encoder as in the original SD. However, determining c G and h Gp q is more complex. We will discuss the details of c G and h Gp q in the following sections. 3.2 Semantic PPR-based Neighbor Sampling A straightforward approach to developing c Gpviq involves using the entire local subgraph of vi. However, this is impractical due to the exponential growth in size with each additional hop, leading to excessively long context sequences. To address this, we leverage both graph structure and node semantics to select informative c G. Structure Proximity: Personalized Page Rank (PPR). Inspired by [10], we first adopt PPR [15] to identify related nodes from a graph structure perspective. PPR processes the graph structure to derive a ranking score Pi,j for each node vj relative to node vi, where a higher Pi,j indicates a greater degree of similarity between vi and vj. Let P P Rnˆn be the PPR matrix of the graph, where each row Pi,: represents a PPR vector toward a target node vi. The matrix P is determined by: P β ˆ AP p1 βq I. (4) where β is the reset probability for PPR and ˆ A is the normalized adjacency matrix. Once P is computed, we define the PPR-based graph condition c Gppr of node vi as the top-Kppr PPR neighbors of node vi: c Gpprpviq argmax c GpprpviqĂV,|c Gpprpviq| Kppr vj Pc Gpprpviq Pi,j. (5) Semantic Proximity: Similarity-based Reranking. However, solely relying on PPR may result in a graph condition set containing images (e.g., scenery pictures) that are not semantically related to our target node (e.g., a picture titled running horse ). To address this, we propose using a semantic-based similarity calculation function Simpd, pq (e.g., CLIP) to rerank vj P c Gpprpviq based on the relatedness of pvj to dvi. The final graph condition c Gpviq is calculated by: c Gpviq argmax c GpviqĂc Gpprpviq,|c Gpviq| K vj Pc Gpviq Simpdvi, pvjq. (6) 3.3 Graph Encoding with Text Conditions After we derive c Gpviq from the previous step, the problem comes to how can we design h Gp q to extract meaningful representations from c Gpviq. Here we focus more on how to utilize the image features from c Gpviq (i.e., tpvj|vj P c Gpviqu) since we find they are more informative for vi image generation compared with text features from c Gpviq (i.e., tdvj|vj P c Gpviqu) (shown in Section 4.3). Simple Baseline: Encoding with Pretrained Image Encoders [31]. A straightforward way to obtain representations for vj P c Gpviq is to directly apply some pretrained image encoders gimgp q (e.g., CLIP [31]): hvj gimgppvjq P Rd, h Gpc Gpviqq rhvjsvj Pc Gpviq P Rdˆlc G , (7) where denotes the concatenation operation. However, this simple design has two significant limitations: 1) The encoding for each pvjpvj P c Gpviqq is isolated from others in c Gpviq and failed to capture the image-image graph dependency. For example, the style extraction from one picture (pvj) can benefit from the other pictures created by the same artist (in c Gpviq). 2) The encoding for each pvj is independent to dvi, which fails to capture the text-image graph dependency. For example, when we are creating a picture titled running horse (dvi), it is desired to offer more weight on horse pictures in c Gpviq rather than scenery pictures. Graph-QFormer. To address these limitations, we propose Graph-QFormer as h Gp q to learn representations for c G while considering the graph dependency information. As shown in Figure 2, Graph-QFormer consists of two Transformer [35] modules motivated by [26]: (1) a self-attention module that facilitates deep mutual information exchange between previous layer hidden states, capturing image-image dependencies and (2) a cross-attention module that weights samples in c G using text guidance, capturing text-image dependencies. Let Hptq c Gpviq P Rdˆlc G denote the hidden states outputted by the t-th Graph-QFormer layer. We use the token embeddings of dvi as the input query embeddings to provide text guidance: Hp0q c Gpviq rx1, ..., x|dvi|s. (8) where xk is the k-th token embedding in dvi and lc G |dvi|. The multi-head self-attention layer (MHASAT) is calculated by H1ptq c Gpviq MHASATrq Hpt 1q c Gpviq, k Hpt 1q c Gpviq, v Hpt 1q c Gpviqs, (9) where q, k, v denotes query, key, and value channels in the Transformer. The output H1ptq c Gpviq is then inputted to the multi-head cross-attention layer (MHACAT), calculated by Hptq c Gpviq MHACATrq H1ptq c Gpviq, k Zc Gpviq, v Zc Gpviqs, (10) where Zc Gpviq rgimgppvjqsvj Pc Gpviq P Rdˆn represents the image embeddings extracted from a fixed pretrained image encoder and n is the number of embeddings. Finally we adopt h Gpc Gpviqq Hp Lq c Gpviq, where L is the number of layers in Graph-QFormer. Connection between INSTRUCTG2I and GNNs. As illustrated in Figure 2, INSTRUCTG2I employs a Transformer-based architecture as the graph encoder. However, it can also be interpreted as a Graph Neural Network (GNN) model. GNN models [38] primarily use a propagation-aggregation paradigm to obtain node representations (Npiq denotes the neighbor set of i): apl 1q ij PROPplq hpl 1q i , hpl 1q j , @j P Npiq ; hplq i AGGplq hpl 1q i , tapl 1q ij |j P Npiqu . Similarly, in INSTRUCTG2I, Eq.(4)(5)(6) can be regarded as the propagation function PROPplq, while the aggregation step AGGplq corresponds to the combination of Eq.(9) and Eq.(10). 3.4 Controllable Generation The concept of classifier-free guidance, introduced by [18], enhances the performance of conditional image synthesis by modifying the noise prediction, eθp q, with the output from an unconditional model. This is formulated as: ˆϵθpzt, cq ϵθpzt, q s pϵθpzt, cq ϵθpzt, qq, where spą 1q is the guidance scale. The intuition is that ϵθ learns the gradient of the log image distribution and increasing the contribution of ϵθpcq ϵθp q will enlarge the convergence to the distribution conditioned on c. In our task, the score network ˆϵθpzt, c G, c T q is conditioned on both text c T di and the graph condition c G. We compose the score estimates from these two conditions and introduce two guidance scales, s T and s G, to control the contribution strength of c T and c G to the generated samples respectively. Our modified score estimation function is: ˆϵθpzt, c G, c T q ϵθpzt, , q s T pϵθpzt, , c T q ϵθpzt, , qq s G pϵθpzt, c G, c T q ϵθpzt, , c T qq. (11) For cases requiring fine-grained control over multiple graph conditions (i.e., different edges), we extend the formula as follows: ˆϵθpzt, c G, c T q ϵθpzt, , q s T pϵθpzt, , c T q ϵθpzt, , qq ÿ spkq G pϵθpzt, cpkq G , c T q ϵθpzt, , c T qq, (12) where cpkq G is the k-th graph condition. For example, to create an artwork that combines the styles of Monet and Van Gogh, the neighboring artworks by Monet and Van Gogh on the graph would be cp1q G and cp2q G , respectively. Further details on the derivation of our classifier-free guidance formulations can be found in Appendix A.3. 4 Experiments 4.1 Experimental Setups Datasets. We conduct experiments on three MMAGs from distinct domains: ART500K [27], Amazon [16], and Goodreads [37]. ART500K is an artwork graph with nodes representing artworks and edges indicating same-author or same-genre relationships. Each artwork node includes a title (text) and a picture (image). Amazon is a product graph where nodes represent products and edges denote co-view relationships. Each product is associated with a title (text) and a picture (image). Goodreads is a literature graph where nodes represent books and edges convey similar-book semantics. Each book node contains a title and a front cover image. Dataset statistics can be found in Appendix A.4. Baselines. We compare INSTRUCTG2I with two groups of baselines: 1) Text-to-image methods: This includes Stable Diffusion 1.5 (SD-1.5) [32] and SD 1.5 fine-tuned on our datasets (SD-1.5 FT). 2) Image-to-image methods: This includes Instruct Pix2Pix [2] and Control Net [41], both initialized with SD 1.5 and fine-tuned on our datasets. We use the most relevant neighbor, as selected in Section 3.2 as the input image for these baselines, allowing them to partially utilize graph information. Metrics. As indicated in Section 2.2, our evaluation mainly concerns the consistency of synthesized images with the ground truth image on the node. Therefore, our evaluation adopts the CLIP [31] and DINOv2 [29] score for instance-level similarity, in addition to the conventional FID [17] metric for image generation. For the CLIP and DINOv2 scores, we utilize CLIP and DINOv2 to obtain representations for both the generated and ground truth images and then calculate their cosine similarity. For FID, we calculate the distance between the distribution of the ground truth images and the distribution of the generated images. 4.2 Main results Quantitative Evaluation. The quantitative results are presented in Table 1 and Figure 3. From Table 1, we observe the following: 1) INSTRUCTG2I consistently outperforms all the baseline methods, highlighting the importance of graph information in image synthesis on MMAGs. 2) Although Instruct Pix2Pix and Control Net partially consider graph context, they fail to capture the semantic signals from the graph comprehensively. In Figure 3, we plot the aver- Table 1: Quantitative evaluation of different methods on ART500K, Amazon, and Goodreads datasets. The CLIP score denotes the image-image score. INSTRUCTG2I significantly outperforms the best baseline with p-value < 0.05 and consistently outperforms all the other common baselines in image synthesis, supporting the benefits of graph conditioning. ART500K Amazon Goodreads Model CLIP score DINOv2 score CLIP score DINOv2 score CLIP score DINOv2 score SD-1.5 58.83 25.86 60.67 32.61 42.16 14.84 SD-1.5 FT 66.55 34.65 65.30 41.52 45.81 18.97 Instruct Pix2Pix 65.66 33.44 63.86 41.31 47.30 20.94 Control Net 64.93 32.88 59.88 34.05 42.20 19.77 INSTRUCTG2I 73.73 46.45 68.34 51.70 50.37 25.54 Prompt: Thicker(fuller(hair(instantly(thick(serum Prompt: The(Crater(and(The(Clouds Ground-truth Sampled Neighbors (a) Ours (b) Stable Diffusion (c) Instruct Pix2Pix (d) Control Net Prompt: Painting of(My(Wife(And(Daughter Figure 4: Qualitative evaluation. Our method exhibits better consistency with the ground truth by better utilizing the graph information from neighboring nodes ( Sampled Neighbors in the figure). age DINOv2 (x-axis, Ò) and FID score (y-axis, Ó) across the three datasets. INSTRUCTG2I outperforms most baselines on both metrics and achieves the best trade-off between them. SD 1.5 SD 1.5 FT Instruct Pix2Pix Control Net 0.2 0.25 0.3 0.35 0.4 0.45 DINOv2 Score Figure 3: INSTRUCTG2I achieves the best trade-off between DINOv2 (Ò) and FID (Ó) scores. Instruct Pix2Pix obtains a better FID score than INSTRUCTG2I because it takes an in-distribution image as input, constraining the output image to stay close to the original distribution. Qualitative Evaluation. We conduct a qualitative evaluation by randomly selecting some generated cases. The results are shown in Figure 4, where we provide the sampled neighbor images from the graph, text prompts, and the ground truth images. From these results, we observe that INSTRUCTG2I generates images that best fit the semantics of the text prompt and context from the graph. For instance, when generating a picture for the crater and the clouds , the baselines either capture only the content ( crater and clouds ) without the style learned from the graph (Stable Diffusion and Instruct Pix2Pix) or adopt a similar style but lose the desired content (Control Net). In contrast, INSTRUCTG2I effectively learns from the neighbors on the graph and conveys the content accurately. Table 2: Ablation study on graph condition variants and Graph-QFormer. ART500K Amazon Goodreads Model CLIP score DINOv2 score CLIP score DINOv2 score CLIP score DINOv2 score INSTRUCTG2I 73.73 46.45 68.34 51.70 50.37 25.54 - Graph-QFormer 72.53 44.16 66.97 48.18 47.91 24.74 + Graph SAGE 72.26 43.06 66.07 43.40 46.68 21.91 + GAT 72.60 43.32 66.73 46.58 46.57 21.45 IP2P w. neighbor images 65.89 33.90 63.19 40.32 47.21 21.55 SD FT w. neighbor texts 69.72 38.64 65.55 43.51 47.47 22.68 Ours (Semantic PPR- based Sampling) Semantics-based Sampling PPR-based Sampling Random Sampling The Horse of the Frieze Text Prompt Ground Truth Sampled Neighbors Sampled Neighbors Figure 5: Ablation study on semantic PPR-based neighbor sampling. The results indicate that both structural and semantic relevance proposed by our method effectively improve the image generation quality and consistency with the graph context. 4.3 Ablation Study Study of Graph Condition for SD Variants. In INSTRUCTG2I, we introduce graph conditions into SD by encoding the images from c G into graph prompts, which serve as conditions together with text prompts for SD s denoising step. In this section, we demonstrate the significance of this design by comparing it with other variants that utilize graph conditions in SD: Instruct Pix2Pix (IP2P) with neighbor images and SD finetuned with neighbor texts. For the first variant, we perform mean pooling on the latent representations of images in c G, according to the IP2P s setting, and use this as the input image representation for IP2P. This variant has the same input information as INSTRUCTG2I. For the second variant, we utilize text information from neighbors instead of images, concatenate it with the text prompt, and fine-tune the SD. The results are shown in Table 2, where INSTRUCTG2I consistently outperforms both variants. This demonstrates the advantage of leveraging image features from c G and the effectiveness of our model design. Study of Graph-QFormer. We first demonstrate the effectiveness of Graph-QFormer by replacing it with the simple baseline mentioned in Eq.(7), denoted as - Graph-QFormer . We then compare it with graph neural network (GNN) baselines including Graph SAGE [13] and GAT [36], integrated into INSTRUCTG2I in the same manner. The results, presented in Table 2, show that INSTRUCTG2I with Graph-QFormer consistently outperforms both the ablated version and GNN baselines. This demonstrates the effectiveness of Graph-QFormer design. Study of the Semantic PPR-based Neighbor Sampling. We propose a semantic PPR-based sampling method that combines structure and semantics for neighbor sampling on graphs, as detailed in Section 3.2. In this section, we demonstrate the effectiveness of this approach by conducting ablation studies that remove either or both components. The results, shown in Figure 5, indicate that our sampling methods effectively identify neighbor images that contribute most significantly to the ground truth in both semantics and style. This underscores the value of integrating both structural and semantic information in our sampling approach. Text Guidance Graph Guidance Sampled Neighbors Text Prompt House&In& Wooded&Area (a) Text and graph guidance study. Pablo Picasso Gustave Courbet (b) Single or multiple graph guidance. Figure 6: Controllable generation study. (a) The ability of INSTRUCTG2I to balance text guidance and graph guidance. (b) Study of multiple graph guidance. Generated artworks with the input text prompt a man playing piano conditioned on single or multiple graph guidance (styles of Picasso and Courbet ). Please refer to Figure 1 for another example between Monet and Kandinsky. 4.4 Controllable Generation Text Guidance & Graph Guidance. In Eq.(11), we discuss the control of guidance from both text and graph conditions. To illustrate its effectiveness, we provide an example in Figure 6(a). The results show that as text guidance increases, the generated image incorporates more of the desired content. Conversely, as graph guidance increases, the generated image adopts a more desired style. This demonstrates the ability of our method to balance content and style through controlled guidance. Multiple Graph Guidance: Virtual Artist. In Eq.(12), we demonstrate how multiple graph guidance can be managed for controllable image generation. We present a use case, virtual artwork creation, to showcase its effectiveness (shown in Figure 6(b)). The goal of this task is to create an image that depicts specific content (e.g., a man playing piano) in the style of one or more artists (e.g., Picasso and Courbet). This is akin to adding a new node to the graph that links to the artwork nodes created by the specified artists and generating an image for this node. The results indicate that when single graph guidance is provided, the generated artwork aligns with that artist s style. As additional graph guidance is introduced, the styles of the two artists blend together. This demonstrates that our method offers the flexibility to meet various control requirements, effectively balancing different types of graph influences. 4.5 Model Behavior Analysis Cross-attention Weight Study in Graph-QFormer. We conduct a cross-attention study for Graph QFormer to understand how different sampled neighbors on the graph are selected based on the text prompt and contribute to the final image generation. We randomly select a case with the text prompt and neighbor images and plot the cross-attention weight map shown in Figure 7. From the weight map, we can find that Graph-QFormer learns to assign higher weight to pictures 1 and 4 which are related to raising and Lazarus in the text prompt respectively. The results indicate that Graph-QFormer effectively learns to select the images that are most relevant to the text prompt. 5 Related works Diffusion Models. Recent advancements in diffusion models have demonstrated significant success in generative applications. Diffusion models [4, 7] generate compelling examples through a step-wise denoising process, which involves a forward process that introduces noise into data distributions and a reverse process that reconstructs the original data [19]. A notable example is the Latent Diffusion Model (LDM) [32], which reduces computational costs by applying the diffusion process Sampled Neighbors The$ Raising$Of$ Lazarus$ Detail Text prompt Figure 7: Study of Graph-QFormer s cross-attention map. Graph-QFormer effectively learns to select the images that are most relevant to the text prompt. in a low-resolution latent space. In the domain of diffusion models, various forms of conditioning are employed to direct the generation process, including labels [6], classifiers [8], texts [28], images [2], and scene graphs [39]. These conditions can be incorporated into diffusion models through latent concatenation [33], cross-attention [1], and gradient control [12]. However, most existing works neglect the relational information between images and cannot be directly applied to image synthesis on MMAGs. Learning on Graphs. Early studies on learning on graphs primarily focus on representation learning for nodes or edges based on graph structures [3, 14]. Methods such as Deepwalk [30] and Node2vec [11] perform random walks on graphs to derive vector representation for each node. Graph neural networks (GNNs) [38, 43] are later introduced as a learnable component that incorporates both initial node features and graph structure. GNNs have been applied to various tasks, including classification [25], link prediction [42], and recommendation [21]. For instance, Graph SAGE [13] employs a propagation and aggregation paradigm for node representation learning, while GAT [36] introduces an attention mechanism into the aggregation process. Recently, research has increasingly focused on integrating text or image features with graph structures [22, 44]. For example, Patton [23] proposes pretraining language models on text-attributed graphs. However, these existing works mainly target representation learning on single-modal graphs and are not directly applicable to the image synthesis from multimodal attributed graph (MMAG) task addressed in this paper. 6 Conclusions In this paper, we identify the problem of image synthesis on multimodal attributed graphs (MMAGs). To address this challenge, we propose a graph context-conditioned diffusion model that: 1) Samples related neighbors on the graph using a semantic personalized Page Rank-based method; 2) Effectively encodes graph information as graph prompts by considering their dependency with Graph-QFormer; 3) Generates images under control with graph classifier-free guidance. We conduct systematic experiments on MMAGs in the domains of art, e-commerce, and literature, demonstrating the effectiveness of our approach compared to competitive baseline methods. Extensive studies validate the design of each component in INSTRUCTG2I and highlight its controllability. Future directions include joint text and image generation on MMAGs and capturing the heterogeneous relations between image and text units on MMAGs. Acknowledgments and Disclosure of Funding This work was supported by the Apple Ph D Fellowship. The research also was supported in part by US DARPA INCAS Program No. HR0011-21-C0165 and BRIES Program No. HR0011-24-3-0325, National Science Foundation IIS-19-56151, the Molecule Maker Lab Institute: An AI Research Institutes program supported by NSF under Award No. 2019897, and the Institute for Geospatial Understanding through an Integrative Discovery Environment (I-GUIDE) by NSF under Award No. 2118329. Any opinions, findings, and conclusions or recommendations expressed herein are those of the authors and do not necessarily represent the views, either expressed or implied, of DARPA or the U.S. Government. The views and conclusions contained in this paper are those of the authors and should not be interpreted as representing any funding agencies. [1] Namhyuk Ahn, Junsoo Lee, Chunggi Lee, Kunhee Kim, Daesik Kim, Seung-Hun Nam, and Kibeom Hong. Dreamstyler: Paint by style inversion with text-to-image diffusion models. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 674 681, 2024. 10 [2] Tim Brooks, Aleksander Holynski, and Alexei A Efros. Instructpix2pix: Learning to follow image editing instructions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18392 18402, 2023. 2, 6, 10 [3] Hongyun Cai, Vincent W Zheng, and Kevin Chen-Chuan Chang. A comprehensive survey of graph embedding: Problems, techniques, and applications. IEEE transactions on knowledge and data engineering, 30(9):1616 1637, 2018. 10 [4] Hanqun Cao, Cheng Tan, Zhangyang Gao, Yilun Xu, Guangyong Chen, Pheng-Ann Heng, and Stan Z Li. A survey on generative diffusion models. IEEE Transactions on Knowledge and Data Engineering, 2024. 9 [5] Eva Cetinic and James She. Understanding and creating art with ai: Review and outlook. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), 18(2): 1 22, 2022. 1 [6] Jian Chen, Ruiyi Zhang, Tong Yu, Rohan Sharma, Zhiqiang Xu, Tong Sun, and Changyou Chen. Label-retrieval-augmented diffusion models for learning from noisy labels. Advances in Neural Information Processing Systems, 36, 2024. 10 [7] Florinel-Alin Croitoru, Vlad Hondru, Radu Tudor Ionescu, and Mubarak Shah. Diffusion models in vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023. 9 [8] Prafulla Dhariwal and Alexander Nichol. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780 8794, 2021. 10 [9] Rohit Gandikota, Joanna Materzynska, Jaden Fiotto-Kaufman, and David Bau. Erasing concepts from diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2426 2436, 2023. 14 [10] Johannes Gasteiger, Aleksandar Bojchevski, and Stephan Günnemann. Predict then propagate: Graph neural networks meet personalized pagerank. ar Xiv preprint ar Xiv:1810.05997, 2018. 4 [11] Aditya Grover and Jure Leskovec. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pages 855 864, 2016. 10 [12] Yingqing Guo, Hui Yuan, Yukang Yang, Minshuo Chen, and Mengdi Wang. Gradient guidance for diffusion models: An optimization perspective. ar Xiv preprint ar Xiv:2404.14743, 2024. 10 [13] Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. In NIPS, pages 1024 1034, 2017. 8, 10 [14] William L Hamilton, Rex Ying, and Jure Leskovec. Representation learning on graphs: Methods and applications. ar Xiv preprint ar Xiv:1709.05584, 2017. 10 [15] Taher H Haveliwala. Topic-sensitive pagerank. In Proceedings of the 11th international conference on World Wide Web, pages 517 526, 2002. 4 [16] Ruining He and Julian Mc Auley. Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In WWW, pages 507 517, 2016. 6 [17] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017. 6 [18] Jonathan Ho and Tim Salimans. Classifier-free diffusion guidance. ar Xiv preprint ar Xiv:2207.12598, 2022. 2, 6 [19] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840 6851, 2020. 9 [20] Aapo Hyvärinen and Peter Dayan. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005. 14 [21] Bowen Jin, Chen Gao, Xiangnan He, Depeng Jin, and Yong Li. Multi-behavior recommendation with graph convolutional networks. In Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval, pages 659 668, 2020. 10 [22] Bowen Jin, Gang Liu, Chi Han, Meng Jiang, Heng Ji, and Jiawei Han. Large language models on graphs: A comprehensive survey. ar Xiv preprint ar Xiv:2312.02783, 2023. 10 [23] Bowen Jin, Wentao Zhang, Yu Zhang, Yu Meng, Xinyang Zhang, Qi Zhu, and Jiawei Han. Patton: Language model pretraining on text-rich networks. ar Xiv preprint ar Xiv:2305.12268, 2023. 10 [24] Wei Jin, Haitao Mao, Zheng Li, Haoming Jiang, Chen Luo, Hongzhi Wen, Haoyu Han, Hanqing Lu, Zhengyang Wang, Ruirui Li, et al. Amazon-m2: A multilingual multi-locale shopping session dataset for recommendation and text generation. Advances in Neural Information Processing Systems, 36, 2024. 1 [25] Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. ar Xiv preprint ar Xiv:1609.02907, 2016. 10 [26] Junnan Li, Dongxu Li, Silvio Savarese, and Steven Hoi. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In International conference on machine learning, pages 19730 19742. PMLR, 2023. 5, 16 [27] Hui Mao, Ming Cheung, and James She. Deepart: Learning joint representations of visual arts. In Proceedings of the 25th ACM international conference on Multimedia, pages 1183 1191. ACM, 2017. 6 [28] Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob Mc Grew, Ilya Sutskever, and Mark Chen. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. ar Xiv preprint ar Xiv:2112.10741, 2021. 10 [29] Maxime Oquab, Timothée Darcet, Théo Moutakanni, Huy Vo, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, et al. Dinov2: Learning robust visual features without supervision. ar Xiv preprint ar Xiv:2304.07193, 2023. 6 [30] Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 701 710, 2014. 10 [31] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748 8763. PMLR, 2021. 4, 5, 6 [32] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. Highresolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10684 10695, 2022. 1, 2, 6, 9, 14 [33] Shelly Sheynin, Adam Polyak, Uriel Singer, Yuval Kirstain, Amit Zohar, Oron Ashual, Devi Parikh, and Yaniv Taigman. Emu edit: Precise image editing via recognition and generation tasks. ar Xiv preprint ar Xiv:2311.10089, 2023. 10 [34] Anwaar Ulhaq, Naveed Akhtar, and Ganna Pogrebna. Efficient diffusion models for vision: A survey. ar Xiv preprint ar Xiv:2210.09292, 2022. 14 [35] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. 5 [36] Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. Graph attention networks. In ICLR, 2018. 8, 10 [37] Mengting Wan, Rishabh Misra, Ndapandula Nakashole, and Julian Mc Auley. Fine-grained spoiler detection from large-scale review corpora. In ACL, pages 2605 2610, 2019. 6 [38] Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. A comprehensive survey on graph neural networks. IEEE transactions on neural networks and learning systems, 32(1):4 24, 2020. 5, 10 [39] Ling Yang, Zhilin Huang, Yang Song, Shenda Hong, Guohao Li, Wentao Zhang, Bin Cui, Bernard Ghanem, and Ming-Hsuan Yang. Diffusion-based scene graph to image generation with masked contrastive pre-training. ar Xiv preprint ar Xiv:2211.11138, 2022. 10 [40] Hu Ye, Jun Zhang, Sibo Liu, Xiao Han, and Wei Yang. Ip-adapter: Text compatible image prompt adapter for text-to-image diffusion models. ar Xiv preprint ar Xiv:2308.06721, 2023. 16 [41] Lvmin Zhang, Anyi Rao, and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3836 3847, 2023. 2, 6 [42] Muhan Zhang and Yixin Chen. Link prediction based on graph neural networks. Advances in neural information processing systems, 31, 2018. 10 [43] Jie Zhou, Ganqu Cui, Shengding Hu, Zhengyan Zhang, Cheng Yang, Zhiyuan Liu, Lifeng Wang, Changcheng Li, and Maosong Sun. Graph neural networks: A review of methods and applications. AI open, 1:57 81, 2020. 10 [44] Jing Zhu, Yuhang Zhou, Shengyi Qian, Zhongmou He, Tong Zhao, Neil Shah, and Danai Koutra. Multimodal graph benchmark. ar Xiv preprint ar Xiv:2406.16321, 2024. 10 [45] Haomin Zhuang, Yihua Zhang, and Sijia Liu. A pilot study of query-free adversarial attack against stable diffusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2384 2391, 2023. 14 A.1 Limitations In this work, we focus on node image generation from multimodal attributed graphs, utilizing Stable Diffusion 1.5 as the base model for INSTRUCTG2I. Due to computational constraints, we leave the exploration of larger diffusion models, such as SDXL, for future work. Additionally, we model the graph as homogeneous, not accounting for heterogeneous node and edge types. Considering that different types of nodes and edges convey distinct semantics, future research could investigate how to perform Graph2Image on heterogeneous graphs. A.2 Ethical Considerations While stable diffusion models [32] have demonstrated advanced image generation capabilities, studies highlight several drawbacks, such as the uncontrollable generation of NSFW content [9], vulnerability to adversarial attacks [45], and being computationally intensive and time-consuming [34]. In INSTRUCTG2I, we address these challenges by introducing graph conditions into the image generation process. However, since INSTRUCTG2I employs stable diffusion as the backbone model, it remains susceptible to these limitations. A.3 Classifier-free Guidance In Section 3.4, we discuss controllable generation to balance text and graph guidances (c T and c G) as well as managing multiple graph guidances (cpkq G ). We introduce s T and s G to control the strength of text conditions and graph conditions and have the modified score estimation shown as follows (copied from Eq.(11)): ˆϵθpzt, c G, c T q ϵθpzt, , q s T pϵθpzt, , c T q ϵθpzt, , qq s G pϵθpzt, c G, c T q ϵθpzt, , c T qq. In this section, we will provide mathematical derivation on how these modified score estimations are developed. Noted that INSTRUCTG2I learns Ppz|c G, c T q, the distribution of image latents z conditioned on text information c T and graph information c G, which can be expressed as: Ppz|c G, c T q Ppz, c G, c T q Ppc G, c T q Ppc G|c T , zq Ppc T |zq Ppzq Ppc G, c T q . (13) INSTRUCTG2I learns and estimates the score [20] of the data distribution, which can also be interpreted as the gradient of the log distribution probability. By taking a log on both sides of Eq.(13), we can attain the following equation: logp Ppz|c G, c T qq logp Ppc G|c T , zqq logp Ppc T |zqq logp Ppzqq logp Ppc G, c T qq. (14) After calculating the derivation on both sides of Eq.(14), we can obtain: Blogp Ppz|c G, c T qq Bz Blogp Ppc G|c T , zqq Bz Blogp Ppc T |zqq Bz Blogp Ppzqq This corresponds to our classifier-free guidance equation shown in Eq.(11), where s T controls how the data distribution shifts toward the zone where Ppc T |zq assigns a high likelihood to c T and s G determines how the data distribution leans toward the region where Ppc G|c T , zq assigns a high likelihood to c G. Although there are other ways to derive the modified score estimation function (e.g., switching c T and s G or making it symmetric), we empirically find that our derivation contributes to both advanced performance (since Ppc T |zq is well learned in the base model) and high efficiency (since the denoising operation only needs to be conducted three times rather than four times compared with symmetric setting). If given multiple graph conditions, we utilize spkq G to control the strength for each of them and have the derived score estimation function as follows (copied from Eq.(12)): ˆϵθpzt, c G, c T q ϵθpzt, , q s T pϵθpzt, , c T q ϵθpzt, , qq ÿ spkq G pϵθpzt, cpkq G , c T q ϵθpzt, , c T qq. Table 3: Dataset Statistics Dataset # Node # Edge ART500K 311,288 643,008,344 Amazon 178,890 3,131,949 Goodreads 93,475 637,210 Table 4: Hyper-parameter configuration for model training. Parameter ART500K Amazon Goodreads Optimizer Adam W Adam W Adam W Adam ϵ 1e-8 1e-8 1e-8 Adam pβ1, β2q (0.9, 0.999) (0.9, 0.999) (0.9, 0.999) Weight decay 1e-2 1e-2 1e-2 Batch size per GPU 16 16 16 Gradient Accumulation 4 4 4 Epochs 10 10 30 Resolution 256 256 256 Learning rate 1e-5 1e-5 1e-5 Backbone SD Stable Diffusion 1.5 If multiple graph conditions are given, Eq.(13) then becomes: Ppz|cp1q G , ..., cp Mq G , c T q Ppz, cp1q G , ..., cp Mq G , c T q Ppcp1q G , ..., cp Mq G , c T q Ppcp1q G , ..., cp Mq G |c T , zq Ppc T |zq Ppzq Ppcp1q G , ..., cp Mq G , c T q , (16) where M is the total number of graph conditions. Assume cpkq G are independent from each other, then we can attain: Ppz|cp1q G , ..., cp Mq G , c T q ś k Ppcpkq G |c T , zq Ppc T |zq Ppzq Ppcp1q G , ..., cp Mq G , c T q . (17) Similar to Eq.(15), we can obtain: Blogp Ppz|cp1q G , ..., cp Mq G , c T qq Bz ÿ Blogp Ppcpkq G |c T , zqq Bz Blogp Ppc T |zqq Bz Blogp Ppzqq This corresponds to the classifier-free guidance equation shown in Eq.(12), where spkq G determines how the data distribution leans toward the region where Ppcpkq G |c T , zq assigns a high likelihood to the graph condition cpkq G . A.4 Datasets The statistics of the three datasets can be found in Table 3. Since Amazon and Goodreads both have multiple domains, we select one from each of them considering the graph size: Beauty domain from Amazon and Mystery domain from Goodreads. A.5 Experimental Settings We randomly mask 1,000 nodes as testing nodes from the graph for all three datasets and serve the remaining nodes and edges as the training graph. In implementing INSTRUCTG2I, we initialize the text encoder and U-Net with the pretrained parameters from Stable Diffusion 1.5 1. We use the pretrained CLIP image encoder as our fixed image encoder to extract features from raw images. For Graph-QFormer, we empirically find that initializing it with the CLIP text encoder parameters can improve performance compared with random initialization. 1https://huggingface.co/runwayml/stable-diffusion-v1-5 We use Adam W as the optimizer to train INSTRUCTG2I. The training of all methods including INSTRUCTG2I and baselines on ART500K and Amazon are conducted on two A6000 GPUs, while that on Goodreads is performed on four A40 GPUs. Each image is encoded as four feature vectors with the fixed image encoder following [40] and we insert one cross-encoder layer after every two self-attention layers in Graph-QFormer following [26]. The detailed hyperparameters are in Table 4. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: The contributions and scope is reflected in the abstract and introduction. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: We discuss the limitations of this work in the conclusion section and Appendix section A.1. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: The theoretical derivation of controllable generation is shown in Appendix A.3. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: We provide the code at https://github.com/Peter Griffin Jin/ Instruct G2I. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: We provide the code at https://github.com/Peter Griffin Jin/ Instruct G2I. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: Detailed experimental settings are discussed in Appendix A.5. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: We provide the p-value in section 4.2. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: The compute resources information is given in Appendix A.5. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: We strictly follow the code of ethics. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: This is discussed in Appendix A.2. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: The paper poses no such risks Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: The licenses for the assets are discussed in Appendix A.4. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [Yes] Justification: The license is provided together with the code in https://github.com/ Peter Griffin Jin/Instruct G2I. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.