# revisiting_injective_attacks_on_recommender_systems__d377c102.pdf Revisiting Injective Attacks on Recommender Systems Haoyang LI1, Shimin DI1 , Lei CHEN2 1Computer Science and Engineering, HKUST, Hong Kong, China 2Data Science and Analytics, HKUST(GZ), Guangzhou, China {hlicg,sdiaa,leichen}@cse.ust.hk Recent studies have demonstrated that recommender systems (Rec Sys) are vulnerable to injective attacks. Given a limited fake user budget, attackers can inject fake users with carefully designed behaviors into the open platforms, making Rec Sys recommend a target item to more real users for profits. In this paper, we first revisit existing attackers and reveal that they suffer from the difficulty-agnostic and diversity-deficit issues. Existing attackers concentrate their efforts on difficult users who have low tendencies toward the target item, thus reducing their effectiveness. Moreover, they are incapable of affecting the target Rec Sys to recommend the target item to real users in a diverse manner, because their generated fake user behaviors are dominated by large communities. To alleviate these two issues, we propose a difficulty and diversity aware attacker, namely DADA. We design the difficulty-aware and diversity-aware objectives to enable easy users from various communities to contribute more weights when optimizing attackers. By incorporating these two objectives, the proposed attacker DADA can concentrate on easy users while also affecting a broader range of real users simultaneously, thereby boosting the effectiveness. Extensive experiments on three real-world datasets demonstrate the effectiveness of our proposed attacker. 1 introduction In the era of big data, recommender systems (Rec Sys) [8, 28, 36, 39, 53] are a key component to help users locate items that they may be interested in. Generally, given the user behaviors, such as item purchase history and movie viewing history, Rec Sys learn the latent tendency of users regarding each item, and then recommend the top-N matched items for each user. Over the past decade, various Rec Sys have been proposed to provide services for users. Recent studies [17, 24, 33, 45, 51] demonstrate that Rec Sys are vulnerable to adversarial attacks, e.g., fake user behaviors. It is because the training of Rec Sys relies on user behaviors and attackers can easily inject fake users into openworld web platforms. Specifically, attackers want to promote an item to real users, which can increase the opportunity that users purchase this target item and thus bring economic profits. To achieve such goals, attackers generally inject a limited number of fake users with designed behaviors, i.e., injective attacks. Then, the target Rec Sys may be affected with these injected fake users, and then recommends the target item into the top-N item list of real users. Based on the techniques that are employed to design the behaviors of injected fake users, existing attackers [4, 6, 10, 11, 12, 20, 30, 38, 41, 45, 49, 51, 52] can be classified into three types: heuristicsbased [4, 30, 49], neural network-based [10, 38, 45, 52], and gradient-based attackers [6, 11, 12, 20, 41, 51]. Heuristics-based attackers [4, 30, 49] sample popular items together with the target item as the behaviors of each fake user to increase the co-occurrence between popular items and the target item, i.e., trying to increase the popularity of the target item. Then, Rec Sys can be affected, Corresponding author 36th Conference on Neural Information Processing Systems (Neur IPS 2022). and will recommend the target item to more users. However, such a way does not directly optimize the attack objective, reducing the effectiveness of attackers [41, 52]. Instead, neural network-based attackers [10, 38, 45, 52] utilize neural networks to generate influential fake user behaviors. They maximize the attack objective by optimizing the parameters of neural networks. Gradient-based attackers [6, 11, 12, 20, 41, 51] adopt the continuous relaxation to project user behaviors from the discrete space into the continuous space. This enables attackers to parameterize the fake user behaviors and maximize the attack objective by optimizing them with gradients. Though existing attackers can achieve reasonably good attacking performance, they still suffer from two crucial issues from the local and global views: difficulty-agnostic and diversity-deficit. First, existing attackers do not consider the promotion difficulty of each user regarding the target item. Particularly, users who have a higher (resp. lower) tendency toward the target item are referred to easier (resp. more difficult) users. According to the formulated attack objective of existing attackers, difficult users will contribute more weights to optimize attackers parameters, e.g., neural network parameters for generating fake users. It makes attackers put more efforts to recommend the target item to difficult users. However, the Rec Sys platforms will usually monitor the user registration, i.e., attackers only have a very limited budget for generating fake users. Thus, attackers that focus on difficult users may only achieve small attacking gain, i.e., the target item can only be recommended to a small number of users. Second, the generated fake behaviors of existing attackers cannot influence real users diversely. Specifically, real users with different preferences are dispersed into different communities. Communities with more users may dominate the optimization of attackers. Consequently, the generated fake users cannot affect Rec Sys to promote the target item to various communities, decreasing the effectiveness of attackers. To address difficulty-agnostic and diversity-deficit issues, we propose a Difficulty-Aware and Diversity-Aware attacker, namely DADA, which takes the promotion difficulty and diversity into consideration simultaneously. First, to handle the difficulty-agnostic issue, we propose a novel difficulty-aware objective that more emphasizes easier users than difficult users when optimizing attack parameters. Moreover, we propose a diversity-aware attack objective that allows the generated fake user behaviors to affect the target Rec Sys to recommend the target item to real users diversely, thereby avoiding the diversity-deficit issue. By incorporating both attack objectives, our proposed attacker DADA can concentrate on easy users and generate fake user behaviors that are capable of affecting real users in a diverse manner. Overall, our contributions are summarized as follows: We revisit attackers on Rec Sys and reveal that existing attackers suffer from difficulty-agnostic and diversity-deficit issues. More specifically, the objective designs in existing attackers make them more focus on promoting the target item to those difficult users in the dominant community. However, it is hard to handle difficult users with only a limited number of generated fake users. Besides, the generated fake users tend to fall into the dominant community, which is not helpful to affect real users in other communities. We propose two novel attack objectives (difficulty-aware and diversity-aware objectives) that force the Rec Sys attacker to attack easy users distributed in various communities. Generally, the proposed objectives could enable the easy users and unaffected real users to contribute more weights when optimizing the parameters of the Rec Sys attacker. Experiments on three real-world datasets demonstrate that the proposed attacker DADA achieves superior performance compared with state-of-the-art attackers. Besides, extensive experiments show that the generated fake users indeed focus on easy users in a diverse way. 2 Related Work Formally, let U and I denote users and items, respectively. R 2 {0, 1}|U| |I| denotes the records of user-item interactions. Specifically, each entry R[u][i] denotes whether the user u 2 U interacts with the item i 2 I, and Iu = {i : R[u][i] = 1} denotes the interacted item set of each user u 2 U. 2.1 Recommender System Recently, various Rec Sys [47, 48] have been proposed and deployed in the website applications, which recommend the items for users that match the preference of users. Formally, given users U, items I, and user behaviors R 2 {0, 1}|U| |I|, Rec Sys M learns the latent tendency score Table 1: Summary of existing Rec Sys attackers. X and indicate that the attacker can and cannot satisfy the property, respectively. t is the target item that attackers are interested in. The function f(R[u][i], R[u][t]) = 1 if both R[u][i] = 1 and R[u][t] = 1, and g(x) = 1 1+exp( x/b). b, b1, and b2 are hyperparameters, and a and c are constants. S0 and S are learned tendency before and after attacks, respectively. We discuss them in detail in Sec. 2.2. Type Attacker Attack Objective Oatk( ) Formulation Difficulty Diversity Heuristic-based Popular [29] - Co Vis [49] P u2U f(R[u][i], R[u][t])) a) Neural network-based Poison Rec [38] 1 |Ur| u2Ur I(t, Γu) Copy Attack [10] 1 |Ur| u2Ur I(t, Γu) Trial Attack [45] P i2Γu ln(g(S[u][i] S[u][t])) Gradient-based u2Ur S[u][t] SGLD [20] P u2Ur S[u][t] SRWA [12] P u2Ur S[u][t] TNA [11] P i2Γu g(S[u][i] S[u][t]) PAPU [51] P i2Γu g(S[u][t] S[u][i]) Rev Adv [41] P u2Ur log( exp(S[u][t]) P i2I exp(S[u][i]) ) DADA (Ours) u2Ur t/ 2Γu log(1 b1 S[u][t]) +λ σ(b2 (S[u][t] S0[u][t] c)) X X S = M (R) 2 R|U| |I| between users and items. The parameters can be optimized as follows: Ltra(M (R), R), (1) where Ltra(M (R), R) measures the training loss of Rec Sys model M based on latent tendency S and user behaviors R. Then, based on the learned tendency S, the Rec Sys M recommends the top-N un-interacted items Γu with the highest tendency scores for each user u 2 U, i.e., Γu I \ Iu and |Γu| = N. Generally, current Rec Sys can be classified into three categories, i.e., neighborhoodbased [27, 35], matrix-factorization-based [18, 32], and neural-network-based [15, 50]. Take the fundamental matrix-factorization Rec Sys WRMF [18] as an example, it maps users U and items I into latent representations U 2 R|U| dx and I 2 R|I| dx, respectively. Then, it predicts the tendency S = UIT between users and items. Particularly, WRMF takes = {U, I} as the parameter and reformulates Eq. (1) into arg min ={U,I} u2U,i2I wui(S[u][i] R[u][i])2 + (k Uk2 2). where k k2 is the l2 norm distance, wui is the weight hyperparamter to differentiate the existing and un-existing interactions, and is a regularization hyperparameter. More details of recommender systems can be referred to comprehensive surveys [47, 48]. 2.2 Injective Attacks on Recommender System Given a target item that attackers are interested in, attackers can inject fake users with carefully crafted behaviors into platforms. The attacker s goal is to affect the Rec Sys to promote the target item into the top-N recommendation item list of as many real users as possible. Note that attackers can also demote a target item, i.e., maximizing the number of real users whose top-N recommendation list does not include the target item. In this paper, we focus on promotion, because attacks for demotion can be constructed by a similar technique as promotion [10, 12, 38, 49, 52]. Formally, let U r and U f denote real users and fake users, respectively. Given the real user behaviors Rr 2 {0, 1}|U r| |I|, the target item t, and the target Rec Sys model M , the goal is to generate the fake user data Rf 2 {0, 1}|U f | |I| with the constraints, injection budget |U f| δ and the maximum number k Rf[u]k0 of items that each fake user can interact with, which can maximize the attack objective as follows: Rf Oatk(M (R); U r, t), (2) s.t. = arg min Ltra(M (R), R), |U f| δ, k Rf[u]k0 for any u 2 U f, where Oatk( ) is the attack objective, Ltra( ) is the training loss of Rec Sys M as Eq. (1), and 2 {0, 1}(|U r|+|U f |) |I| is matrix concatenation. Based on this formulation, various attackers against Rec Sys are mainly different from three aspects: (1). The formulation of the attack objective Oatk; (2). Technique used to optimize Eq. (2); (3). Knowledge accessed by attackers. Formulation of Oatk. Hit ratio [10, 38] is a common formulation, i.e., Oatk( ) = 1 |U r| u2U r I(t, Γu), where indicator I(t, Γu) = 1 if t is recommended in the top-N item list of u, i.e., t 2 Γu. Also, several attackers [11, 12, 20, 41, 45, 51, 52] try to maximize the latent tendency between real users and the target item, and propose several attack objective variants. The objective details of these attackers are listed in Tab. 1. For clarification, we specifically list the part of the attack objective that is related to the learned latent tendency S between real users U r and items i 2 I, and omit unrelated parts, such as the regularization of attacker neural network parameters. Note that several attackers [11, 41, 51] try to minimize the attack loss Latk. For generality, we set their attack objective Oatk = Latk and use two notations interchangeably in this paper. Technique. It is a non-trivial task to generate the exact optimal fake user behaviors Rf that can achieve the maximum attack objective due to the exponential candidate search space, i.e., O(|I| |U f |). As shown in Tab. 1, regarding the technique that generates behaviors for injected fake users, there are three types of attackers: heuristics-based, neural network-based, and gradient-based attackers. As introduced in Sec. 1, heuristics-based attackers [4, 30, 49] cannot cover various behavior patterns and do not directly optimize the attack objective, leading to unsatisfied performance [52, 41]. Instead, neural network-based attackers [10, 38, 45, 52] propose neural networks A' parameterized with ' to learn probability distributions of selected items for each fake user. Then, they sample fake user behaviors Rf A'(Rr) accordingly. Formally, the parameter ' can be optimized by maximizing the attack objective as max', ERf A'(Rr)[Oatk(M (R); U r, t)]. Besides, gradient-based attackers A [6, 11, 12, 20, 41, 51] relax the discrete fake user behaviors into continuous Rf 2 R|U f | |I|. Then, they directly parameterize Rf and maximize attack objective as max Rf , Oatk(M ( R); U r, t). After finishing the optimization of Rf, they use a projector function to project the continuous user behaviors Rf into discrete fake behaviors Rf. In this paper, we design a gradient-based attacker to generate fake user behaviors. For generality, we use ' to denote the parameterized Rf. Specifically, existing gradient-based attackers A use gradient ascent approach [34] to optimize ' by maximizing the attack objective Oatk( ) as follows @Oatk(M (R); u, t) where is the learning rate, and Oatk(M (R); u, t) denotes the objective value of user u regarding the target item t. Particularly, when a user u provides more gradients @Oatk(M (R);u,t) @' to optimize parameter ', the attacker will prioritize maximizing the user s objective. For simplicity, we use Oatk(U r, t) and Oatk(u, t) to denote Oatk(M (R); U r, t) and Oatk(M (R); u, t), respectively. Knowledge. Regarding the knowledge accessed by attackers, existing attackers can be classified into two categories. White-box attackers [11, 12, 20] take real user behaviors and the target Rec Sys parameters as inputs to generate fake user behaviors. Unlike white-box attackers, gray-box attackers [10, 38, 41, 45, 51, 52] assume that the target Rec Sys parameters are inaccessible. Instead, the first type of gray-box attackers [10, 38] assume that the top-N item lists of real users recommended by the target Rec Sys are available. Also, the second type of gray-box attackers [41, 45, 51, 52] only access user behaviors and use a surrogate Rec Sys ˆ Mˆ to replace the target Rec Sys M in Eq. (2), which is expected to mimic the target Rec Sys. In summary, the second type of gray-box attackers are more realistic than white-box and the first type of gray-box. The reason is that the target Rec Sys parameters and the recommendation lists of real users are difficult to obtain. Instead, due to the openness of web platforms, it is easy to collect real user behaviors [41, 51], such as movie viewing history. More discussions on user behavior collection can be referred to Appx. B.5. Thus, we focus on the second type of gray-box attacks on Rec Sys in this paper. 3 DADA Attacker In this section, we introduce our attacker DADA. We first propose two attack objectives that consider the promotion difficulty and the diversity of users, respectively. Then, we formulate a novel difficulty and diversity-aware attack objective that enables our attacker DADA to simultaneously focus on easy users and affect real users diversely, and propose an efficient greedy-based algorithm to optimize it. (a) Average gradient on each user group. (b) HR@50 on each user group. (c) Fake user distribution on each community. (d) HR@50 on each community. Figure 1: Difficulty-agnostic and diversity-deficit analysis. 3.1 Difficulty-aware Attack Objective 3.1.1 Motivational Observations of Difficulty-agnostic As discussed in Sec. 1, given a target item and a set of real users, existing Rec Sys attackers more focus on promoting the target item to difficult users when designing fake user behaviors. We first demonstrate several motivational observations in existing works. Specifically, we employ the representative and state-of-the-art gradient-based attackers, including Rev Adv [41], TNA [11], and PAPU [51], to inject fake users into the ML-100K dataset [14], where fake user budget δ = 0.01|U r|. Specifically, we use the representative Rec Sys WRMF [18] as the target Rec Sys M for evaluation. We first train WRMF on ML-100K without attacks, and output the latent tendency S between each user u 2 U r and each item i 2 I. We then randomly select an item t 2 I as the target item. For simplicity and clarification, we separate real users into three groups as an example. Specifically, if the latent tendency of a user u toward the target item S[u][t] is at the top of {33.3%, 33.3% 66.7%, 66.7% 100%} largest among all items I, it will be regarded as {easy, medium, difficult} users, respectively. Then, we can obtain three groups {U e, U m, U d}. For clarification, we show the result of Rev Adv in Fig. 1 and similar results of TNA and PAPU in Appx. A.1.1. For each group U g 2 {U e, U m, U d}, we show its average gradient for optimizing the attacker parameter ' in Fig. 1 (a), i.e., u2Ug kr'Latk(u,t)k1 |U g| (|U f | |I|) . We can easily observe that the gradient value is directly proportional to the degree of difficulty. In other words, the more difficult it is, the greater gradient it is. It indicates that Rev Adv puts more efforts to promote the target item to difficult and medium users than easy users. Then, we investigate the hit ratio of 50 (HR@50) performance on three different user groups. As illustrated in Fig. 1 (b), despite that Rev Adv puts more effort on medium and difficult users, the HR@50 of them is only marginally improved. Moreover, putting more effort on medium and difficult users leads that the HR@50 of easy users increases unstably. Therefore, with a limited budget for fake users, it is preferable to concentrate on easy users, as it is easier to promote the target item for them. 3.1.2 Insights on Difficulty-agnostic Issue Motivated by the observations, we plan to design a new difficulty-aware attack objective, which enables attackers to focus on easy users primarily and achieve more recommendation gain. The basic idea is to let easier users outweigh difficult users when optimizing the attacker s parameter '. As introduced in Eq. (3), the gradients for optimizing attack parameter ' that each user provides are related to the attack objective Oatk. Particularly, due to the chain rule, the gradients @Oatk(u,t) each user u 2 U r can be formulated as @Oatk(u,t) @' = @Oatk(u,t) @S[u][t] @S[u][t] @' , where @Oatk(u,t) @S[u][t] only relates to the attacker objective design. @S[u][t] @' only relies on the attacker parameters, which is fixed and unrelated to attack objective Oatk( ). Thus, we can increase the @Oatk(u,t) @S[u][t] of easier user to let them more determine the optimization direction of attacker parameters. Besides, if a user has received the target item in the top-N recommendation list, it indicates that the Rec Sys has promoted the target item to this user successfully. Thus, this user should not provide any gradient, because increasing its tendency toward the target item will not have any improvement on the recommendation gain. Based on the above analysis, the difficulty-aware attack objective should satisfy two difficulty properties: (1) given two users u and v, if the tendency score S[v][t] < S[u][t] and t /2 Γu, @S[v][t] < @Oatk(u,t) @S[u][t] . (2) if the target item t 2 Γu, @Oatk(u,t) @S[u][t] = 0. As shown in Tab. 1, the attack objectives of existing attackers do not satisfy both properties simultaneously. In this paper, based on the two properties, we formulate a difficulty-aware attack objective. Formally, given the tendency ˆS learned by the surrogate Rec Sys ˆ Mˆ , the target item t, the difficulty-aware attack objective regarding each real user u 2 U r is formulated as follows: Dict( ˆ Mˆ (R); u, t) = log(1 b1 ˆS[u][t]), t /2 Γu where b1 is a positive hyperparameter. Furthermore, we use the following Proposition 1 to show that Eq. (4) enables easier users to more dominate the the attack parameter optimization than difficult users. Due to space limit, we put the proof in Appx. A.2.1. Proposition 1. Our difficulty-aware attack objective satisfies difficulty property 1 and property 2. 3.2 Diversity-aware Attack Objective 3.2.1 Motivational Observations of Diversity-deficit As introduced in Sec. 1, existing attackers have the diversity-deficit issue that the generated fake behaviors are dominated by large communities, where users in each community have similar preferences. Specifically, as introduced in Eq. (3), each real user u 2 U r will provide gradients to optimize the attackers parameters '. Since similar real users have similar preferences on items, real users can form several communities {Cj}nc j=1, where each community Cj has nj real users. Thus, from the community perspective, the optimization process of the attacker parameter ' via maximizing the attack objective Oatk( ) can be reformulated as follows: ' = '+ Pnc @Oatk(M (R);u,t) @' . Intuitively, users in the same community who are similar will give the gradients in similar directions. Thus, there is a potential risk that the large communities Ck will dominate the optimization direction of attacker parameter ' via P @Oatk(M_ (R);u,t) @' . Consequently, this community will dominate the generated fake user behaviors. The generated fake user behaviors can only manipulate the target Rec Sys to promote the target item to the dominated communities. In another words, the target Rec Sys is incapable of promoting the target item to more users in other communities diversely. Similar to Sec. 3.1.1, we demonstrate several motivational observations. We use state-of-the-art gradient-based attackers Rev Adv [41], TNA [11], and PAPU [51], and neural network-based attacker Trial Attack [45], to generate 0.05|U r| fake users and inject them into the ML-100K dataset [14]. Then, we train the target Rec Sys WRMF and study the attacking results. For simplicity and clarification, we put the result of Rev Adv in Fig. 1 and similar results of TNA, PAPU, and Trial Attack in Appx. A.1.2. Following [41], we extract the representations of real and fake users learned by WRMF, and use PCA [1] to visualize them. Based on the result of PCA, we separate real users into three communities by KMeans [19]. As shown in Fig. 1 (c), the fake users (blue points) tend to form a cluster that is distributed in the community 1 (yellow points). It indicates that community 1 dominates the generated fake users. It is because users in the same community will generate gradients with similar directions. Thus, large communities with more users are more likely to dominate the optimization directions of the attacker parameter. Consequently, Rec Sys can only be affected by the generated fake user behaviors to recommend the target item to more users in the community 1. To further verify this, we list the hit ratio 50 (HR@50) of each community before and after injecting fake users in Fig. 1 (d). It shows that the generated fake users improve the HR@50 of the community 1 significantly, while maintaining similar HR@50 in the other two communities. To enable Rev Adv to affect users diversely, we add our formulated diversity-aware attack objective (see Sec. 3.2.2) into Rev Adv s objective. As shown in Fig. 1 (d), Rev Adv+Diverse improves HR@50 of community 2 and 3. Particularly, HR@50 on community 1 increases as well. As described in Sec. 3.1.1, Rev Adv more focuses on difficult users, thus the generated fake user behaviors may not be able to affect the target Rec Sys to recommend the target item to other less difficult users in community 1. Fortunately, the diverse-aware objective allows Rev Adv to put more efforts on these users, thereby improving the performance of community 1. 3.2.2 Insights on Diversity-deficit Issue Motivated by these observations, we propose a diversity-aware attack objective that enables fake user behaviors to affect real users diversely. The basic idea is that if the real users are less affected by fake user behavior, the diversity-aware objective will concentrate on these real users in the subsequent gradient computation process. Then, the attacker will be optimized to generate fake user behaviors that can affect these users. In such a way, users in different communities can be affected diversely. Specifically, the influence of fake user behaviors on each user regarding the target item can be measured by the difference between tendencies before and after attacks. Formally, let S0 and S denote the tendencies learned by Rec Sys M before and after attacks, the influence of fake user behaviors Rf on user u regarding the target item t is computed by 4S[u][t] = S[u][t] S0[u][t]. To enable attackers to affect more real users, the attack objective should focus on users with a smaller 4S[u][t] when optimizing the attacker parameter '. Specifically, according to the chain rule and @4S[u][t] @S[u][t] = 1, the gradients @Oatk(u,t) @' regarding the attacker parameter ' in Eq. (3) of each user u 2 U r can be formulated as @Oatk(u,t) @' = @Oatk(u,t) @4S[u][t] @S[u][t] @' . Thus, we can increase the @Oatk(u,t) @4S[u][t] of real users with smaller influence 4S[u][t] to let them more determine the attacker parameter optimization direction. Similar to Sec. 3.1.2, the diversity-aware attacker objective should satisfy two diversity properties: (1) given two users u and v, if 4S[v][t] > 4S[u][t], and t /2 Γu, @Oatk(v,t) @4S[v][t] < @Oatk(u,t) @4S[u][t] ; (2) if the target item t has been recommended to a user u, the user u should have no effect on the attackers parameter optimization, i.e., if t 2 Γu, @Oatk(u,t) @4S[u][t] = 0. Obviously, existing attackers do not satisfy both properties because they do not consider the difference of tendency between before and after attacks. Based on the two properties, we formulate a diversityaware attack objective. Formally, given the target item t, and the real user u 2 U r, let ˆS0 and ˆS denote the tendencies learned by the surrogate Rec Sys ˆ Mˆ before and after injecting fake user behaviors Rf, the diversity-aware objective on each user u 2 U r regarding the target item t is computed as: Div(ˆS[u][t]; u, t, ˆS0[u][t]) = σ(b2 (ˆS[u][t] ˆS0[u][t] c)), t /2 Γu 0, t 2 Γu , (5) where c = minv2U r (ˆS[v][t] ˆS0[v][t]) is a constant for normalization, and b2 is a positive hyperparameter to scale the tendency difference. σ(x) = 1 1+exp( x) is the Sigmoid function. Additionally, we use Proposition 2 to show that Eq. (5) enables users with a smaller 4ˆS[u][t] to contribute more weights to the attack parameter optimization. Due to space limit, we put the proof in Appx. A.2.2. Proposition 2. Our diversity-aware attack objective satisfies diversity property 1 and property 2. 3.3 Difficulty and Diversity Aware Attacker 3.3.1 Problem Formulation Based on two measurements proposed in Eq. (4) and Eq. (5), we formally formulate the Difficulty Aware and Diversity-Aware (DADA) Rec Sys attacking problem as follows. Definition 1 (DADA Problem). Given the real user behaviors Rr 2 {0, 1}|U r| |I|, the target item t, and the surrogate Rec Sys ˆ Mˆ , the goal is to generate the fake user behaviors Rf 2 {0, 1}|U f | |I| with the constraints, injection budget |U f| δ and the maximum number k Rf[u]k0 of items that each fake user can interact with, which can maximize the attack objective as follows: Dict( ˆ Mˆ (R); u, t) + λ Div(ˆS[u][t]; u, t, ˆS0[u][t]) (6) s.t. ˆ = arg min Ltra( ˆ Mˆ (R), R), |U f| δ, k Rf[u]k0 for any u 2 U f, where λ is a trade-off hyperparamter, R 2 {0, 1}(|U r|+|U f |) |I| concatenates the real and fake user representations, ˆS and ˆS0 are the learned latent tendencies of the surrogate Rec Sys ˆ Mˆ on R and Rr, respectively. Ltra( ) is the training loss of ˆ Mˆ as in Eq. (1). 3.3.2 Greedy Algorithm Due to the exponential search space O(|I| |U f |), it is nontrivial to find the exact optimal fake user behaviors for the difficulty-aware and diversity-aware optimization problem in Eq (6). Thus, we propose a gradient-based greedy algorithm to maximize the objective in Alg. 1 in Appx. A.3. With the help of gradients, our algorithm can give a good direction to find effective fake user behaviors. Specifically, we first initialize the fake user behaviors Rf 2 {0, 1}|U f | |I| (line 1). In each epoch l, we first train the surrogate model ˆ Mˆ on R, and then compute the attack objective Oatk( ) in Eq. (6) (line 4-5). Second, we relax the discrete fake user behaviors Rf into a continuous one Rf 2 R|U f | |I|, and take it as our attacker s parameter ' (line 6). Then, we compute the gradients r Rf Oatk( ) according to [41] (line 7). Finally, we update the Rf and utilize a projector to map it into discrete behaviors Rf, where each fake user can select items at most (line 8). Formally, given the continuous user behaviors Rf, the maximum number of items that each fake user u 2 U f can interact with, and the pre-defined threshold th, the projector is formulated as follows: Proj( Rf; , th)[u][i] = 0, Rf[u][i] < th _ i /2 Pu 1, Rf[u][i] > th i 2 Pu where Pu is toplargest item set of user u according to the value in Rf[u]. After finishing L epochs, we output Rf for injection. We analyze the time complexity of Alg. 1 in Appx. A.4. 4 Experiments 4.1 Experiment Setting Datasets. We evaluate our attacker on widely used three real-world datasets: Movie Lens-100K (ML-100K) [14], Movie Lens-1M (ML-1M) [14], and Gowalla [5]. We put the descriptions of three datasets in Appx. B.1. Attacker Baselines. To verify the effectiveness of our proposed attacker, we compare our proposed attacker DADA with the three types of attackers in Tab. 1. Specifically, heuristic-based attackers include Random [12, 51] and Popular [29, 51], neural network-based attackers include Trial Attack [45], and gradient-based attackers include Co Vis [49], PGA [20], SGLD [20], SRWA [12], TNA [11], Rev Adv [41], and PAPU [51]. Particularly, we do not compare Poison Rec [38] and Copy Attack [10] in Tab. 1, because they need to access the predictions of the target Rec Sys for real users. Also, we compare two variants: DADA-DICT and DADA-DIV, which only optimize the difficulty and diversity-aware objective in Eq. (4) and Eq. (5), respectively. Targeted Recommender Systems. Following [41, 51], we choose three representative target Rec Sys to evaluate the attacking performance of Rec Sys attackers, i.e., matrix factorization-based WRMF [18], neural network-based NCF [16], and graph neural network-based Light GCN [15]. Target Items and Evaluation Metric. Following [41, 51], we uniformly sample 5 items as a target item set It and measure the hit ratio Hit@N of It on the real users U r. If one of these target items appears in the top-N recommended item list of a user u 2 U r, the target item set is considered as a hit. Besides, we use the Normalized Discounted Cumulative Gain (NDCG) to measure the rank quality of the target items in the recommendation list. Specifically, we set the relevance score of target items and the other items for all real users as 1 and 0, respectively. Then, a higher NDCG value indicates better attack performance, where target items are ranked higher in the recommendation list. The reported results are averaged over five different runs. Hyperparamter Setting. Due to datasets with different sizes, we set the fake user budget δ = β1 |U r| and the maximum number of selected items = β2 |I|, where β1 and β2 are the parameters to control δ and . For default, we set β1 = 0.01 for all datasets. For unnoticeable, we set the default to be similar to the average number of interacted items of real users as shown in Tab. 4, i.e., β2 is {0.05, 0.05, 0.002} for ML-100K, ML-1M, and Gowalla, respectively. The details about hyper-parameter settings are listed in Appx. B.2. 4.2 Experiment Results We report the HR performance of attackers on ML-100K and ML-1M in Tab. 2 and Tab. 3. To better show the improvement, we add the improvement ratio of HR@10 for each attacker compared with the result in clean data. Due to space limit, we report similar results of Gowalla in Appx. B.3. We report the NDCG results on three datasets in Appx. B.4. Under the same target Rec Sys, the bold and underlined numbers indicate the first and second best performance, respectively. Baseline Comparison. As shown in Tab. 2 and Tab. 3, our attacker DADA outperforms all baselines under different target Rec Sys on both ML-100K and ML-1M datasets. Specifically, heuristics-based Table 2: Evaluation on ML-100K dataset under fake user budget δ = 0.01|U r|. Attacker WRMF [18] NCF [16] Light GCN [15] HR@10 HR@50 HR@100 HR@10 HR@50 HR@100 HR@10 HR@50 HR@100 Clean 0.0901 0.3606 0.5677 0.1094 0.4799 0.7462 0.0977 0.3813 0.7941 Random [51, 12] 0.0907 (+0.67%) 0.3701 0.5536 0.1134 (+3.66%) 0.5031 0.7511 0.0987 (+1.02%) 0.3887 0.7824 Popular [29] 0.0912 (+1.22%) 0.3743 0.5695 0.1145 (+4.66%) 0.5221 0.7574 0.0998 (+2.15%) 0.3918 0.7973 Co Vis [49] 0.0965 (+7.10%) 0.3722 0.5778 0.1139 (+4.11%) 0.5585 0.7798 0.0985 (+0.82%) 0.3960 0.7946 Trial Attack [45] 0.1002 (+11.21%) 0.3902 0.5911 0.1203 (+9.96%) 0.5998 0.7868 0.0998 (+2.15%) 0.3992 0.7983 SRWA [12] 0.0906 (+0.55%) 0.3721 0.5732 0.1172 (+7.13%) 0.5124 0.7521 0.0991 (+1.43%) 0.3927 0.7981 PGA [20] 0.0923 (+2.44%) 0.3752 0.5769 0.1211 (+10.69%) 0.5483 0.7668 0.0987 (+1.02%) 0.3971 0.7946 SGLD [20] 0.0923 (+2.44%) 0.3721 0.5716 0.1187 (+8.50%) 0.5530 0.7675 0.0966 (-1.13%) 0.3824 0.7952 TNA [11] 0.0974 (+8.10%) 0.3801 0.5843 0.1199 (+9.60%) 0.5812 0.7826 0.0924 (-5.42%) 0.3928 0.7962 Rev Adv [41] 0.1017 (+12.87%) 0.4167 0.6285 0.1229 (+12.34%) 0.6128 0.7885 0.1008 (+3.17%) 0.4181 0.8026 PAPU [51] 0.1039 (+15.32%) 0.4231 0.6361 0.1405 (+28.43%) 0.6334 0.7983 0.1012 (+3.58%) 0.4191 0.7998 DADA-DICT 0.1145 (+27.08%) 0.4380 0.6471 0.1418 (+29.62%) 0.6387 0.8130 0.1019 (+4.30%) 0.4212 0.8061 DADA-DIV 0.1078 (+19.64%) 0.4218 0.6411 0.1375 (+25.69%) 0.6292 0.8017 0.1011 (+3.48%) 0.4196 0.8033 DADA 0.1220 (+35.41%) 0.4453 0.6585 0.1513 (+38.30%) 0.6623 0.8225 0.1040 (+6.45%) 0.4286 0.8127 Table 3: Evaluation on ML-1M dataset under fake user budget δ = 0.01|U r|. Attacker WRMF [18] NCF [16] Light GCN [15] HR@10 HR@50 HR@100 HR@10 HR@50 HR@100 HR@10 HR@50 HR@100 Clean 0.0195 0.0970 0.1912 0.0158 0.0925 0.1965 0.0257 0.1008 0.1572 Random [51, 12] 0.0196 (+0.51%) 0.1002 0.2030 0.0174 (+10.13%) 0.0948 0.1983 0.0264 (+2.72%) 0.1028 0.1618 Popular [29] 0.0196 (+0.51%) 0.1007 0.2028 0.0202 (+27.85%) 0.0976 0.2008 0.0268 (+4.28%) 0.1035 0.1692 Co Vis [49] 0.0212 (+8.72%) 0.1061 0.2148 0.0198 (+25.32%) 0.1073 0.2180 0.0279 (+8.56%) 0.1057 0.1793 Trial Attack [45] 0.0218 (+11.79%) 0.1084 0.2203 0.0213 (+34.81%) 0.1088 0.2159 0.0295 (+14.79%) 0.1115 0.1864 SRWA [12] 0.0202 (+3.59%) 0.1017 0.2060 0.0194 (+22.78%) 0.1021 0.2061 0.0272 (+5.84%) 0.1031 0.1711 PGA [20] 0.0206 (+5.64%) 0.1021 0.2072 0.0201 (+27.21%) 0.1017 0.2048 0.0276 (+7.39%) 0.1046 0.1715 SGLD [20] 0.0209 (+7.18%) 0.1020 0.2025 0.0192 (+21.52%) 0.0992 0.2008 0.0267 (+3.89%) 0.1029 0.1702 TNA [11] 0.0221 (+13.33%) 0.1068 0.2182 0.0195 (+23.42%) 0.1049 0.2065 0.0282 (+9.73%) 0.1074 0.1854 Rev Adv [41] 0.0227 (+16.41%) 0.1168 0.2412 0.0224 (+41.77%) 0.1109 0.2197 0.0318 (+23.74%) 0.1206 0.2037 PAPU [51] 0.0246 (+26.15%) 0.1126 0.2462 0.0221 (+39.87%) 0.1125 0.2173 0.0334 (+29.96%) 0.1257 0.2077 DADA-DICT 0.0253 (+29.74%) 0.1217 0.2581 0.0230 (+45.57%) 0.1290 0.2219 0.0343 (+33.46%) 0.1312 0.2184 DADA-DIV 0.0236 (+21.03%) 0.1175 0.2460 0.0227 (+43.67%) 0.1146 0.2201 0.0331 (+28.79%) 0.1214 0.2049 DADA 0.0286 (+46.67%) 0.1407 0.2746 0.0233 (+47.47%) 0.1315 0.2313 0.0362 (+40.86%) 0.1407 0.2308 approaches, such as Random, Popular, and Co Vis, achieve unsatisfied results. Because they only consider to increase the co-occurrence between several items and the target items, which does not relate to the attack objective directly and cannot cover various behavior patterns [41, 51, 52]. Instead, neural network-based and gradient-based attackers optimize the attack objective directly. Compared with Trial Attack, Rev Adv and PAPU, several gradient-based attackers, including SRWA, PGA, SGLD, and TNA, achieve inferior performance. It is because when computing the gradients to design fake user behaviors, these approaches approximate the gradients in a biased way instead of exact computing gradients [41]. Our attacker DADA outperforms the above baselines significantly, because it enables the fake user behaviors to affect real users in a diverse manner. Ablation Study. As shown in Tab. 2 and Tab. 3, DADA-DICT achieves the second best results, outperforming all baselines. It is because the difficulty-aware attack objective allows DADA-DICT to concentrate on easy users with high tendencies toward the target items. Moreover, DADA-DIV achieves satisfied performance. It outperforms current state-of-the-art attackers, including Rev Adv and PAPU, in most cases. It is because the diversity-aware attack objective allows the generated fake user behaviors to affect real users more diversely, avoiding large communities dominating the generated fake users. Nevertheless, DADA-DIV does not outperform DADA-DICT, because DADADIV cannot differentiate easy users from all real users and put more efforts on them. Furthermore, DADA that considers both user difficulty and diversity can perform the best. In such a way, it can potentially affect easy users from different communities, thereby boosting attacker s effectiveness. 4.3 Parameter Sensitivity In this subsection, we evaluate DADA under different parameters. As shown in Tab 2 and Tab. 3, the results of DADA and baselines on different target Rec Sys and hit ratios have similar trends. For clarification, we use WRMF and HR@50 for evaluation on ML-100K and ML-1M datasets. Fake User Budget δ. We study the effect of fake user budget δ = β1|U r| by varying β1 2 {0.01, 0.03, 0.05, 0.07, 0.09}. For clarification, we only compare the effective baselines, including Trial Attack, TNA, Rev Adv, and PAPU. As shown in Fig. 2 (a) and (b), HR@50 of all attackers increases as the fake user budget increases on both ML-100K and ML-1M datasets. Furthermore, (a) δ on ML-100K. (b) δ on ML-1M. (c) on ML-100K. (d) on ML-1M. Figure 2: Evaluation on the fake user budget and maximum selected items. DADA outperforms all baselines under different attack budgets, i.e., the red solid line is above other lines, demonstrating its effectiveness. Particularly, as the fake user budget increases, our attacker and baselines improve HR@50 on ML-100K more than that on ML-1M. The reason is that ML-100K only has 1682 items, while ML-1M has 3706 items. As a result, it is more difficult to promote target items into the top-50 recommendation list of each user in ML-1M than ML-100K. Maximum Selected Items . We study the effect of different maximum number of selected items = β2|I| by varying β2 2 {0.01, 0.03, 0.05, 0.07, 0.09}. As shown in Fig. 2 (c) and (d), as the maximum number of items increases, the performance of attackers first increases and then decreases on both datasets. It indicates that selecting suitable items can increase the attacker s power. But more items will reduce the effectiveness of attackers, since the valuable items are limited, and extra items will introduce noise [12, 29, 51]. Furthermore, compared with these powerful baselines, DADA can achieve stable and the best performance under different maximum number of selected items. 4.4 More Insights via Empirical Study Due to space limit, we include more experiments in Appx. B for more insights. In Appx. B.5, to better mimic the real-world applications, we demonstrate the generality and practicability of our attacker on partial and noisy user behaviors. In Appx. B.6, we evaluate our attacker under different item popularity. In Appx. B.7, we more comprehensively reveal the effectiveness of the difficultyaware objective and quantitatively measure the diversity brought by the diversity-aware objective. In Appx. B.8, we evaluate our attacker under more parameter settings, including extremely limited fake user budget and λ in Eq (6). In Appx. B.9, we analyze the hit ratio improvement under different target Rec Sys. In Appx. B.10, we provide visualizations on gradients provided by different users and fake user distributions. Finally, in Appx. B.11, we list several future directions based on our observations. 5 Conclusion In this paper, we revisit the problem of injective attacks against recommender systems and reveal two issues of existing attackers: difficulty-agnostic and diversity-deficit issues. To address these two issues, we propose DADA, a difficulty and diversity aware attacker. Within a limited fake user budget, DADA enables the generated fake user behaviors to concentrate on easy users who have high tendencies toward the target item, and affect real users in a more diverse manner. Extensive experiments on three real-world datasets demonstrate the effectiveness of our proposed attacker. In this paper, we mainly reveal the weakness of existing recommender systems and discuss how to attack them. For the sake of society, we discuss how to defense and resist such attacks in Appx. B.12. 6 Acknowledgement Lei Chen s work is partially supported by National Key Research and Development Program of China Grant No. 2022YFE0200500 and 2018AAA0101100, the Hong Kong RGC GRF Project 16213620, CRF Project C6030-18G, C1031-18G, C5026-18G, CRF C2004-21GF, AOE Project Ao E/E603/18, RIF Project R6020-19, Theme-based project TRS T41-603/20R, China NSFC No. 61729201, Guangdong Basic and Applied Basic Research Foundation 2019B151530001, Hong Kong ITC ITF grants ITS/044/18FX and ITS/470/18FX, Microsoft Research Asia Collaborative Research Grant, HKUST-NAVER/LINE AI Lab, Didi-HKUST joint research lab, HKUST-Webank joint research lab grants and HKUST Global Strategic Partnership Fund (2021 SJTU-HKUST). Shimin Di s work is supported by the JC STEM Lab of Data Science Foundations funded by The Hong Kong Jockey Club Charities Trust. [1] Hervé Abdi and Lynne J Williams. Principal component analysis. Wiley interdisciplinary reviews: computational statistics, 2(4):433 459, 2010. [2] Ofer Arazy, Nanda Kumar, and Bracha Shapira. Improving social recommender systems. IT professional, 11(4):38 44, 2009. [3] Atilim Gunes Baydin et al. Automatic differentiation in machine learning: a survey. JMLR, 18:1 43, 2018. [4] Robin Burke et al. Limited knowledge shilling attacks in collaborative filtering systems. In IJCAI), pages 17 24, 2005. [5] Eunjoon Cho et al. Friendship and mobility: user movement in location-based social networks. In SIGKDD, pages 1082 1090, 2011. [6] Konstantina Christakopoulou and Arindam Banerjee. Adversarial attacks on an oblivious recommender. In Rec Sys, pages 322 330, 2019. [7] Anahita Davoudi and Mainak Chatterjee. Detection of profile injection attacks in social recommender systems using outlier analysis. In 2017 IEEE International Conference on Big Data (Big Data), pages 2714 2719. IEEE, 2017. [8] Yashar Deldjoo, Tommaso Di Noia, and Felice Antonio Merra. A survey on adversarial recommender systems: from attack/defense strategies to generative adversarial networks. ACM Computing Surveys (CSUR), 54(2):1 38, 2021. [9] Shimin Di, Yanyan Shen, and Lei Chen. Relation extraction via domain-aware transfer learning. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & Data Mining, pages 1348 1357, 2019. [10] Wenqi Fan et al. Attacking black-box recommendations via copying cross-domain user profiles. In ICDE, pages 1583 1594. IEEE, 2021. [11] Minghong Fang, Neil Zhenqiang Gong, and Jia Liu. Influence function based data poisoning attacks to top-n recommender systems. In WWW, pages 3019 3025, 2020. [12] Minghong Fang, Guolei Yang, Neil Zhenqiang Gong, and Jia Liu. Poisoning attacks to graph- based recommender systems. In ACSAC, pages 381 392, 2018. [13] Matthias Fey, Jan E Lenssen, Frank Weichert, and Jure Leskovec. Gnnautoscale: Scalable and expressive graph neural networks via historical embeddings. In International Conference on Machine Learning, pages 3294 3304. PMLR, 2021. [14] F Maxwell Harper et al. The movielens datasets: History and context. TIIS, 5(4):1 19, 2015. [15] Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yongdong Zhang, and Meng Wang. Lightgcn: Simplifying and powering graph convolution network for recommendation. In SIGIR, pages 639 648, 2020. [16] Xiangnan He et al. Neural collaborative filtering. In WWW, pages 173 182, 2017. [17] Hai Huang, Jiaming Mu, Neil Zhenqiang Gong, Qi Li, Bin Liu, and Mingwei Xu. Data poisoning attacks to deep learning based recommender systems. ar Xiv preprint ar Xiv:2101.02644, 2021. [18] Yehuda Koren, Robert Bell, and Chris Volinsky. Matrix factorization techniques for recom- mender systems. Computer, 42(8):30 37, 2009. [19] K Krishna et al. Genetic k-means algorithm. Cybernetics, 29(3):433 439, 1999. [20] Bo Li et al. Data poisoning attacks on factorization-based collaborative filtering. Neur IPS, 29:1885 1893, 2016. [21] Haoyang Li and Lei Chen. Cache-based gnn system for dynamic graphs. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pages 937 946, 2021. [22] Haoyang Li, Shimin Di, Zijian Li, Lei Chen, and Jiannong Cao. Black-box adversarial attack and defense on graph neural networks. In ICDE, pages 1017 1030. IEEE, 2022. [23] Haoyang Li, Xueling Lin, and Lei Chen. Fine-grained entity typing via label noise reduc- tion and data augmentation. In International Conference on Database Systems for Advanced Applications, pages 356 374. Springer, 2021. [24] Chen Lin, Si Chen, Hui Li, Yanghua Xiao, Lianyun Li, and Qian Yang. Attacking recommender systems with augmented user profiles. In Proceedings of the 29th ACM international conference on information & knowledge management, pages 855 864, 2020. [25] Xueling Lin, Lei Chen, and Chaorui Zhang. Tenet: Joint entity and relation linking with coherence relaxation. In Proceedings of the 2021 International Conference on Management of Data, pages 1142 1155, 2021. [26] Xueling Lin, Haoyang Li, Hao Xin, Zijian Li, and Lei Chen. Kbpearl: a knowledge base population system supported by joint entity and relation linking. Proceedings of the VLDB Endowment, 13(7):1035 1049, 2020. [27] Greg Linden et al. Amazon. com recommendations: Item-to-item collaborative filtering. IEEE Internet computing, 7(1):76 80, 2003. [28] Jiaqi Ma, Zhe Zhao, Xinyang Yi, Ji Yang, Minmin Chen, Jiaxi Tang, Lichan Hong, and Ed H Chi. Off-policy learning in two-stage recommender systems. In Proceedings of The Web Conference 2020, pages 463 473, 2020. [29] Bamshad Mobasher et al. Toward trustworthy recommender systems: An analysis of attack models and algorithm robustness. TOIT, 7(4):23 es, 2007. [30] Michael P O Mahony, Neil J Hurley, and Guénolé CM Silvestre. Recommender systems: Attack types and strategies. In AAAI, pages 334 339, 2005. [31] Steffen Rendle et al. Bpr: Bayesian personalized ranking from implicit feedback. ar Xiv preprint ar Xiv:1205.2618, 2012. [32] Steffen Rendle, Walid Krichene, Li Zhang, and John Anderson. Neural collaborative filtering vs. matrix factorization revisited. In Rec Sys, pages 240 248, 2020. [33] Dazhong Rong, Shuai Ye, Ruoyan Zhao, Hon Ning Yuen, Jianhai Chen, and Qinming He. Fedrecattack: Model poisoning attack to federated recommendation. ar Xiv preprint ar Xiv:2204.01499, 2022. [34] Sebastian Ruder. An overview of gradient descent optimization algorithms. ar Xiv preprint ar Xiv:1609.04747, 2016. [35] Badrul Sarwar, George Karypis, Joseph Konstan, and John Riedl. Item-based collaborative filtering recommendation algorithms. In Proceedings of the 10th international conference on World Wide Web, pages 285 295, 2001. [36] Lalita Sharma and Anju Gera. A survey of recommendation system: Research challenges. IJETT, 4(5):1989 1992, 2013. [37] DI Shimin, YAO Quanming, Yongqi ZHANG, and CHEN Lei. Efficient relation-aware scoring function search for knowledge graph embedding. In 2021 IEEE 37th International Conference on Data Engineering (ICDE), pages 1104 1115. IEEE, 2021. [38] Junshuai Song et al. Poisonrec: an adaptive data poisoning framework for attacking black-box recommender systems. In ICDE, pages 157 168. IEEE, 2020. [39] Jiaxi Tang, Francois Belletti, Sagar Jain, Minmin Chen, Alex Beutel, Can Xu, and Ed H. Chi. Towards neural mixture recommender for long range dependent user sequences. In The World Wide Web Conference, pages 1782 1793, 2019. [40] Jiaxi Tang and Ke Wang. Personalized top-n sequential recommendation via convolutional sequence embedding. In WSDM, pages 565 573, 2018. [41] Jiaxi Tang, Hongyi Wen, and Ke Wang. Revisiting adversarially learned injection attacks against recommender systems. In Rec Sys, pages 318 327, 2020. [42] Xin Wang, Ziwei Zhang, and Wenwu Zhu. Automated graph machine learning: Approaches, libraries and directions. ar Xiv preprint ar Xiv:2201.01288, 2022. [43] Zhili Wang, Shimin Di, and Lei Chen. Autogel: An automated graph neural network with explicit link information. Advances in Neural Information Processing Systems, 34:24509 24522, 2021. [44] Chenwang Wu, Defu Lian, et al. Fight fire with fire: Towards robust recommender systems via adversarial poisoning training. In SIGIR, pages 1074 1083, 2021. [45] Chenwang Wu, Defu Lian, et al. Triple adversarial learning for influence based poisoning attack in recommender systems. In SIGKDD, pages 1830 1840, 2021. [46] Huijun Wu, Chen Wang, Yuriy Tyshetskiy, et al. Adversarial examples for graph data: deep insights into attack and defense. In IJCAI, pages 4816 4823, 2019. [47] Le Wu et al. A survey on neural recommendation: From collaborative filtering to content and context enriched recommendation. ar Xiv preprint ar Xiv:2104.13030, 2021. [48] Shiwen Wu, Fei Sun, Wentao Zhang, and Bin Cui. Graph neural networks in recommender systems: a survey. ar Xiv preprint ar Xiv:2011.02260, 2020. [49] Guolei Yang, Neil Zhenqiang Gong, and Ying Cai. Fake co-visitation injection attacks to recommender systems. In NDSS, 2017. [50] Rex Ying et al. Graph convolutional neural networks for web-scale recommender systems. In SIGKDD, pages 974 983, 2018. [51] Hengtong Zhang et al. Data poisoning attack against recommender system using incomplete and perturbed data. In SIGKDD, pages 2154 2164, 2021. [52] Hengtong Zhang, Yaliang Li, Bolin Ding, and Jing Gao. Practical data poisoning attack against next-item recommendation. In WWW, pages 2458 2464, 2020. [53] Shuai Zhang, Lina Yao, Aixin Sun, and Yi Tay. Deep learning based recommender system: A survey and new perspectives. CSUR, 52(1):1 38, 2019. [54] Ziwei Zhang, Xin Wang, and Wenwu Zhu. Automated machine learning on graphs: A survey. ar Xiv preprint ar Xiv:2103.00742, 2021. [55] Huan Zhao, Lanning Wei, and Quanming Yao. Simplifying architecture search for graph neural network. ar Xiv preprint ar Xiv:2008.11652, 2020. Please do not modify the questions and only use the provided macros for your answers. Note that the Checklist section does not count towards the page limit. In your paper, please delete this instructions block and only keep the Checklist section heading above along with the questions/answers below. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [Yes] This paper mainly reveals the potential way to attack some recommendation systems. We discuss some existing defense models and future directions in Appx. B.12 (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experi- mental results (either in the supplemental material or as a URL)? [No] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experi- ments multiple times)? [No] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [Yes] (c) Did you include any new assets either in the supplemental material or as a URL? [No] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]