# coexposure_maximization_in_online_social_networks__734f7bd1.pdf Co-exposure maximization in online social networks Sijing Tu Department of Computer Science KTH Royal Institute of Technology Stockholm, Sweden sijing@kth.se Cigdem Aslay Department of Computer Science Aarhus University Aarhus, Denmark cigdem@cs.au.dk Aristides Gionis Department of Computer Science KTH Royal Institute of Technology Stockholm, Sweden argioni@kth.se Social media has created new ways for citizens to stay informed on societal matters and participate in political discourse. However, with its algorithmically-curated and virally-propagating content, social media has contributed further to the polarization of opinions by reinforcing users existing viewpoints. An emerging line of research seeks to understand how content-recommendation algorithms can be re-designed to mitigate societal polarization amplified by social-media interactions. In this paper, we study the problem of allocating seed users to opposing campaigns: by drawing on the equal-time rule of political campaigning on traditional media, our goal is to allocate seed users to campaigners with the aim to maximize the expected number of users who are co-exposed to both campaigns. We show that the problem of maximizing co-exposure is NP-hard and its objective function is neither submodular nor supermodular. However, by exploiting a connection to a submodular function that acts as a lower bound to the objective, we are able to devise a greedy algorithm with provable approximation guarantee. We further provide a scalable instantiation of our approximation algorithm by introducing a novel extension to the notion of random reverse-reachable sets for efficiently estimating the expected co-exposure. We experimentally demonstrate the quality of our proposal on real-world social networks. 1 Introduction Social media have created new ways for citizens to stay informed and participate in societal discourse. However, despite enabling users to access a variery of information, social media has been linked to increased societal polarization [22], by amplifying the phenomenon of echo chambers [4, 27], where users are only exposed to information from like-minded individuals, and of filter bubbles [35, 37], where algorithms only present personalized content that agrees with the user s viewpoint. To address these concerns, an emerging line of research seeks to understand how content-recommendation algorithms can be re-designed to mitigate societal polarization amplified by social-media interactions. Recent work includes developing methods for balancing [25] and diversifying [3] information exposure, while considering the tendency of the recommended content to spread through the online social network under a stochastic information propagation model. In this paper, we take a step in this direction and consider the problem of breaking filter bubbles through the information-propagation lens. Following related work that has considered the problem 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. of viral marketing for multiple items in online social networks [1, 2, 12, 20, 21, 29], we consider a setting with a centralized authority (host) that is responsible for allocating seed nodes to campaigns. We assume that two campaigns, supporting opposing sides of a controversial social issue, approach the host to benefit from its viral-marketing service. By drawing on the equal-time rule of political campaigning in the pre-digital era, our aim is to devise a seed-set allocation framework so that the expected number of users who are exposed to both campaigns, through the propagation of information in the social network, is maximized. In a traditional viral-marketing setting, both the campaigners and host would be interested in reaching out to the maximum number of people without any consideration to co-exposure [1, 2, 12, 20, 21, 29]. However, the new era of fake news and polarization has brought to the fore the fact that, algorithmic principles that work well for commercial advertising have unintended consequences when applied to political advertising, due to their commercial focus that prioritize revenue, resulting in biased and imbalanced campaigning. Such adverse effects have led to social-media platforms being held accountable for having political bias in their services. In some cases, popular social media have stopped their political advertising service at the expense of losing revenue. Thus, we assume the host has an incentive to respect balance and objectivity considerations, due to enforced legislations or social conscience. As a step of addressing the aforementioned challenges, we formally introduce co-exposure maximization (COEM) as the problem of assigning seed sets to each campaign such that the expected number of users co-exposed to campaigns under a stochastic information propagation model is maximized. We show that the COEM problem is NP-hard and NP-hard to approximate within a factor better than 1 1/e. Although the co-exposure function is neither submodular nor supermodular, we propose a greedy algorithm that exploits a connection to a submodular function that acts as the lower bound of the objective and obtain bounded approximation guarantees. Due to the #P-hardness of expected spread computation, we introduce a novel extension to the notion of random reverse-reachable sets [9] for efficiently estimating the expected co-exposure. Finally, we experimentally evaluate our algorithm on several real-world datasets and demonstrate its superiority over several baselines. Omitted proofs and implementation are provided as supplementary material. 2 Related work Our work relates to the emerging line of research on breaking filter bubbles in social media through information-propagation lens. There have been a number of studies on the effects of echo chambers [4, 26] and filter bubbles [4, 17, 22, 37]. In particular, it has been observed that news stories containing opinion-challenging information spread less than other news [26] and filtering of content by a social-network owner to increase user engagement can significantly increase societal polarization [17]. Recent approaches to breaking filter bubbles focus on making recommendations to individuals of opposing viewpoints [23, 24, 31], targeting users so as to reduce the polarization of opinions and bridge opposing views by considering opinion-formation models [16, 33, 34], or addressing these issues under information-propagation models [3, 25] as we do in our work. Aslay et al. [3] study the related problem of diversifying exposure to information that is propagating in a social network. Their problem formulation assumes that the leanings of users and news articles are quantified in the interval [ 1, 1] and are known. The goal is to find an assignment of articles to seed users to maximize the total diversity over all users in the network. The diversity of a user is defined to be a function that takes as input the leanings of the set of news articles that the user is exposed as well as the learning of the user. Therefore, users who are exposed to only one article, which has a different leaning from their own, still contribute to the value of the diversity objective. Translating this formulation to our setting, by considering two articles with leanings 1 and 1, implies that their objective function can potentially achieve a relatively high value while the co-exposure being equal to 0. Thus, their work does not guarantee that co-exposure is maximized. Moreover, Aslay et al. [3] also propose an extension to random reverse-reachable sets [9] for scalable estimation of their objective function. For this task, they sample random sets defined over user-article pairs while our sampling domain is user-user pairs as we explain in Section 5. Due to the difference in the objective functions, hence, the estimation task, the sample-complexity results and the sample of random sets obtained for one problem cannot be used to solve the other. The work most related to ours is the one by Garimella et al. [25], in which they consider a similar information-propagation setting with two opposing campaigns. Different from our work, they assume a set of initial seed sets to be given for each campaign and aim to recruit additional seed sets to maximize the expected number of users exposed either to both or none of the campaigns. The seed sets are not required to be disjoint, which makes their setting less realistic, as the seeds of a campaign are more likely to take one-sided stance. Furthermore, among the three algorithms they propose, only one provides an approximation guarantee for the setting where campaigns have different propagation probabilities, while the other two algorithms rely on limiting assumptions, such as, campaigns having same propagation probabilities and choosing common seeds for both campaigns. Another difference with the current work is that the co-exposure function we consider here is arguably a more natural choice for reducing polarization, as it accounts solely for nodes that are informed by both sides of a controversial issue. Finally, we note that Garimella et al. [25] use computationally prohibitive Monte Carlo simulations while we propose an efficient co-exposure estimation framework. 3 Problem definition Ingredients. The input to our problem consists of: (i) a directed social graph G = (V, E) with |V | = n nodes and |E| = m edges, where a directed edge (u, v) indicates that node v follows node u, thus, v can see and propagate posts by u; (ii) two campaigners, campaigning for opposing sides of a controversial issue, with their campaigns hereafter referred as campaign r (red) and campaign b (blue), and their seed set budgets denoted by kr Z+ and kb Z+, respectively; and (iii) campaign-specific propagation probabilities pr uv and pb uv, for all (u, v) E, representing the probability that a post from node u will propagate to node v in the respective campaigns. Given the budgets kr and kb of the campaigners, the host is in charge of selecting disjoint seed sets Sr and Sb for advertiser r and b, respectively, while respecting their seed set budgets, i.e., |Sr| kr and |Sb| kb. The goal, on a high level, is to select seeds for the two campaigns, within the allocated budgets, so as to maximize the number of nodes in the network who are exposed to both campaigns. Propagation model. We assume that the propagation of each campaign follows the independentcascade (IC) model [28], each with campaign-specific propagation probabilities pr uv and pb uv, for all directed edges (u, v) E. We assume that the propagation of a campaign through the edge (u, v) is independent of v s activation status on the other campaign. This way, we are able to take into account the tendency of nodes to adopt information they agree or disagree with, and model a realistic setting where users can adopt more than one campaign. This is in contrast to the competitive propagation models that assume that a user can adopt only one campaign [7, 10, 15, 41, 32]. Thus, once a node u becomes active at time t on campaign r (respectively, b), it has one shot to activate each inactive out-neighbor v at time t + 1, with probability pr uv (respectively, pb uv), independently of the history thus far. We say that a node u is exposed to a specific campaign if u is activated on that campaign, either by an in-neighbor that is active on the same campaign, or by directly being a seed node of the campaign. Possible-world semantics. Given any two sets of seeds Sr and Sb for the two campaigns, a single possible world represents an outcome of the stochastic propagation processes starting from the nodes in Sr and Sb. To formalize the possible-world semantics of our problem, we adopt the edge-colored multigraph representation introduced by Aslay et al. [3]. Accordingly, we define a directed edgecolored multigraph G = (V, E, p) from G = (V, E), by creating a parallel edge (u, v)i associated with color ci and probability pi uv, for each campaign i {r, b}. This way, G can be regarded as a probability distribution over all the possible subgraphs of (V, E). The probability of a possible world w G, obtained by sampling each (u, v)i independently with probability pi uv, is thus given by (u,v)i w pi uv Y (u,v)i E\w (1 pi uv). Let Iw(Sr) (respectively, Iw(Sb)) denote the set of nodes that are reachable from the nodes in Sr (respectively, Sb) in w, by using the edges associated with color cr (respectively, cb). Let also Cw(Sr, Sb) = |Iw(Sr) Iw(Sb)| denoted to set of nodes co-exposed to both campaigns in w. Then, the expected number of nodes co-exposed to both campaigns is given by E[C(Sr, Sb)] = X w G Pr[w] |Iw(Sr) Iw(Sb)|. We are now ready to formally define the problem we study in this paper. Problem 3.1 (CO-EXPOSURE MAXIMIZATION (COEM)). Given a directed social graph G = (V, E), two opposing campaigns r and b, campaign-specific propagation probabilities pr uv and pb uv, for all (u, v) E, and two positive integers kr and kb, find two disjoint seed sets Sr and Sb, such that |Sr| kr and |Sb| kb, and the expected number of nodes co-exposed to the r and b campaigns is maximized. That is, maximize Sr,Sb V E [C(Sr, Sb)] subject to Sr Sb = , |Sr| kr, |Sb| kb. 4 Theoretical analysis In this section, we establish the complexity of the COEM problem, and we theoretically analyze its objective function. We start by showing that COEM is NP-hard to approximate within a factor better than 1 1/e. Theorem 4.1. It is NP-hard to approximate the COEM problem within a factor better than 1 1 Next we proceed to analyze the properties of the objective of the COEM problem. Our first observation is that when the seed set of one of the campaigns is fixed, e.g., Sr, the objective function E[C(Sr, Sb)] is submodular in Sb, and vice versa. This observation follows from the submodularity of E[I] under the independent-cascade model [28]. However, the COEM problem requires to evaluate E[C(Sr, Sb)] over the pair of seed sets. Bi-submodularity [38, 36] extends the concept of submodularity to bi-set functions, i.e., set functions with two arguments. We first show that E[C], which is a bi-set function, is not bi-submodular. Lemma 4.2. The function E[C] : 2V 2V Z 0 is a non-decreasing bi-set function, which is not bi-submodular. In the rest of the section, we will first provide an equivalent univariate formulation of our objective, by defining a function f : 2E Z 0 on the ground set E of ordered pairs of nodes. We will then show that E[f] is neither submodular nor supermodular. For non-decreasing functions that are not submodular, the greedy algorithm, although being useful in practice, might not provide theoretical performance guarantees. However, in cases that the function is close to being submodular, the performance of the greedy algorithm tends to improve. The deviation of a function from submodularity is typically captured by the submodularity ratio of the function [18, 19]. This observation has been used in the literature to show that the greedy algorithm enjoys a tight approximation guarantee of 1 γc (1 e γcγr) for maximizing a non-submodular function subject to a cardinality constraint, where γr and γc are the submodularity ratio and the generalized curvature of the non-submodular function, respectively [8]. We note that such efforts are dedicated to approximating a non-submodular set function under a cardinality constraint. As we will see next, the constraints of our problem form an independence system on the ground set E of ordered pairs of nodes. We also note that E[f] has a submodularity ratio of 0 for many problem instances, limiting the possibility of obtaining a bounded approximation when greedily maximizing the objective function. Hence, instead, we will exploit a connection of E[f] to a submodular function, that acts as its lower bound, to obtain bounded approximation guarantees. We now start formalizing our approach built on a set system defined on ordered pairs of nodes, which we name as set-of-pairs system, and establish its equivalence to the pairs of seed sets. Let (S r, S b ) denote the optimal pair of seed sets maximizing the expected co-exposure. Being a non-decreasing function, the maximum value of the objective function is attained when kr and kb seed nodes are selected for campaigns b and r, respectively. This implies that |S r| = kr and |S b | = kb. Without loss of generality we assume that kr kb. Let O1 = {(Sr, Sb) | Sr Sb = , |Sr| = kr, |Sb| = kb, Sr, Sb V } denote the set of feasible seed set pairs of maximal size. Notice that (S r, S b ) O1. For any (Sr, Sb) O1, it follows that 1 |Sb| kr . This implies that we can construct a set of pairings of the nodes in Sr and Sb such that each element in Sr corresponds to at least one and at most kb kr elements in Sb. Let E = {V V } \ {(v, v) | v V } denote a ground set of ordered pairs, where (u, v) E represents the pairing of a node u, selected as a seed node for campaign r with a node v selected as a seed node for campaign b. We now formally define set-of-pairs system defined on E, and establish its relation to O1. Definition 4.1 (Set-of-pairs system). Let (E, I) be a set system where E = {V V }\{(v, v) | v V } is the ground set and I is a collection of subsets of E. For any X I, let Xr = S{r | (r, b) X} and Xb = S{b | (r, b) X}. We say that (E, I) is a set-of-pairs system if for any set X I the following conditions hold: (i) |Xr| kr; (ii) |Xb| = |X| kb; (iii) Xr Xb = ; (iv) for each r0 Xr, |S{b | (r0, b) X}| kb Lemma 4.3. Let O2 = {(Xr, Xb) | X I}. Then O1 O2. Let f : 2E Z 0 be a function defined as f(X) = |I(Xr) I(Xb)| where I(Xr) and I(Xb) are the random variables representing the set of the nodes exposed to campaigns r and b, respectively. Since (S r, S b ) O1, it follows from Lemma 4.3 that (S r, S b ) O2. Thus, the COEM problem can be equivalently formulated as max X I E[f(X)]. (1) Lemma 4.4. The function f : 2E Z defined above is a non-decreasing set function, which is neither submodular nor supermodular. Next, we give a formal definition of the submodularity ratio and formalize the problem instances in which γr = 0. Definition 4.2 (Submodularity ratio [8]). The submodularity ratio of a non-negative set function F is the largest scalar γr such that X e X\L F(L {e}) F(L) γr (F(L X) F(L)) , for all L, X E. The set function F is submodular if and only if γr = 1. For each u V , let σ(u) denote the set of nodes that node u can reach in G. Assume that there exists at least two ordered pairs (r1, b1) and (r2, b2) in E, where σ(ri) σ(bi) = , for i = 1, 2, and there exist i, j {1, 2}, with i = j, such that σ(ri) σ(bj) = . Let L = and X = {(r1, b1), (r2, b2)}. Notice that, in any realization of the stochastic propagation process starting from the seed sets Xr and Xb, we have f({(ri, bi)}) = 0, hence, E[f({(ri, bi)})] = 0. Moreover, given our assumption that σ(ri) σ(bj) = in at least one of the cross pairings, we have E[f(X)] > 0, yielding 0 γr E[f(X)], thus, γr = 0. We have shown that E[f(X)] is not submodular, with a submodularity ratio of 0 in some problem instances. We now introduce a function g : 2E Z 0, defined as g(X) = | (r,b) X(I(r) I(b))|, which we show is submodular and differs from f(X) within a multiplicative factor in any realization of the stochastic propagation process. Lemma 4.5. The function g : 2E Z defined above is a submodular non-decreasing set function. Notice that submodularity of g implies submodularity of E[g] since non-negative linear combination of submodular functions is also submodular. We now prove the relation between the optimal solutions maximizing E[f] and E[g] subject to X I. Lemma 4.6. Let X0 = arg max X I E[g(X)], X = arg max X I E[f(X)], and kr kb. Then E[f(X )] kr E[g(X0)]. Lemma 4.6 suggests that any algorithm that provides an approximation guarantee for the problem of maximizing E[g] over the set system (E, I), provides also a bounded approximation guarantee for the COEM problem. We now study the properties of the set-of-pairs system (E, I). We first provide the preliminary definitions. Definition 4.3 (Independence system). A set system (E, I) is an independence system if I is nonempty and satisfies the downward-closure property, i.e., X I and Y X imply Y I. Algorithm 1: Pairs-Greedy Input :G = (V, E, p), (E, I). Output :XG 2 while E = do 3 y = arg maxx:XG {x} I E[g(XG {x})] E[g(XG)] 4 E E \ {y} 5 XG = XG {y} 7 return XG Definition 4.4 (p-system [11]). An independence system (E, I) is said to be a p-system, if max Y E max J:J is a base of Y |J| min J:J is a base of Y |J| p, where any subset J of Y is a base of Y if J I and for all e Y \ J it is J {e} I. We now show that (E, I) is an independence system. Lemma 4.7. The set-of-pairs system (E, I) is an independence system. Lemma 4.8. The independence system (E, I) is a 2 kb kr -system. Combining the result of Lemma 4.8 that (E, I) is a 2 kb kr -system with the monotonicity and submodularity of E[g], the greedy algorithm (Algorithm 1) provides an 1 1+2 kb kr -approximation [11]. Let XG E denote the solution returned by Algorithm 1. Let also XG r = S{r | (r, b) XG} and XG b = S{b | (r, b) XG}. We now show that, the pair (XG r , XG b ) of seed sets provides 1/((1 + 2 kb kr )kr)-approximation to the optimal solution of the COEM problem. Theorem 4.9. Let XG be the solution returned by Algorithm 1 and let XG f = {(r, b) | r XG r , b XG b } denote all possible ordered pairings between XG r and XG b . Then E[f(XG f )]) 1 (1 + 2 kb kr )kr E[f(X )]. 5 Algorithms Efficient implementation of the greedy algorithm is a challenge as the computation of the expected spread E[I(S)] for any given S is a #P-hard problem under the independent-cascade model [14]. A common practice is to estimate the expected spread using Monte Carlo simulations [28]. However, accurate estimation requires a large number of Monte Carlo simulations, which is prohibitively expensive, and results in a time complexity of O(kbmn2r) when using r rounds of Monte Carlo simulations in Algorithm 1. Considerable effort has been devoted to developing scalable influence-maximization algorithms. Borgs et al. [9] made a breakthrough by introducing the notion of sampling reverse-reachable (RR) sets, and proposed a quasi-linear time randomized algorithm for the influence-maximization problem. Tang et al. improved it to a near-linear time randomized algorithm, called two-phase influence maximization (TIM) [39], and subsequently proposed IMM [40] with a tightened lower bound on the sample size required to estimate the expected spread with high probability. Random RR-sets are critical for efficient estimation of the expected influence spread. However, they are designed for the standard influence-maximization problem. In this section, we introduce a non-trivial generalization of RR-sets, which we name reverse-reachable pairs sets (RRP-sets), and accordingly devise an estimator for accurate estimation of E[g(X)] for any X I. Definition 5.1 (Random reverse-reachable pairs (RRP) set). Let w G be any possible world and let v V be a node uniformly sampled at random. A random RRP-set R E is defined as the set of ordered pairs that can reach node v via the colored edges in w, i.e., R = {(r, b) : v Iw(r) Iw(b)}. A RRP-set can be sampled efficiently by first sampling a node v V uniformly at random, then performing a randomized breadth-first search starting from v in G as follows. Let N in(v) denote the set of in-neighbors of node v in G. Initially, for each campaign i {r, b}, create an empty breadth-first search queue Qi, and for each z N in(v), insert z into Qi with probability pi zv. The following loop is executed until Qi is empty: dequeue a node u from Qi and examine its incoming edges of color i: for each edge (w, u)i, insert w into Qi with probability pi wu. Let Ai denote the set of nodes dequeued from Qi. Then, a RRP-set R is constructed from the set of all possible ordered pairings between Ar and Ab, i.e., R = {(r, b) | r Ar, b Ab}. Let R denote a sample of random RRP-sets generated by the procedure described above. Given a sample R, let FR(X) = P R R 1[R X = ]/|R| denote the fraction of RRP-sets that have non-empty intersection with X. Next, we show that, for any given X E, we can estimate E[g(X)] by estimating E[FR(X)]. We have the following. Lemma 5.1. For any X E, we have E[g(X)] = n E[FR(X)], where the expectation is taken over the randomness in v V and w G. We now present our algorithm TCEM (two-phase co-exposure maximization) that provides an approximate-greedy solution XG to the problem of maximizing E[g(X)] using a sample R. The TCEM algorithm operates in two phases: (i) sampling phase, which determines the size of the sample required for accuracy of estimations and generates the sample R; (ii) the greedy pair selection phase, where, at each iteration, a feasible pair maximizing FR( XG {x}) FR( XG) is added to the solution XG. Next we show that, given a sample R of random RRP-sets from which we can obtain accurate estimations of E[FR(X)] for all X I with high probability, we can solve COEM accurately with high probability. Let OPT = E[g(X0)], and Ibase I be the maximal independent sets of (E, I). Theorem 5.2. Assume the greedy pair selection phase of TCEM receives as input a sample R of random RRP-sets such that |n FR(X) E[g(X)]| < ϵ holds for any X Ibase with probability at least 1 n ℓ/|Ibase|. Then, the algorithm TCEM returns an 1 (1+2 kb kr )kr ϵ -approximate solution to the COEM problem, with probability at least 1 n ℓ, and runs in time O(P Lemma 5.3. The size of Ibase satisfies |Ibase| n kr(τ+1) (kr(τ+1))! kr! (τ!)kr , where τ = kb kr . The bound is tight when kb mod kr = 0. 3 + 2 (ℓln n + ln2 + ln|Ibase|) We now give a lower bound on the size of R so that Equation (2) holds for all X Ibase. Lemma 5.4. Let R be such that |R| λ OPT. Then, Equation (2) holds for all X Ibase with probability at least 1 n ℓ. From the previous lemma, R should satisfy |R| λ/OPT, however OPT is unknown and NPhard to compute. To circumvent this problem, we follow the approach employed in previous work [3, 40] and exploit a connection to the martingale theory to adaptively estimate a lower bound of OPT by performing a statistical test B(y). We perform the test iteratively on O(log2 n) values of y = n/2, n/4, . . . , 1, such that, if OPT < y then B(y) = false. We now give the details of the sampling phase of TCEM which, by employing the statistical test, identifies a lower bound LB and generates the final sample on which the approximate greedy solution XG will be computed. The algorithm starts by initializing R = , a less stringent error parameter ϵ2 ϵ, and a naïve lower bound LB = 1. Then, it enters a for-loop with at most log2 n iterations. In the i-th iteration, the algorithm computes y = n/2i and derives 3 + 2 (ℓln n + ln log2 n + ln|Ibase|) n Then the algorithm inserts more random RRP-sets into R until |R| θi and greedily computes a solution XG i on this sample. If R satisfies the following stopping condition n FR( XG i ) (1 + ϵ2) y, (3) the algorithm sets LB = n FR( XG i ) 1+ϵ2 and terminates the for-loop. If this is the case, the algorithm generates more random RRP-sets into R until |R| λ/LB and returns R as input to the greedy pair selection phase that computes XG. Otherwise, the algorithm proceeds in the (i + 1)-th iteration. If after O(log2 n) iterations the algorithm cannot set LB, then the naïve lower bound is used. Our main result is the following. Theorem 5.5. With probability at least 1 n ℓ, TCEM returns a sample R such that |R| λ/OPT. 6 Experimental evaluation We evaluate our method against different baselines on real-world networks. We measure the coexposure size as a function of the available budget. We also evaluate the scalability of our method. The confidence and accuracy parameters are set to ℓ= 1 and ϵ = 0.2. Our experiments are performed on a server with a 2 10 core Xeon E5 2680 v2 2.80 GHz processor, with 256 GB memory. Datasets. We use the following networks: Flixster [6], Last.FM [5], Net HEPT [13], and Wiki Vote [30]. Basic statistics of these networks are reported in the supplementary. For each network, we use three different methods to assign independent-cascade parameters. First, we use the weightedcascade model [28], in which the probability of edge (u, v) is set to 1/d(v) for both campaigns, where d(v) is the in-degree of node v. The resulting networks are denoted by adding prefix _wc, e.g., Flixster_wc. As second and third, following [25], we devise homogeneous and heterogeneous setting using arbitrary edge probabilities: campaign-specific propagation probabilities on an edge are set to be equal in the former while they differ in the latter setting. For all the networks except Flixster, we use the trivalency model [14] to assign the homogeneous and heterogeneous edge probabilities by drawing randomly from {0.1, 0.01, 0.001} for each edge. For Flixster, we use probabilities learned in [6]. The resulting networks are denoted by adding the prefixes _het and _hom, e.g., Flixster_het and Flixster_hom. Baselines. We compare our method, TCEM, with four baselines: Degree-One, Degree-Two, MNI and Balance Exposure. The first two baselines consider the nodes in decreasing order of out-degree. For Degree-One, kr seeds are assigned to one campaign and kb to the other; for Degree-Two seeds are assigned in a round-robin fashion. MNI1 solves arg max X I|N (Xr) N (Xb)|, where N (Xi) is the union of Xi and Xi s out-neighbors, and I is given in Definition 4.1. Balance Exposure is the greedy method proposed by Garimella et al. [25], which we use without initial seeds. Budget selection. Our method takes as input seed budgets kr and kb, while Balance Exposure takes as input a single budget k, and returns optimal kr and kb for that k. To ensure a fair comparison, we first execute Balance Exposure with varying k = 50, 100, 150, 200, and use the returned values kr and kb as input for the other methods. Results. Co-exposure results for different networks are shown in Figure 1. We observe that TCEM outperforms the baselines in most networks, while different independent-cascade models and network topologies have significant impact on performance. For heterogeneous and homogeneous independentcascade models, TCEM is the best-performing method. For the weighted-cascade model, the three local algorithms that use out-degree information can perform slightly better than TCEM; this happens for instance in Wiki Vote_wc. This result is consistent with the empirical observation that nodes with high out-degree obtain large expected spread under the weighted-cascade model [28], resulting in large expected co-exposure values. However, such local baselines do not have robust performance as they are all outperformed by TCEM when different independent-cascade parameters are used. We observe that the Balance Exposure algorithm is consistently outperformed by all the algorithms in all the settings. In Figure 2 we report the memory consumption and running time of TCEM, as a function of kr + kb, for the Flixster_het dataset. We fix τ = 2 and kr = 20. We observe that memory and time increase linearly, or better, with kr + kb. As an indicative performance result, when kr = 20 and kb = 200, TCEM requires 48 GB and runs in 150 minutes. 1Short for maximum neighborhood intersection. 50 100 150 210 kr + kb exp. co-exposure 50 100 150 210 kr + kb 0 10 20 30 40 50 60 70 80 90 exp. co-exposure 50 100 150 210 kr + kb exp. co-exposure 50 100 150 210 kr + kb exp. co-exposure Balance Exposure (a) Flixster_het (b) Net HEPT_het (c) Wiki Vote_het (d) Flixster_hom 50 100 150 210 kr + kb 110 120 130 140 150 160 170 180 190 200 exp. co-exposure 50 100 150 210 kr + kb exp. co-exposure 50 100 150 210 kr + kb exp. co-exposure 50 100 150 210 kr + kb exp. co-exposure (e) Last.FM_hom (f) Wiki Vote_hom (g) Net HEPT_wc (h) Wiki Vote_wc Figure 1: Co-exposure results for different networks for varying kr + kb. 0 50 100 150 200 250 kr + kb Memory (GB) 0 50 100 150 200 250 kr + kb Time (sec x 103) 0 50 100 150 200 250 kr + kb Memory (GB) 0 50 100 150 200 250 kr + kb 2 3 4 5 6 7 8 9 Time (sec x 103) (a) Memory (τ = 2) (b) Time (τ = 2) (c) Memory (kr = 20) (d) Time (kr = 20) Figure 2: Memory consumption and running time of TCEM on the Flixster_het. 7 Conclusions In this paper, we address the problem of maximizing co-exposure in social networks. We show that the problem is NP-hard and the objective function is neither submodular nor supermodular. By exploiting a connection to a submodular function that acts as a lower bound to the objective, we devise an approximation algorithm with provable guarantee. We further propose TCEM, a scalable instantiation of our approximation algorithm that can efficiently estimate the expected co-exposure. Several directions for future work open ahead. First, it would be interesting to improve the approximation guarantee for the problem we define. Second, we would like to extend our approach to account for the social advertising setting [2] in which the advertisers, with a limited monetary budget, are required to pay a monetary amount to the host for each engagement to their virally propagating campaign. Acknowledgments Part of this work was done while the authors were with Aalto University, in Finland. This research is supported by the Academy of Finland projects AIDA (317085) and MLDB (325117), the ERC Advanced Grant REBOUND (834862), the EC H2020 RIA project So Big Data (871042), and the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation. Broader impact Our work addresses the problem of maximizing co-exposure of information in online social networks via viral-marketing strategies. We are interested in situations where opposing campaigns are propagated in different parts of a social network, with users in one side not being aware of the content and arguments seen on the other side. Although, the focus of our work is mainly theoretical, and a number of modeling considerations has been stripped out for the sake of mathematical rigor, applying this kind of ideas in practice may have significant impact towards reducing polarization on societal issues, and offering users a more balanced news diet and the possibility to participate in constructive deliberation. On the other hand, one needs to be careful how our framework will be applied in practice. One potential source of misuse is when misinformation or disinformation is offered to counter true facts. Here we assume that this aspect is orthogonal to our approach, and that the social-network platform needs to mitigate this danger by providing mechanisms of information validation, fact checking, and ethical compliance of the content before allowing it to circulate in the network. Another issue is that, users often do not understand why they see a particular item in their feed; the system content-filtering and prioritization algorithm is opaque to them. In the context of our proposal, since we are suggesting to make content recommendations to selected users, it is important that transparent mechanisms are in place for the users to opt in participating in such features, to understand why they receive these recommendations, and in general, to be able to control their content. [1] Cigdem Aslay, Wei Lu, Francesco Bonchi, Amit Goyal, and Laks VS Lakshmanan. Viral marketing meets social advertising: ad allocation with minimum regret. Proceedings of the VLDB Endowment, 8(7):814 825, 2015. [2] Cigdem Aslay, Francesco Bonchi, Laks VS Lakshmanan, and Wei Lu. Revenue maximization in incentivized social advertising. Proceedings of the VLDB Endowment, 10(11):1238 1249, 2017. [3] Cigdem Aslay, Antonis Matakos, Esther Galbrun, and Aristides Gionis. Maximizing the diversity of exposure in a social network. In 2018 IEEE International Conference on Data Mining (ICDM), pages 863 868. IEEE, 2018. [4] Eytan Bakshy, Solomon Messing, and Lada Adamic. Exposure to ideologically diverse news and opinion on facebook. Science, 2015. [5] Nicola Barbieri and Francesco Bonchi. Influence maximization with viral product design. In Proceedings of the 2014 SIAM International Conference on Data Mining, pages 55 63. SIAM, 2014. [6] Nicola Barbieri, Francesco Bonchi, and Giuseppe Manco. Topic-aware social influence propagation models. Knowledge and information systems, 37(3):555 584, 2013. [7] Shishir Bharathi, David Kempe, and Mahyar Salek. Competitive influence maximization in social networks. In International workshop on web and internet economics, pages 306 311. Springer, 2007. [8] Andrew An Bian, Joachim M Buhmann, Andreas Krause, and Sebastian Tschiatschek. Guarantees for greedy maximization of non-submodular functions with applications. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 498 507. JMLR. org, 2017. [9] Christian Borgs, Michael Brautbar, Jennifer Chayes, and Brendan Lucier. Maximizing social influence in nearly optimal time. In Proceedings of the twenty-fifth annual ACM-SIAM symposium on Discrete algorithms, pages 946 957. SIAM, 2014. [10] Allan Borodin, Yuval Filmus, and Joel Oren. Threshold models for competitive influence in social networks. In WINE, pages 539 550, 2010. [11] Gruia Calinescu, Chandra Chekuri, Martin Pál, and Jan Vondrák. Maximizing a monotone submodular function subject to a matroid constraint. SIAM Journal on Computing, 40(6): 1740 1766, 2011. [12] Parinya Chalermsook, Atish Das Sarma, Ashwin Lall, and Danupon Nanongkai. Social network monetization via sponsored viral marketing. ACM SIGMETRICS Performance Evaluation Review, 43(1):259 270, 2015. [13] Wei Chen, Yajun Wang, and Siyu Yang. Efficient influence maximization in social networks. In Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 199 208, 2009. [14] Wei Chen, Chi Wang, and Yajun Wang. Scalable influence maximization for prevalent viral marketing in large-scale social networks. In Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 1029 1038, 2010. [15] Wei Chen, Alex Collins, Rachel Cummings, Te Ke, Zhenming Liu, David Rincon, Xiaorui Sun, Yajun Wang, Wei Wei, and Yifei Yuan. Influence maximization in social networks when negative opinions may emerge and propagate. In Proceedings of the 2011 siam international conference on data mining, pages 379 390. SIAM, 2011. [16] Xi Chen, Jefrey Lijffijt, and Tijl De Bie. Quantifying and minimizing risk of conflict in social networks. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1197 1205, 2018. [17] Uthsav Chitra and Christopher Musco. Analyzing the impact of filter bubbles on social network polarization. In Proceedings of the 13th International Conference on Web Search and Data Mining, pages 115 123, 2020. [18] Abhimanyu Das and David Kempe. Submodular meets spectral: greedy algorithms for subset selection, sparse approximation and dictionary selection. In Proceedings of the 28th International Conference on International Conference on Machine Learning, pages 1057 1064, 2011. [19] Abhimanyu Das and David Kempe. Approximate submodularity and its applications: subset selection, sparse approximation and dictionary selection. The Journal of Machine Learning Research, 19(1):74 107, 2018. [20] Samik Datta, Anirban Majumder, and Nisheeth Shrivastava. Viral marketing for multiple products. In Data Mining (ICDM), 2010 IEEE 10th International Conference on, pages 118 127. IEEE, 2010. [21] Nan Du, Yingyu Liang, Maria-Florina Balcan, Manuel Gomez-Rodriguez, Hongyuan Zha, and Le Song. Scalable inuence maximization for multiple products in continuous-time diffusion networks. The Journal of Machine Learning Research, 18(1):30 74, 2017. [22] Seth Flaxman, Sharad Goel, and Justin M Rao. Filter bubbles, echo chambers, and online news consumption. Public opinion quarterly, 80(S1):298 320, 2016. [23] Kiran Garimella, Gianmarco De Francisci Morales, Aristides Gionis, and Michael Mathioudakis. Factors in recommending contrarian content on social media. In Proceedings of the 2017 ACM on Web Science Conference, pages 263 266, 2017. [24] Kiran Garimella, Gianmarco De Francisci Morales, Aristides Gionis, and Michael Mathioudakis. Reducing controversy by connecting opposing views. In Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, pages 81 90, 2017. [25] Kiran Garimella, Aristides Gionis, Nikos Parotsidis, and Nikolaj Tatti. Balancing information exposure in social networks. In Advances in Neural Information Processing Systems, pages 4663 4671, 2017. [26] R. Kelly Garrett. Echo chambers online?: Politically motivated selective exposure among internet news users. JCMC, 14(2), 2009. ISSN 1083-6101. doi: 10.1111/j.1083-6101.2009. 01440.x. URL http://dx.doi.org/10.1111/j.1083-6101.2009.01440.x. [27] R Kelly Garrett. Echo chambers online?: Politically motivated selective exposure among internet news users. Journal of Computer-Mediated Communication, 14(2):265 285, 2009. [28] David Kempe, Jon Kleinberg, and Éva Tardos. Maximizing the spread of influence through a social network. In Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining, pages 137 146. ACM, 2003. [29] Arijit Khan, Benjamin Zehnder, and Donald Kossmann. Revenue maximization by viral marketing: A social network host s perspective. In 2016 IEEE 32nd international conference on data engineering (ICDE), pages 37 48. IEEE, 2016. [30] Jure Leskovec and Andrej Krevl. SNAP Datasets: Stanford large network dataset collection. http://snap.stanford.edu/data, June 2014. [31] Q Vera Liao and Wai-Tat Fu. Expert voices in echo chambers: effects of source expertise indicators on exposure to diverse opinions. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pages 2745 2754, 2014. [32] Wei Lu, Wei Chen, and Laks VS Lakshmanan. From competition to complementarity: Comparative influence diffusion and maximization. Proceedings of the VLDB Endowment, 9(2), 2015. [33] Antonis Matakos, Sijing Tu, and Aristides Gionis. Tell me something my friends do not know: Diversity maximization in social networks. Knowledge and Information Systems, 62(9): 3697 3726, 2020. [34] Cameron Musco, Christopher Musco, and Charalampos E Tsourakakis. Minimizing polarization and disagreement in social networks. In Proceedings of the 2018 World Wide Web Conference, pages 369 378, 2018. [35] Dimitar Nikolov, Diego FM Oliveira, Alessandro Flammini, and Filippo Menczer. Measuring online social bubbles. Peer J Computer Science, 1:e38, 2015. [36] Naoto Ohsaka and Yuichi Yoshida. Monotone k-submodular function maximization with size constraints. In Advances in Neural Information Processing Systems, pages 694 702, 2015. [37] Eli Pariser. The Filter Bubble: What the Internet Is Hiding from You. Penguin UK, 2011. [38] Ajit Singh, Andrew Guillory, and Jeff Bilmes. On bisubmodular maximization. In Artificial Intelligence and Statistics, pages 1055 1063, 2012. [39] Youze Tang, Xiaokui Xiao, and Yanchen Shi. Influence maximization: Near-optimal time complexity meets practical efficiency. In Proceedings of the 2014 ACM SIGMOD international conference on Management of data, pages 75 86. ACM, 2014. [40] Youze Tang, Yanchen Shi, and Xiaokui Xiao. Influence maximization in near-linear time: A martingale approach. In Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, pages 1539 1554, 2015. [41] Isabel Valera and Manuel Gomez-Rodriguez. Modeling adoption and usage of competing products. In ICDM, 2015. doi: 10.1109/ICDM.2015.40. URL https://doi.org/10.1109/ ICDM.2015.40.