# protecting_geolocation_privacy_of_photo_collections__dddf3277.pdf The Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20) Protecting Geolocation Privacy of Photo Collections Jinghan Yang, Ayan Chakrabarti, Yevgeniy Vorobeychik Computer Science & Engineering, Washington University in St. Louis {jinghan.yang, ayan, yvorobeychik}@wustl.edu People increasingly share personal information, including their photos and photo collections, on social media. This information, however, can compromise individual privacy, particularly as social media platforms use it to infer detailed models of user behavior, including tracking their location. We consider the specific issue of location privacy as potentially revealed by posting photo collections, which facilitate accurate geolocation with the help of deep learning methods even in the absence of geotags. One means to limit associated inadvertent geolocation privacy disclosure is by carefully pruning select photos from photo collections before these are posted publicly. We study this problem formally as a combinatorial optimization problem in the context of geolocation prediction facilitated by deep learning. We first demonstrate the complexity both by showing that a natural greedy algorithm can be arbitrarily bad and by proving that the problem is NP-Hard. We then exhibit an important tractable special case, as well as a more general approach based on mixedinteger linear programming. Through extensive experiments on real photo collections, we demonstrate that our approaches are indeed highly effective at preserving geolocation privacy. 1 Introduction A person s physical whereabouts the city, state, or country they are in or have traveled to is sensitive information that many desire to keep private. With routine information about everyday activities increasingly being shared online and on social media, there is a heightened risk that such data may unintentionally leak information about the geographic location of individuals. For example, disclosure of a person s current location or travel history to law-enforcement agencies, hostile governments, employers, advertising companies, etc. could result in significant real harms. In this work, we specifically address the risk of geolocation privacy from collections of photographs shared by users online. The IM2GPS method (Hays and Efros 2008) was the first to demonstrate that, even without explicit embedded geotags, image content sometimes contains enough cues to allow accurate estimation of the location where the image was taken. The accuracy of image geolocation algorithms has Copyright c 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: Illustration of our problem and approach. This photo collection has 16 photos, all taken in Barcelona, which is also the correctly predicted location when all photos are used (top left map). Greedily removing photos to ensure that Barcelona is not among top-5 predicted locations corresponds to the selection highlighted in dashed green, and the resulting top-5 predictions are on the top right map. Photos chosen by our approach are identified by a red box, and the result is shown on the lower map. since improved considerably, most recently through the use of deep convolutional neural networks (CNNs) (Weyand, Kostrikov, and Philbin 2016). Nevertheless, geolocation prediction accuracy from individual images is still limited, and the content of a single photograph is typically insufficient to unambiguously resolve the location of the photographer. However, as we show, geolocation predictions can be significantly more accurate when aggregating information across a collection of photographs known to have been taken in the same geographic location (e.g., a person s photoalbum from a specific trip, or images from different people known to be traveling together or attending the same event). While individual images can appear to have been taken in any of a large set of possible locations, the intersection of these sets for different images is often sufficient to narrow down the true location of the image collection. We consider the problem of censoring photo collections in order to preserve geolocation privacy: selecting a minimal set of images to delete from a given collection, such that the true location is not included in the most likely locations predicted from the remaining images. We consider different variants of this task: (1) defeating top-1 accuracy while minimizing deletions, to ensure the most likely predicted location is incorrect; (2) defeating top-k accuracy, such that the true location is not among the k most likely predicted locations; and (3) maximizing the rank of the true location in the list of likely predicted locations, under a fixed budget on the number of deleted images. We consider both whitebox and black-box versions of these tasks. In the white-box version, the privacy-preserving algorithm has access to the true location classifier. In the black-box version, in contrast, the algorithm instead uses a proxy classifier. We show that, except for the top-1 variant, this problem is NP-hard, because the optimal deletion set should not just censor the images that are most indicative of the true location, but also maintain the likelihood of plausible alternatives. We also specifically show that a naive greedy algorithm can be arbitrarily sub-optimal. Then, we propose a polynomial-time algorithm for the top-1 variant and present a mixed-integer linear programming (MILP) formulation for the other variants to allow the use of standard MILP solvers. Finally, we present an extensive experimental evaluation of our proposed approaches using images and albums downloaded from Flickr. 1 In particular, our experiments show that our methods significantly outperform simple greedy deletions in all task variants, and in many cases admit optimal solutions with only a small number of deleted images. Our work thus highlights the risk to geolocation privacy that exists from photo collections commonly shared by users today, while also providing an effective solution that comes at the relatively low cost of censoring a small fraction of images. Due to space limitation, some proofs are deferred to the extended version . Our code and extended version could be found in the project page 2. Related Work Sharing photographs with embedded location information can allow images to tell a richer story of 1https://www.flickr.com 2https://github.com/jinghan Y/geo Privacy Album the places and events they represent. Consequently, many cameras (especially those on mobile phones) allow users the option of embedding GPS coordinates in image EXIF tags. Hays and Efros (2008) were the first to exploit such data to train geolocation predictors, and show that geographic localization is possible from image information alone. Since then, a number of methods have made steady progress on improving the accuracy of image location estimators through the use of deep convolutional neural networks (CNN). Weyand, Kostrikov, and Philbin (2016) formulate geolocation as a classification problem, by dividing the globe into a discrete number of cells and assigning a separate label for each. Howard et al. (2017) demonstrated versions of these classifiers that can run on mobile devices, while Kim, Dunn, and Frahm (2017) considered learning representations that exploit contextual information. Vo, Jacobs, and Hays (2017) presented a study confirming the efficacy of Weyand, Kostrikov, and Philbin, as well as of using deep network features for localization by retrieval. Tian, Chen, and Shah (2017) devised a matching-based geolocation approach specifically designed for street-view images. Mura (2017) and Hal (2013) discuss the dangers of using geolocation for commercial purposes and the importance of geolocation privacy. In the security domain, most existing works propose ways to anonymize location data by replacing the associated name with an untraceable ID (Beresford and Stajano 2003; Gruteser and Grunwald 2003). Other work addresses geolocation privacy of user interfaces (Doty and Wilde 2010). Andr es et al. (2013), Xiao and Xiong (2015) and Ho and Ruan (2011) develop mechanisms for preserving geolocation privacy using differential privacy techniques which add noise to actual location information; as such, they do not deal with the issue of inferring location from a photo collection, which is the focus of our work. However, these efforts do not address the issue of location inference based on published images and photo collections. Our work is also broadly related to adversarial machine learning (Vorobeychik and Kantarcioglu 2018), although the specific problem and techniques differ considerably. In particular, our focus is not to exploit divergences between learned classifiers and human perception (by creating imperceptible perturbations that fool classifiers (Szegedy et al. 2013; Nguyen, Yosinski, and Clune 2015)), but rather to censor a small number of images to remove information that can be leveraged for geolocation. 2 Problem Formulation Geolocation Prediction from Images Predicting geographic location from a single image is often naturally cast as a classification task (Weyand, Kostrikov, and Philbin 2016). Possible locations on the globe are discretized into a discrete set C = [c1, c2, . . . c M] of M geospatial regions, and a classifier is trained to predict a probability distribution P(c|X) from an input image X for c C. Here, P(c|X) denotes the probability or classifier s belief that the image X was taken in the discretized location c. While the accuracy of per-image location classifiers is still limited, we consider the effect of aggregating location infor- mation across images. Formally, we consider the task of estimating location from a collection A = [X1, X2, . . . XN] of N images that are known to have all been taken at the same location c. Without making any assumption on the relationship between images in A (i.e., treating them as conditionally independent given c), and further assuming that all locations c C are equally likely a-priori, the probability distribution of the collection A is simply given by the product of the individual image distributions: P(c|A) = N i=1 P(c|Xi) c C N i=1 P(c |Xi) (1) For convenience, we define M dimensional score vectors {Si} as the log probabilities of locations from individual images, with the jth element defined by Sj i = log P(cj|Xi). Similarly, we define S = N i=1 Si as the score corresponding to the image collection. This implies that Sj = log P(cj|A) + z, where z is a scalar corresponding to the normalization constant in (1) (i.e., it is added to all elements of the vector S). Note that since P(cj|A) > P(cj |A) Sj > Sj , a location cj has the same relative rank in both the probability and score vectors. Thus, the location scores for an image collection are given simply by summing score vectors, as returned by a classifier, from individual images. As our experiments will show, aggregating information in this way from even a relatively small number of photos leads to significant gains in location accuracy compared to single images. Location Privacy by Selecting Images for Deletion When using the distributions P(c| ) to make inferences about location, one can choose to look at either the most likely location i.e., arg maxc P(c| ) or the set of k most likely locations. Specifically, let H(cj, S) define the set of locations that have higher scores in S than a given cj: H(cj, S) = {j : cj C, Sj Sj, j = j }. (2) If the true location for a collection is ct, then our location estimate is exactly correct (top-1 accurate) if |H(ct, S)| = 0, and top-k accurate if |H(ct, S)| < k. Our goal is to develop a privacy protection method that, given an image set A and per-image scores {Si}, selects a small set of images D A to delete from A to yield a censored image set A = A \ D, so that score vector S from the remaining images, i.e., S = i:Xi A Si, leads to incorrect location estimates. Next, we define different variants of this task for different objectives trading-off utility (number of deletions) and privacy. Top-k Guarantee: The first variant seeks to minimize the number of deletions, while ensuring that the resulting scores from the censored image set push the true location ct out of the top-k location results, for a specified k, i.e., min D |D|, s.t. |H(ct, S )| k. (3) A special case of this is when k = 1, when we wish to ensure that the true location isn t predicted to be the most likely. Fixed Deletion Budget: Another variant works with a fixed budget d on the number of images to delete, while trying to maximize the rank of the true label in the resulting score vector S , i.e., max D |H(ct, S )|, s.t. |D| d. (4) Here, we call the resulting value of |H(ct, S )| for the optimized deletion set as the protected-k , since it defeats top-k accuracy for all values of k up to and including |H(ct, S )|. Incorporating User Preference: In some cases, a user may want to specify a set of images U A that should not be deleted. In this case, we add the constraint D A \ U to the formulations (3) and (4) above. Black-box Protection: So far, we have assumed that we have access to the location classifier that will be used for inference, and therefore to the corresponding scores {Si}. However, in many real-world settings, our privacy protection algorithm will not have access to the actual classifier. In such cases, we propose using scores {Si} from a proxy classifier, and adding a scalar margin θ > 0 to the scores St i of the true location in each image. These modified scores are used to solve the optimization problems (3) and (4), and capture uncertainty about the true value of the scores Sj i (we can set θ such that the score advantage St i Sj i , of the true location over any other location in a given image from our proxy, is likely to be within θ of the true advantage). 3 Computational Hardness It is tempting to view the task of selecting images for deletion as computationally easy that it can be solved simply with a greedy selection of images with highest probabilities or scores for the true location. However, this is not the case since optimal deletion must both reduce the cumulative score of the true location, and maintain high scores for possible alternate locations. Deleting an image with high scores for other locations, as well as the true one, can often lead to no change, or even a decrease, in the relative rank of the true location. Conversely, it may sometimes be optimal to delete an image with a moderate true location score, when it is the only one in the collection with a very low score for an otherwise plausible alternate location. In this section, we discuss the sub-optimality of greedy selection, and then prove that optimal deletion is NP-hard. Consider the greedy selection method that constructs a deletion set D by selecting images with the highest values of the score St i for the true location ct. To defeat top-k accuracy in (3) for a given k, it sequentially deletes images with decreasing values St i, stopping when ct is no longer in the top-k. Given a fixed budget d for deletions in (4), it simply deletes the d images with the highest scores St i. Observation 1. Greedy selection can be arbitrarily suboptimal. Specifically, in cases with just three possible discretized locations (i.e., |C| = 3), there exists a set of score vectors corresponding to log-probabilities for N +3 images, N 1, where the first N deletions by greedy selection leave the true location as the most likely prediction, whereas deleting two optimally selected images is sufficient to make it the least likely prediction (i.e., defeat top-2 accuracy). Broadly, the above set corresponds to N score vectors with uniform scores for all locations (Sj i = log(1/3), j), and three additional vectors with lower than uniform truelocation scores (St i < log(1/3)). From the latter, two vectors each feature a very low score for each of the incorrect locations, respectively. The greedy method selects the images with uniform distributions leading to no change in rank of the true location, while deleting the latter two images makes the true location the least likely prediction. The extended version provides the explicit construction of these score vectors and a complete proof. Beyond worst-case sub-optimality, our experiments will show that greedy selection also yields poorer results than optimal deletion with our method in practice, with typical score vectors from real images. Figure 1 shows an example set of sixteen real images where, to achieve a top-5 location privacy guarantee, greedy selection requires twelve deletions while the optimal solution requires only four. We next show that the optimization problems in our privacy formulation are in fact NP-hard. Theorem 1. The optimization problem in (3), to minimize deletions for a top-k guarantee, is NP-hard k 2. Theorem 2. The optimization problem in (4), for maximizing protected-k under a fixed deletion budget, is NP-hard. We begin by introducing and proving a lemma that relates to the NP-hardness of optimal deletion for making the true location less likely than a given set of alternate locations. We then use this lemma to prove Theorem 1, and subsequently, prove Theorem 2. Lemma 1. Given an image collection A with true label ct, and a target set H C, ct / | H| of alternate locations, with fixed size k = | H|, it is NP-hard k 2 to find the smallest deletion set D A , s.t. S t S j, cj H, where S are score vectors for A = A \ D. Definition 1. Deletion for Two Alternate Locations (DTAL). Given: a collection A, pair of alternate labels H = {cp, cq}, and true label ct / H. Question: is there any deletion set D A , s.t. S t S j, cj H ? Note that although scores correspond to log-probabilities in our formulation, we place no restrictions on Sj i above, since given arbitrary score values, we can subtract an imagespecific constant log j exp(Sj i ) from the scores for all labels to yield valid log-probabilities. This results in an identical decision problem, as the same constant occurs on both sides of all inequality constraints. Definition 2. Knapsack Problem (KP). Given: Values {vi > 0} and weights {wi > 0} for a set of objects O, with a capacity constraint W and value bound V . Question: Is there a subset of objects Q O with total weight at most W > 0 and total value at least V > 0 ? Proof of Lemma 1. It is easy to see DTAL is in NP. We prove DTAL is NP-complete by reduction from KP. Given an instance of KP, we form an instance of DTAL by creating a collection A with |O| + 1 images. We set St i = 0 for all images. For 1 i |O|, we set the scores Sp i = vi and Sq i = wi. For the final image with i = |O + 1|, we set Sp i = V and Sq i = W. Any deletion set returned by DTAL can not include the last image X|O|+1, since St i > Sq i for all other images. Given this solution, we can get a corresponding solution for KP as the set of all objects for which the corresponding image Xi was not included in the deletion set. Similarly, a solution of the KP instance corresponds to a solution of DTAL, where the deletion set includes images corresponding to all objects not included in the KP solution. Since DTAL which asks if a deletion set, without any constraint on its size is NP-complete, it follows that the task of determining if a deletion set of bounded size exists is also NP-complete. The latter corresponds to the decision version of the optimization problem in Lemma 1 for k = 2. Therefore, the optimization in Lemma 1 is NP-hard for k = 2. It is then easy to show the problem is NP-hard also for any k > 2, since if there were a polynomial-time solver for k > 2, we could call it to solve the k = 2 version by setting the scores for all but two locations to , in all images. Proof of Theorem 1. The optimization problem in Theorem 1 requires making a similar decision as in Lemma 1, except that it also requires determining the best set of alternate labels H where | H| = k, ct / H. It is easy to see that this is at least as hard as the fixed alternate target set variant. Specifically, if we have a solver for the problem in Theorem 1, we can use it to solve the problem in Lemma 1 by simply setting the scores Sj i = , i, j : j = ct, j / H. Proof of Theorem 2. The fixed deletion budget optimization problem in Theorem 2 is at least as hard as the top-k guarantee problem in Theorem 1. If we have a solver for the fixed budget problem, we can call that solver sequentially with deletion budget d = (1, 2, . . . N), and stop whenever the returned protected-k satisfies the desired top-k guarantee. 4 Proposed Approach We now present algorithms for solving the privacypreserving optimization problems in our formulation. We begin with a generic approach that maps our objectives and constraints to a mixed integer linear program (MILP). This formulation is naturally also NP-hard, but it allows us to use standard MILP solvers that use various heuristics to find a solution efficiently. Finally, we had shown that the top-k guarantee problem is NP-hard k 2. For the important special case k = 1, we show that this problem can in fact be solved in polynomial time.t Mixed Integer Linear Programming Given a collection A with set of locations C, true location ct, and score vectors {Si}, we let zi {0, 1} denote whether a photo is deleted (zi = 0) or kept (zi = 1). We let hj {0, 1}, j = t denote whether location cj has a lower (hj = 0) or higher (hj = 1) score S j than that of the true location S t in the censored collection A . We enforce consistency of these values of hj to the assignment zi by adding the constraints: i zi Sj i St t 0, j = t. (5) Under these definitions, the protected-k of a given assignment is given by j =t hj, and the number of deletions by i(1 zi). Thus, our problem variant (3) where we minimize the number of deleted photos that ensure top-k location privacy corresponds to solving max {zi},{hj} j =t hj k, (6) since minimizing the number of deleted photos is equivalent to maximizing the number of photos remaining in the censored collection. Analogously, the problem of maximizing protected-k with a fixed deletion budget d corresponds to: max {zi},{hj} j =t hj, s.t i (1 zi) d. (7) Both problems above are over the binary variables zi {0, 1}, hj {0, 1}, under the additional constraints in (5). Note that the formulations above involve bilinear constraints (5), as these feature products of hj and zi. However, since these are both binary variables, it is straightforward to linearize the constraints using Mc Cormick inequalities (see the extended version for details). User Preference: User specified constraints that a set of images U not be deleted can be incorporated simply by adding the constraints zi = 1, Xi U to the MILP. Black-Box Protection: The margin θ described in Sec. 2, that accounts for discrepancies between the proxy and unknown true classifiers, can also be incorporated in our MILP by simply modifying the constraints in (5) above to: i zi Sj i St t θ 0, j = t. (8) Polynomial-Time Algorithm for top-1 Guarantee As we ve shown, the optimization problems in our formulation are NP-hard for k 2. Conversely, the special case of minimizing deletions for top-1 accuracy admits a solution in polynomial time. We begin by noting that, for this case, we must only ensure that the true location ct has a lower score than any other single location cj, j = t. First, note that for a specific choice of cj, the assignment zj i {0, 1} with fewest deletions that ensures cj has a higher score than ct is given by { zj i } = arg max {zi} i zi(Sj i St i) 0, (9) where zj i = 0 indicates that Xi is deleted, as previously. This can be solved by sorting the images Xi in descending order of the score advantage (St i Sj i ) of the true location ct over cj, and deleting images (setting zj i = 0) sequentially till the constraint above is satisfied. The solution to the original problem is then given by searching over all possible cj = ct, i.e., minj =t( i 1 zj i ). Thus, we solve (9) for each possible alternate location cj = ct, compare the number of deletions required for each, and select the answer with the fewest deletions. The overall algorithm runs in O(MN log N) time (where N = |A| is the number of images, and M = |C| the number of locations). The full algorithm is included in the extended version. 5 Experiments We now present an extensive evaluation of our method for all variants of our geolocation privacy problem, measuring their success at preserving privacy against a state-of-the-art deep neural network-based location classifier with experiments on real photographs downloaded from Flickr. Preliminaries Datasets and Discretized Locations: We conduct experiments on a set of 750k images downloaded from Flickr, where we ensure each image has a geotag to specify groundtruth location. We randomly split these into training and test sets, of size 90% and 10% of total images respectively. Following Weyand, Kostrikov, and Philbin (2016), we cast geolocation as a multi-class classification problem over a set of discretized locations C. This set is formed by an adaptive scheme to partition the globe into contiguous cells, such that each cell contains the same number of training images. This means that cell sizes vary depending on the density of photos (and indirectly, of population) in the corresponding region (e.g., oceans tend to inhabit large cells, while large metropolitan areas may be partitioned among multiple cells). We use a discretization into 512 cells (i.e., |C| = 512). Photo Collections: Photo collections for evaluation are constructed by randomly forming sets of images with the same discretized location from the test set. We consider collections of size 16, 32, 64, and 128, where collections for each size are formed as 10 random partitions of the entire testset. This gives us roughly 45k collections of size 16, 21k of size 32, 10k of size 16, and 5k of size 128. Further, we collected 10k Flickr user designated albums , each with images all from the same discretized location, and with no image overlap with the training set. The album sizes ranged between 16 and 250 (33 photos on average), with all 512 locations represented. In our evaluations, we treat these albums as photo collections, although images in photo albums tend to be more correlated than random collections. Location Classifier: We use a deep convolutional neural network with a standard VGG-16 (Vo, Jacobs, and Hays 2017) architecture to produce the probability distribution over locations from a single input image (the last layer of the network is modified from the standard architecture to have 512 channels, to match the size of C). We initialize the weights of all but the last layer of the network from a model trained for Image Net (Russakovsky et al. 2015) classification, and then train the network with a standard crossentropy loss on the entire training set (except for the results on the black-box setting, as described later). Collection Type top-1 top-5 protected-k Single Image 0.20 0.38 59.58 R-16 0.97 1.00 0.06 R-32 0.99 1.00 0.01 R-64 1.00 1.00 0.00 R-128 1.00 1.00 0.00 Albums 0.20 0.39 53.51 Table 1: Geolocation performance on single images vs image collections. We show top-1 and top-5 accuracies (fraction of instances when the true location is in the top 1 or 5 predictions), and average protected-k (where rank of true location is k + 1) for all cases. Note lower protected-k represents more accurate geolocation. Here R-N refers to randomly constructed collections of size N, while Albums corresponds to user designated photo albums. Geolocation Performance of Images vs Collections We begin with experiments that demonstrate that the threat to location privacy is significantly heightened when aggregating information across multiple images. We evaluate perimage location accuracy from our trained classifier in Table 1, and compare it to accuracies of predictions from different collections using (1) as a collection-based classifier. While performance is relatively modest with single images, it increases significantly when we use photo collections, even with only 16 images, yielding a significant privacy risk. Note that accuracies for user designated albums are considerably lower because images in these albums are often correlated with similar content in many images, since most albums have images all taken by a single photographer and in similar environs. Nevertheless, some user albums may still pose a privacy risk, and our top-k guarantee approach can be used on these, since it would delete photos only in albums that compromised location privacy. Top-k Guarantee with Minimum Deletions We now consider the effectiveness of our approach, starting with the problem in (3), where the goal is to minimize the number of deleted photos while ensuring that the true location is not in the top-k most likely predictions. We present these results in Table 2, for k = 1 and 5. We find that when using our approach, we typically need to delete only a modest 30% of the photos on average to achieve a top-1 guarantee, and < 50% of photos for the stronger top-5 guarantee note that these averages are computed only over collections where the original predictions (prior to any deletion) were already accurate in the top-1 and top-5 sense, respectively. We also see that the optimal solution from our approach always requires fewer deletions than greedy selection usually by a significant margin. Thus, our method ensures location privacy while still allowing users to share a significant fraction of their images. Maximum protected-k under Deletion Budget Next, we consider our approach for our second problem variant in (4) where, given a fixed budget on the number of Collection top-1 top-5 Type Optimal Greedy Optimal Greedy R-16 0.32 0.37 0.45 0.53 R-32 0.31 0.43 0.47 0.59 R-64 0.31 0.49 0.48 0.64 R-128 0.31 0.53 0.49 0.68 Albums 0.21 0.33 0.31 0.39 Table 2: Average fraction of deleted photos selected to achieve top-1 and top-5 geolocation privacy. We compute averages only over instances where predictions on the original whole collection was accurate in the top-1 / top-5 sense. Coll. Deletion Budget (% Photos) Type Method 12.5% 25% 37.5% 50% R-16 Optimal 0.67 3.00 10.20 29.12 Greedy 0.46 1.88 6.22 17.67 R-32 Optimal 0.17 1.14 4.84 17.49 Greedy 0.07 0.42 1.90 7.15 R-64 Optimal 0.05 0.57 2.87 10.20 Greedy 0.02 0.09 0.57 2.80 R-128 Optimal 0.03 0.38 2.06 8.07 Greedy 0.01 0.03 0.21 1.29 Albums Optimal 77.09 103.18 - - Greedy 70.78 86.31 - - Table 3: Average protected-k, such that the true label is not in top-k, for different deletion budgets as percentages of collection size. Note higher protected-k implies greater privacy. images to delete from each collection, the goal is to maximize the value of k such that the true location is not in the top-k predictions. Table 3 reports results for this setting on our various collection types, from optimal selection with our approach as well as from the greedy selection method (this time, averaging over all collection instances). We find that again, optimal selection with our approach yields significantly better privacy protection than the greedy baseline. Moreover, the same fraction of deleted images provides greater protection when the size of the collection is small. Note that average protected-k values are higher for the user albums, because the corresponding values for the uncensored collections were already high (Table 1), so much so that it is unnecessary to consider deletion budget above 25%, with k already above 100 at that point (hence the blank entries in the table). In addition, we check if the photos deleted for a given collection for one deletion budget, are also deleted given a larger budget. As our results in Table 4 indicate, this is not the case: when more deletions are possible, the optimal solution may involve keeping some of the photos that would have been deleted under a lower budget. This indicates that any sequential greedy strategy is likely sub-optimal. Top-1 Top-5 R-16 R-32 R-64 R-128 Figure 2: Effect of user-specified constraint to retain a chosen photo, while providing top-k guarantee. Rank of the chosen photo is with respect to its score for the true location St i, success rate indicates fraction of collections for which a feasible deletion set was found, and fraction of photos removed is reported averaged over successful cases. Coll. Deletion Budgets Type 12.5% 25% 25% 50% R-16 0.27 0.10 R-32 0.32 0.15 R-64 0.34 0.19 R-128 0.33 0.22 Table 4: Fraction of images that were deleted with a lowerdeletion budget, but NOT with a higher one. User Preference Next we consider the case where we incorporate userspecified constraints to retain certain photos in each collection. Naturally, the effectiveness of privacy protection will depend on how informative that photo is towards location: preserving privacy may be impossible without deleting images with clearly recognizable landmarks. For a systematic evaluation, we always choose one photo to retain per collection, with different experiments where we consistently choose the photo with the same rank in terms of true location scores (rank 1 implying photo with highest St i). Figure 2 presents these results for different collection types, for the tasks of top-1 and -5 guarantees. We show the success rate (fraction of collections where it was possible to achieve a guarantee without deleting the specified photo) and the average fraction of deletions required (for successful cases). For comparison, we show the baseline fraction of deletions from Table 2. We see that putting constraints does degrade performance requiring more deletions and sometimes leading the problem to be infeasible. This is more so for small collections. However, the rank of the photo plays a major role, and the performance drop is negligible when constraining photos that have higher rank (i.e., lower scores top-1 top-5 Figure 3: Effectiveness of black-box attacks as a function of the margin θ for R-128 collections. for the true location), especially for larger collections. Such settings are likely to match user expectations of maintaining privacy while retaining location neutral photographs. Black-Box Protection Finally, we evaluate the case when we do not have access to true classifier scores but only a proxy. For this case, we train two VGG network classifiers, each on two separate halves of the training set. We then use one classifier to derive our proxy scores and apply our approach for top-k guarantees, and evaluate its success by determining the fraction of collections for which this guarantee holds when using the other classifier. These results are shown in Figure 3, where we report both success rate and the number of deletions required for different values of margin θ. We see that our success at defeating an unknown classifier increases with higher margins, albeit at the cost of an increased number of deletions. Running Time Although our optimization problem is NP-hard, we find that a modern MILP solver, CPLEX, returns solutions in reasonable time: 9.2s for a R-128 collection and 4.8s for R-64, for the top-5 guarantee setting. 6 Conclusion Our work demonstrated that photo collections pose a clear risk to location privacy simply from aggregating the outputs of existing deep neural network-based location classifiers and proposed a framework for ameliorating this risk, through an algorithm for selecting optimal photos for deletion. While we considered that entire set was available to us prior to selection, an interesting direction of future work relates to online versions of the selection problem choosing to delete or keep/post a photo without knowing what other photos may be added to the collection. It is also worth highlighting the assumptions and limitations of our formulation. Our selection method requires access to at least a reasonable proxy for the model that an adversary may use for location inference this is implicit in the choice of the margin θ in the black-box setting. We also assume there is no source of location information other than the images themselves, and accounting for side-channels requires extensions to our framework (which could potentially be modeled as a score offset provided by the side-channel). We also assume conditional independence between different images in a collection, in deriving the collection-level location score as the sum of per-image scores. However, it is conceivable that a more sophisticated location classifier could learn to model user behavior, and leverage higherorder statistics to find, for example, that the presence of certain combinations of images in a collection are more indicative of certain locations. Acknowledgments This work was partially supported by the National Science Foundation CAREER award (IIS-1905558) and NVIDIA. References Andr es, M.; Bordenabe, N.; Chatzikokolakis, K.; and Palamidessi, C. 2013. Geo-indistinguishability: differential privacy for location-based systems. In CCS 2013, 901 914. Association for Computing Machinery. Beresford, A. R., and Stajano, F. 2003. Location privacy in pervasive computing. IEEE Pervasive Computing 2(1):46 55. Doty, N., and Wilde, E. 2010. Geolocation privacy and application platforms. In Proceedings of the 3rd ACM SIGSPATIAL International Workshop on Security and Privacy in GIS and LBS, SPRINGL 10, 65 69. New York, NY, USA: ACM. Gruteser, M., and Grunwald, D. 2003. Anonymous usage of location-based services through spatial and temporal cloaking. In Proceedings of the 1st International Conference on Mobile Systems, Applications and Services, Mobi Sys 03, 31 42. New York, NY, USA: ACM. Hal, T. J. V. 2013. Taming the golden goose: Private companies, consumer geolocation data, and the need for a class action regime for privacy protection. 15 Vand. J. Ent. & Tech. L. 713. Hays, J., and Efros, A. A. 2008. im2gps: estimating geographic information from a single image. In Proceedings of the IEEE Conf. on Computer Vision and Pattern Recognition (CVPR). Ho, S.-S., and Ruan, S. 2011. Differential privacy for location pattern mining. In Proceedings of the 4th ACM SIGSPATIAL International Workshop on Security and Privacy in GIS and LBS, SPRINGL 11, 17 24. New York, NY, USA: ACM. Howard, A. G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; and Adam, H. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. ar Xiv:1704.04861. Kim, H. J.; Dunn, E.; and Frahm, J.-M. 2017. Learned contextual feature reweighting for image geo-localization. In CVPR. Mura, R. 2017. Geolocation and targeted advertising: Making the case for heightened protections to address growing privacy concerns. 9 Buff. Intell. Prop. L.J. 77. Nguyen, A.; Yosinski, J.; and Clune, J. 2015. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 427 436. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; Berg, A. C.; and Fei-Fei, L. 2015. Image Net Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV). Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; and Fergus, R. 2013. Intriguing properties of neural networks. In Porceedings of the International Conference on Learning Representations. Tian, Y.; Chen, C.; and Shah, M. 2017. Cross-view image matching for geo-localization in urban environments. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 1998 2006. Vo, N.; Jacobs, N.; and Hays, J. 2017. Revisiting im2gps in the deep learning era. ICCV. Vorobeychik, Y., and Kantarcioglu, M. 2018. Adversarial machine learning. Synthesis Lectures on Artificial Intelligence and Machine Learning. Weyand, T.; Kostrikov, I.; and Philbin, J. 2016. Pla Net - Photo Geolocation with Convolutional Neural Networks. Cham: Springer International Publishing. 37 55. Xiao, Y., and Xiong, L. 2015. Protecting locations with differential privacy under temporal correlations. In Proceedings of the 22Nd ACM SIGSAC Conference on Computer and Communications Security, CCS 15, 1298 1309. New York, NY, USA: ACM.