# robust_classification_under_sample_selection_bias__bbb15ab2.pdf Robust Classification Under Sample Selection Bias Anqi Liu Department of Computer Science University of Illinois at Chicago Chicago, IL 60607 aliu33@uic.edu Brian D. Ziebart Department of Computer Science University of Illinois at Chicago Chicago, IL 60607 bziebart@uic.edu In many important machine learning applications, the source distribution used to estimate a probabilistic classifier differs from the target distribution on which the classifier will be used to make predictions. Due to its asymptotic properties, sample reweighted empirical loss minimization is a commonly employed technique to deal with this difference. However, given finite amounts of labeled source data, this technique suffers from significant estimation errors in settings with large sample selection bias. We develop a framework for learning a robust bias-aware (RBA) probabilistic classifier that adapts to different sample selection biases using a minimax estimation formulation. Our approach requires only accurate estimates of statistics under the source distribution and is otherwise as robust as possible to unknown properties of the conditional label distribution, except when explicit generalization assumptions are incorporated. We demonstrate the behavior and effectiveness of our approach on binary classification tasks. 1 Introduction The goal of supervised machine learning is to use available source data to make predictions with the smallest possible error (loss) on unlabeled target data. The vast majority of supervised learning techniques assume that source (training) data and target (testing) data are drawn from the same distribution over pairs of example inputs and labels, P(x, y), from which the conditional label distribution, P(y|x), is estimated as ˆP(y|x). In other words, data is assumed to be independent and identically distributed (IID). For many machine learning applications, this assumption is not valid; e.g., survey response rates may vary by individuals characteristics, medical results may only be available from a non-representative demographic sample, or dataset labels may have been solicited using active learning. These examples correspond to the covariate shift [1] or missing at random [2] setting where the source dataset distribution for training a classifier and the target dataset distribution on which the classifier is to be evaluated depend on the example input values, x, but not the labels, y [1]. Despite the source data distribution, P(y|x)Psrc(x), and the target data distribution, P(y|x)Ptrg(x), sharing a common conditional label probability distribution, P(y|x), all (probabilistic) classifiers, ˆP(y|x), are vulnerable to sample selection bias when the target data and the inductive bias of the classifier trained from source data samples, Psrc(x) P(y|x), do not match [3]. We propose a novel approach to classification that embraces the uncertainty resulting from sample selection bias by producing predictions that are explicitly robust to it. Our approach, based on minimax robust estimation [4, 5], departs from the traditional statistics perspective by prescribing (rather than assuming) a parametric distribution that, apart from matching known distribution statistics, is the worst-case distribution possible for a given loss function. We use this approach to derive the robust bias-aware (RBA) probabilistic classifier. It robustly minimizes the logarithmic loss (logloss) of the target prediction task subject to known properties of data from the source distribution. The parameters of the classifier are optimized via convex optimization to match statistical properties measured from the source distribution. These statistics can be measured without the inaccuracies introduced from estimating their relevance to the target distribution [1]. Our formulation requires any assumptions of statistical properties generalizing beyond the source distribution to be explicitly incorporated into the classifier s construction. We show that the prevalent importance weighting approach to covariate shift [1], which minimizes a sample reweighted logloss, is a special case of our approach for a particularly strong assumption: that source statistics fully generalize to the target distribution. We apply our robust classification approach on synthetic and UCI binary classification datasets [6] to compare its performance against sample reweighted approaches for learning under sample selection bias. 2 Background and Related Work Under the classical statistics perspective, a parametric model for the conditional label distribution, denoted ˆPθ(y|x), is first chosen (e.g., the logistic regression model), and then model parameters are estimated to minimize prediction loss on target data. When source and target data are drawn from the same distribution, minimizing loss on samples of source data, Psrc(x) P(y|x), argmin θ E Psrc(x) P (y|x)[loss( ˆPθ(Y |X), Y )], (1) efficiently converges to the target distribution (Ptrg(x)P(y|x)) loss minimizer. Unfortunately, minimizing the sample loss (1) when source and target distributions differ does not converge to the target loss minimizer. A preferred approach for dealing with this discrepancy is to use importance weighting to estimate the prediction loss under the target distribution by reweighting the source samples according to the target-source density ratio, Ptrg(x)/Psrc(x) [1, 7]. We call this approach sample reweighted loss minimization, or the sample reweighted approach for short in our discussion in this paper. Machine learning research has primarily investigated sample selection bias from this perspective, with various techniques for estimating the density ratio including kernel density estimation [1], discriminative estimation [8], Kullback-Leibler importance estimation [9], kernel mean matching [10, 11], maximum entropy methods [12], and minimax optimization [13]. Despite asymptotic guarantees of minimizing target distribution loss [1] (assuming Ptrg(x) > 0 = Psrc(x) > 0), EPtrg(x)P (y|x)[loss( ˆPθ(Y |X), Y )] = lim n E P (n) src (x) P (y|x) Psrc(X)loss( ˆPθ(Y |X), Y ) Sample reweighted objective function Dataset #1 Dataset #2 Figure 1: Datapoints (with + and o labels) from two source distributions (Gaussians with solid 95% confidence ovals) and the largest data point importance weights, Ptrg(x)/Psrc(x), under the target distributions (Gaussian with dashed 95% confidence ovals). sample reweighting is often extremely inaccurate for finite sample datasets, Psrc(x), when sample selection bias is large [14]. The reweighted loss (2) will often be dominated by a small number of datapoints with large importance weights (Figure 1). Minimizing loss primarily on these datapoints often leads to target predictions with overly optimistic confidence. Additionally, the specific datapoints with large importance weights vary greatly between random source samples, often leading to high variance model estimates. Formal theoretical limitations match these described shortcomings; generalization bounds on learning under sample selection bias using importance weighting have only been established when the first moment of sampled importance weights is bounded, EPtrg(x)[Ptrg(X)/Psrc(X)] < [14], which imposes strong restrictions on the source and target distributions. For example, neither pair of distributions in Figure 1 satisfies this bound because the target distribution has fatter tails than the source distribution in some or all directions. Though developed using similar tools, previous minimax formulations of learning under sample selection bias [15, 13] differ substantially from our approach. They consider the target distribution as being unknown and provide robustness to its worst-case assignment. The class of target distributions considered are those obtained by deleting a subset of measured statistics [15] or all possible reweightings of the sample source data [13]. Our approach, in contrast, obtains an estimate for each given target distribution that is robust to all the conditional label distributions matching source statistics. While having an exact or well-estimated target distribution a priori may not be possible for some applications, large amounts of unlabeled data enable this in many batch learning settings. A wide range of approaches for learning under sample selection bias and transfer learning leverage additional assumptions or knowledge to improve predictions [16]. For example, a simple, but effective approach to domain adaptation [17] leverages some labeled target data to learn some relationships that generalize across source and target datasets. Another recent method assumes that source and target data are generated from mixtures of domains and uses a learned mixture model to make predictions of target data based on more similar source data [18]. 3 Robust Bias-Aware Approach We propose a novel approach for learning under sample selection bias that embraces the uncertainty inherent from shifted data by making predictions that are explicitly robust to it. This section mathematically formulates this motivating idea. 3.1 Minimax robust estimation formulation Minimax robust estimation [4, 5] advocates for the worst case to be assumed about any unknown characteristics of a probability distribution. This provides a strong rationale for maximum entropy estimation methods [19] from which many familiar exponential family distributions (e.g., Gaussian, exponential, Laplacian, logistic regression, conditional random fields [20]) result by robustly minimizing logloss subject to constraints incorporating various known statistics [21]. Probabilistic classification performance is measured by the conditional logloss (the negative conditional likelihood), logloss Ptrg(X)(P(Y |X), ˆP(Y |X)) EPtrg(x)P (y|x)[ log P(Y |X)], of the es- timator, ˆP(Y |X), under an evaluation distribution (i.e., the target distribution, Ptrg(X)P(Y |X), for the sample selection bias setting). We assume that a set of statistics, denoted as convex set Ξ, characterize the source distribution, Psrc(x, y). Using this loss function, Definition 1 forms a robust minimax estimate [4, 5] of the conditional label distribution, ˆP(Y |X), using a worst-case conditional label distribution, ˇP(Y |X). Definition 1. The robust bias-aware (RBA) probabilistic classifier is the saddle point solution of: min ˆ P (Y |X) Δ max ˇ P (Y |X) Δ Ξ logloss Ptrg(X) ˇP(Y |X), ˆP(Y |X) , (3) where Δ is the conditional probability simplex: x X, y Y : P(y|x) 0; y Y P(y |x) = 1. This formulation can be interpreted as a two-player game [5] in which the estimator player first chooses ˆP(Y |X) to minimize the conditional logloss and then the evaluation player chooses distribution ˇP(Y |X) from the set of statistic-matching conditional label distributions to maximize conditional logloss. This minimax game reduces to a maximum conditional entropy [19] problem: Theorem 1 ([5]). Assuming Ξ is a set of moment-matching constraints, EPsrc(x) ˆ P (y|x)[f(X, Y )] = c EPsrc(x)P (y|x)[f(X, Y )], the solution of the minimax logloss game (3) maximizes the target distribution conditional entropy subject to matching statistics on the source distribution: max ˆ P (Y |X) Δ HPtrg(x), ˆ P (y|x)(Y |X) such that: EPsrc(x) ˆ P (y|x)[f(X, Y )] = c. (4) Conceptually, the solution to this optimization (4) has low certainty where the target density is high by matching the source distribution statistics primarily where the target density is low. 3.2 Parametric form of the RBA classifier Using tools from convex optimization [22], the solution to the dual of our constrained optimization problem (4) has a parametric form (Theorem 2) with Lagrange multiplier parameters, θ, weighing Logistic regression Reweighted Robust bias-aware Figure 2: Probabilistic predictions from logistic regression, sample reweighted logloss minimization, and robust bias-aware models ( 4.1) given labeled data ( + and o classes) sampled from the source distribution (solid oval indicating Gaussian covariance) and a target distribution (dashed oval Gaussian covariance) for first-order moment statistics (i.e., f(x, y) = [y yx1 yx2]T ). the feature functions, f(x, y), that constrain the conditional label distribution estimate (4) (derivation in Appendix A). The density ratio, Psrc(x)/Ptrg(x), scales the distribution s prediction certainty to increase when the ratio is large and decrease when it is small. Theorem 2. The robust bias-aware (RBA) classifier for target distribution Ptrg(x) estimated from statistics of source distribution Psrc(x) has a form: ˆPθ(y|x) = e Psrc(x) Ptrg(x) θ f(x,y) Psrc(x) Ptrg(x) θ f(x,y ) , (5) which is parameterized by Lagrange multipliers θ. The Lagrangian dual optimization problem selects these parameters to maximize the target distribution log likelihood: maxθ EPtrg(x)P (y|x)[log ˆPθ(Y |X)]. Unlike the sample reweighting approach, our approach does not require that target distribution support implies source distribution support (i.e., Ptrg(x) > 0 = Psrc(x) > 0 is not required). Where target support vanishes (i.e., Ptrg(x) 0), the classifier s prediction is extremely certain, and where source support vanishes (i.e., Psrc(x) = 0), the classifier s prediction is a uniform distribution. The critical difference in addressing sample selection bias is illustrated in Figure 2. Logistic regression and sample reweighted loss minimization (2) extrapolate in the face of uncertainty to make strong predictions without sufficient supporting evidence, while the RBA approach is robust to uncertainty that is inherent when learning from finite shifted data samples. In this example, prediction uncertainty is large at all tail fringes of the source distribution for the robust approach. In contrast, there is a high degree of certainty for both the logistic regression and sample reweighted approaches in portions of those regions (e.g., the bottom left and top right). This is due to the strong inductive biases of those approaches being applied to portions of the input space where there is sparse evidence to support them. The conceptual argument against this strong inductive generalization is that the labels of datapoints in these tail fringe regions could take either value and negligibly affect the source distribution statistics. Given this ambiguity, the robust approach suggests much more agnostic predictions. The choice of statistics, f(x, y) (also known as features), employed in the model plays a much different role in the RBA approach than in traditional IID learning methods. Rather than determining the manner in which the model generalizes, as in logistic regression, features should be chosen that prevent the robust model from pushing all of its certainty away from the target distribution. This is illustrated in Figure 3. With only first moment constraints, the predictions in the denser portions of the target distribution have fairly high uncertainty under the RBA method. The larger number of constraints enforced by the second-order mixed moment statistics preserve more of the original distribution using the RBA predictions, leading to higher certainty in those target regions. Logistic regression Reweighted Robust bias-aware First moment Second moment Figure 3: The prediction setting of Figure 2 with partially overlapping source and target densities for first-order (top) and second-order (bottom) mixed-moments statistics (i.e., f(x, y) = [y yx1 yx2 yx2 1 yx1x2 yx2 2]T ). Logistic regression and the sample reweighted approach make high-certainty predictions in portions of the input space that have high target density. These predictions are made despite the sparseness of sampled source data in those regions (e.g., the upper-right portion of the target distribution). In contrast, the robust approach pushes its more certain predictions to areas where the target density is less. 3.3 Regularization and parameter estimation In practice, the characteristics of the source distribution, Ξ, are not precisely known. Instead, empirical estimates for moment-matching constraints, c E Psrc(x) P (y|x)[f(X, Y )], are available, but are prone to sampling error. When the constraints of (4) are relaxed using various convex norms, || c E Psrc(x) ˆ P (y|x)[f(X, Y )]|| , the RBA classifier is obtained by 1or 2-regularized maximum conditional likelihood estimation (Theorem 2) of the dual optimization problem [23, 24], θ = argmax θ EPtrg(x)P (y|x) log ˆPθ(Y |X) ||θ|| . (6) The regularization parameters in this approach can be chosen using straight-forward bounds on finite sampling error [24]. In contrast, the sample reweighted approach to learning under sample selection bias [1, 7] also makes use of regularization [9], but appropriate regularization parameters for it must be haphazardly chosen based on how well the source samples represent the target data. Maximizing this regularized target conditional likelihood (6) appears difficult because target data from Ptrg(x)P(y|x) is unavailable. We avoid the sample reweighted approach (2) [1, 7], due to its inaccuracies when facing distributions with large differences in bias given finite samples. Instead, we use the gradient of the regularized target conditional likelihood and only rely on source samples adequately approximating the source distribution statistics (a standard assumption for IID learning): θEPtrg(x)P (y|x)[log ˆPθ(Y |X)] = c E Psrc(x) ˆ P (y|x)[f(X, Y )]. (7) Algorithm 1 is a batch gradient algorithm for parameter estimation under our model. It does not require objective function calculations and converges to a global optimum due to convexity [22]. Algorithm 1 Batch gradient for robust bias-aware classifier learning. Input: Dataset {(xi, yi)}, source density Psrc(x), target density Ptrg(x), feature function f(x, y), measured statistics c, (decaying) learning rate {γt}, regularizer , convergence threshold τ Output: Model parameters θ θ 0 repeat ψ(xi, y) Psrc(x) Ptrg(x)θ f(xi, y) for all: dataset examples i, labels y ˆP(Yi = y|xi) eψ(xi,y) y eψ(xi,y ) for all: dataset examples i, labels y y Y ˆP(Yi = y|xi) f(xi, y) θ θ + γt( L + θ||θ||) until || θ||θ|| + L|| τ return θ 3.4 Incorporating expert knowledge and generalizing the reweighted approach In many settings, expert knowledge may be available to construct the constraint set Ξ instead of, or in addition to, statistics c E Psrc(x) P (y|x)[f(X, Y )] estimated from source data. Expert-provided source distributions, feature functions, and constraint statistic values, respectfully denoted P src(x), f (x, y), and c , can be specified to express a range of assumptions about the conditional label distribution and how it generalizes. Theorem 3 establishes that for empirically-based constraints provided by the expert, EPtrg(x) ˆ P (y|x)[f(X, Y )] = c E Psrc(x) P (y|x)[(Ptrg(X)/Psrc(X))f(X, Y )], corresponding to strong source-to-target feature generalization assumptions, P src(x) Ptrg(x), reweighted logloss minimization is a special case of our robust bias-aware approach. Theorem 3. When direct feature generalization of reweighting source samples to the target distribution is assumed, the constraints become EPtrg(x) ˆ P (y|x)[f(X, Y )] = c E Psrc(x) P (y|x) Ptrg(X) Psrc(X)f(X, Y ) and the RBA classifier minimizes sample reweighted logloss (2). This equivalence suggests that if there is expert knowledge that reweighted source statistics are representative of the target distribution, then these strong generalization assumptions should be included as constraints in the RBA predictor and results in the sample reweighted approach1. Figure 4: The robust estimation setting of Figure 3 (bottom, right) with assumed Gaussian feature distribution generalization (dashed-dotted oval) incorporated into the density ratio. Three increasingly broad generalization distributions lead to reduced target prediction uncertainty. Weaker expert knowledge can also be incorporated. Figure 4 shows various assumptions of how widely sample reweighted statistics are representative across the input space. As the generalization assumptions are made to align more closely with the target distribution (Figure 4), the regions of uncertainty shrink substantially. 1Similar to the previous section, relaxed constraints || c E Psrc(x) ˆ P (y|x)[f(X, Y )]|| , are employed in practice and parameters are obtained by maximizing the regularized conditional likelihood as in (6). 4 Experiments and Comparisons 4.1 Comparative approaches and implementation details We compare three approaches for learning classifiers from biased sample source data: (a) source logistic regression maximizes conditional likelihood on the source data, maxθ E Psrc(x) P (y|x)[log Pθ(Y |X) ||θ||]; (b) sample reweighted target logistic regression minimizes the conditional likelihood of source data reweighted to the target distribution (2), maxθ E Psrc(x) P (y|x)[(Ptrg(x)/Psrc(x)) log Pθ(Y |X) ||θ||]; and robust bias-aware classification robustly minimizes target distribution logloss (5) trained using direct gradient calculations (7). As statistics/features for these approaches, we consider nth order uni-input moments, e.g., yx1, yx2 2, yxn 3, . . ., and mixed moments, e.g., yx1, yx1x2, yx2 3x5x6, . . .. We employ the CVX package [25] to estimate parameters of the first two approaches and batch gradient ascent (Algorithm 1) for our robust approach. 4.2 Empirical performance evaluations and comparisons We empirically compare the predictive performance of the three approaches. We consider four classification datasets, selected from the UCI repository [6] based on the criteria that each contains roughly 1,000 or more examples, has discretely-valued inputs, and has minimal missing values. We reduce multi-class prediction tasks into binary prediction tasks by combining labels into two groups based on the plurality class, as described in Table 1. Table 1: Datasets for empirical evaluation Dataset Features Examples Negative labels Positive labels Mushroom 22 8,124 Edible Poisonous Car 6 1,728 Not acceptable all others Tic-tac-toe 9 958 X does not win X wins Nursery 8 12,960 Not recommended all others We generate biased subsets of these classification datasets to use as source samples and unbiased subsets to use as target samples. We create source data bias by sampling a random likelihood function from a Dirichlet distribution and then sample source data without replacement in proportion to each datapoint s likelihood. We stress the inherent difficulties of the prediction task that results; label imbalance in the source samples is common, despite sampling independently from the example label (given input values) due to source samples being drawn from focused portions of the input space. We combine the likelihood function and statistics from each sample to form na ıve source and target distribution estimates. The complete details are described in Appendix C, including bounds imposed on the source-target ratios to limit the effects of inaccuracies from the source and target distribution estimates. We evaluate the source logistic regression model, the reweighted maximum likelihood model, and our bias-adaptive robust approach. For each, we use first-order and second-order non-mixed statistics: x2 1y, x2 2y, . . . , x2 Ky, x1y, x2y, . . . , x Ky. For each dataset, we evaluate target distribution logloss, E Ptrg(x) P (y|x)[ log ˆP(Y |X)], averaged over 50 random biased source and unbiased target samples. We employ log2 for our loss, which conveniently provides a baseline logloss of 1 for a uniform distribution. We note that with exceedingly large regularization, all parameters will be driven to zero, enabling each approach to achieve this baseline level of logloss. Unfortunately, since target labels are assumed not to be available in this problem, obtaining optimal regularization via crossvalidation is not possible. After trying a range of 2-regularization weights (Appendix C), we find that heavy 2-regularization is needed for the logistic regression model and the reweighted model in our experiments. Without this heavy regularization, the logloss is often extremely high. In contrast, heavy regularization for the robust approach is not necessary; we employ only a mild amount of 2-regularization corresponding to source statistic estimation error. We show a comparison of individual predictions from the reweighted approach and the robust approach for the Car dataset on the left of Figure 5. The pairs of logloss measures for each of the 50 Figure 5: Left: Log-loss comparison for 50 source and target distribution samples between the robust and reweighted approaches for the Car classification task. Right: Average logloss with 95% confidence intervals for logistic regression, reweighted logistic regression, and bias-adaptive robust target classifier on four UCI classification tasks. sampled source and target datasets are shown in the scatter plot. For some of the samples, the inductive biases of the reweighted approach provide better predictions (left of the dotted line). However, for many of the samples, the inductive biases do not fit the target distribution well and this leads to much higher logloss. The average logloss for each approach and dataset is shown on the right of Figure 5. The robust approach provides better performance than the baseline uniform distribution (logloss of 1) with statistical significance for all datasets. For the first three datasets, the other two approaches are significantly worse than this baseline. The confidence intervals for logistic regression and the reweighted model tend to be significantly larger than the robust approach because of the variability in how well their inductive biases generalize to the target distribution for each sample. However, the robust approach is not a panacea for all sample selection bias problems; the No Free Lunch theorem [26] still applies. We see this with the Nursery dataset, in which the inductive biases of the logistic regression and reweighted approaches do tend to hold across both distributions, providing better predictions. 5 Discussion and Conclusions In this paper, we have developed a novel minimax approach for probabilistic classification under sample selection bias. Our approach provides the parametric distribution (5) that minimizes worstcase logloss (Def. 1), and that can be estimated as a convex optimization problem (Alg. 1). We showed that sample reweighted logloss minimization [1, 7] is a special case of our approach using very strong assumptions about how statistics generalize to the target distribution (Thm. 3). We illustrated the predictions of our approach in two toy settings and how those predictions compare to the more-certain alternative methods. We also demonstrated consistent better than uninformed prediction performance using four UCI classification datasets three of which prove to be extremely difficult for other sample selection bias approaches. We have treated density estimation of the source and target distributions, or estimating their ratios, as an orthogonal problem in this work. However, we believe many of the density estimation and density ratio estimation methods developed for sample reweighted logloss minimization [1, 8, 9, 10, 11, 12, 13] will prove to be beneficial in our bias-adaptive robust approach as well. We additionally plan to investigate the use of other loss functions and extensions to other prediction problems using our robust approach to sample selection bias. Acknowledgments This material is based upon work supported by the National Science Foundation under Grant No. #1227495, Purposeful Prediction: Co-robot Interaction via Understanding Intent and Goals. [1] Hidetoshi Shimodaira. Improving predictive inference under covariate shift by weighting the loglikelihood function. Journal of Statistical Planning and Inference, 90(2):227 244, 2000. [2] Roderick J. A. Little and Donald B. Rubin. Statistical Analysis with Missing Data. John Wiley & Sons, Inc., New York, NY, USA, 1986. [3] Wei Fan, Ian Davidson, Bianca Zadrozny, and Philip S. Yu. An improved categorization of classifier s sensitivity on sample selection bias. In Proc. of the IEEE International Conference on Data Mining, pages 605 608, 2005. [4] Flemming Topsøe. Information theoretical optimization techniques. Kybernetika, 15(1):8 27, 1979. [5] Peter D. Gr unwald and A. Phillip Dawid. Game theory, maximum entropy, minimum discrepancy, and robust Bayesian decision theory. Annals of Statistics, 32:1367 1433, 2004. [6] Kevin Bache and Moshe Lichman. UCI machine learning repository, 2013. [7] Bianca Zadrozny. Learning and evaluating classifiers under sample selection bias. In Proceedings of the International Conference on Machine Learning, pages 903 910. ACM, 2004. [8] Steffen Bickel, Michael Br uckner, and Tobias Scheffer. Discriminative learning under covariate shift. Journal of Machine Learning Research, 10:2137 2155, 2009. [9] Masashi Sugiyama, Shinichi Nakajima, Hisashi Kashima, Paul V. Buenau, and Motoaki Kawanabe. Direct importance estimation with model selection and its application to covariate shift adaptation. In Advances in Neural Information Processing Systems, pages 1433 1440, 2008. [10] Jiayuan Huang, Alexander J. Smola, Arthur Gretton, Karsten M. Borgwardt, and Bernhard Schlkopf. Correcting sample selection bias by unlabeled data. In Advances in Neural Information Processing Systems, pages 601 608, 2006. [11] Yaoliang Yu and Csaba Szepesv ari. Analysis of kernel mean matching under covariate shift. In Proc. of the International Conference on Machine Learning, pages 607 614, 2012. [12] Miroslav Dud ık, Robert E. Schapire, and Steven J. Phillips. Correcting sample selection bias in maximum entropy density estimation. In Advances in Neural Information Processing Systems, pages 323 330, 2005. [13] Junfeng Wen, Chun-Nam Yu, and Russ Greiner. Robust learning under uncertain test distributions: Relating covariate shift to model misspecification. In Proc. of the International Conference on Machine Learning, pages 631 639, 2014. [14] Corinna Cortes, Yishay Mansour, and Mehryar Mohri. Learning bounds for importance weighting. In Advances in Neural Information Processing Systems, pages 442 450, 2010. [15] Amir Globerson, Choon Hui Teo, Alex Smola, and Sam Roweis. An adversarial view of covariate shift and a minimax approach. In Joaquin Qui nonero-Candela, Mashashi Sugiyama, Anton Schwaighofer, and Neil D. Lawrence, editors, Dataset Shift in Machine Learning, pages 179 198. MIT Press, Cambridge, MA, USA, 2009. [16] Sinno Jialin Pan and Qiang Yang. A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10):1345 1359, 2010. [17] Hal Daum e III. Frustratingly easy domain adaptation. In Conference of the Association for Computational Linguistics, pages 256 263, 2007. [18] Boqing Gong, Kristen Grauman, and Fei Sha. Reshaping visual datasets for domain adaptation. In Advances in Neural Information Processing Systems, pages 1286 1294, 2013. [19] Edwin T. Jaynes. Information theory and statistical mechanics. Physical Review, 106:620 630, 1957. [20] John Lafferty, Andrew Mc Callum, and Fernando Pereira. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proc. of the International Conference on Machine Learning, pages 282 289, 2001. [21] Martin J. Wainwright and Michael I. Jordan. Graphical models, exponential families, and variational inference. Foundations and Trends in Machine Learning, 1(1-2):1 305, 2008. [22] Stephen Boyd and Lieven Vandenberghe. Convex Optimization. Cambridge University Press, 2004. [23] Miroslav Dud ık and Robert E. Schapire. Maximum entropy distribution estimation with generalized regularization. In Learning Theory, pages 123 138. Springer Berlin Heidelberg, 2006. [24] Yasemin Altun and Alex Smola. Unifying divergence minimization and statistical inference via convex duality. In Learning Theory, pages 139 153. Springer Berlin Heidelberg, 2006. [25] Michael Grant and Stephen Boyd. CVX: Matlab software for disciplined convex programming, version 2.1. http://cvxr.com/cvx, March 2014. [26] David H. Wolpert. The lack of a priori distinctions between learning algorithms. Neural Comput., 8(7):1341 1390, 1996.