# random_tessellation_forests__c383cefd.pdf Random Tessellation Forests Shufei Ge1 shufei_ge@sfu.ca Shijia Wang2,1 shijia_wang@sfu.ca Yee Whye Teh3 y.w.teh@stats.ox.ac.uk Liangliang Wang1 liangliang_wang@sfu.ca Lloyd T. Elliott1 lloyd_elliott@sfu.ca 1Department of Statistics and Actuarial Science, Simon Fraser University, Canada 2School of Statistics and Data Science, LPMC & KLMDASR, Nankai University, China 3Department of Statistics, University of Oxford, UK Space partitioning methods such as random forests and the Mondrian process are powerful machine learning methods for multi-dimensional and relational data, and are based on recursively cutting a domain. The flexibility of these methods is often limited by the requirement that the cuts be axis aligned. The Ostomachion process and the self-consistent binary space partitioning-tree process were recently introduced as generalizations of the Mondrian process for space partitioning with non-axis aligned cuts in the two dimensional plane. Motivated by the need for a multi-dimensional partitioning tree with non-axis aligned cuts, we propose the Random Tessellation Process (RTP), a framework that includes the Mondrian process and the binary space partitioning-tree process as special cases. We derive a sequential Monte Carlo algorithm for inference, and provide random forest methods. Our process is self-consistent and can relax axis-aligned constraints, allowing complex inter-dimensional dependence to be captured. We present a simulation study, and analyse gene expression data of brain tissue, showing improved accuracies over other methods. 1 Introduction Bayesian nonparametric models provide flexible and accurate priors by allowing the dimensionality of the parameter space to scale with dataset sizes [12]. The Mondrian process (MP) is a Bayesian nonparametric prior for space partitioning and provides a Bayesian view of decision trees and random forests [29, 19]. Inference for the MP is conducted by recursive and random axis-aligned cuts in the domain of the observed data, partitioning the space into a hierarchical tree of hyper-rectangles. The MP is appropriate for multi-dimensional data, and it is self-consistent (i.e., it is a projective distribution), meaning that the prior distribution it induces on a subset of a domain is equal to the marginalisation of the prior over the complement of that subset. Self-consistency is required in Bayesian nonparametric models in order to insure correct inference, and prevent any bias arising from sampling and sample population size. Recent advances in MP methods for Bayesian nonparametric space partitioning include online methods [21], and particle Gibbs inference for MP additive regression trees [22]. These methods achieve high predictive accuracy, with improved efficiency. However, the axis-aligned nature of the decision boundaries of the MP restricts its flexibility, which could lead to failure in capturing inter-dimensional dependencies in the domain. Recently, advances in Bayesian nonparametrics have been developed to allow more flexible non-axis aligned partitioning. The Ostomachion process (OP) was introduced to generalise the MP and allow non-axis aligned cuts. The OP is defined for two dimensional data domains [11]. In the OP, the angle 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. Figure 1: A draw from the u RTP prior with domain given by a four dimensional hypercube (x, y, z, w) [ 1, 1]4. Intersections of the draw and the three dimensional cube are shown for w= 1, 0, 1. Colours indicate polytope identity, and are randomly assigned. and position of each cut is randomly sampled from a specified distribution. However, the OP is not self-consistent, and so the binary space partitioning-tree (BSP) process [9] was introduced to modify the cut distribution of the OP in order to recover self-consistency. The main limitation of the OP and the BSP is that they are not defined for dimensions larger than two (i.e., they are restricted to data with two predictors). To relax this constraint, in [10] a self-consistent version of the BSP was extended to arbitrarily dimensioned space (called the BSP-forest). But for this process each cutting hyperplane is axis-aligned in all but two dimensions (with non-axis alignment allowed only in the remaining two dimensions, following the specification of the two dimensional BSP). Alternative constructions of non-axis aligned partitioning for two dimensional spaces and non-Bayesian methods involving sparse linear combinations of predictors or canonical correlation have also been proposed as random forest generalisations [14, 30, 28]. In this work, we propose the Random Tessellation Process (RTP), a framework for describing Bayesian nonparametric models based on cutting multi-dimensional Euclidean space. We consider four versions of the RTP, including a generalisation of the Mondrian process with non-axis aligned cuts (a sample from this prior is shown in Figure 1), a formulation of the Mondrian process as an RTP, and weighted versions of these two methods (shown in Figure 2). By virtue of their construction, all versions of the RTP are self-consistent, and are based on the theory of stable iterated tessellations in stochastic geometry [27]. The partitions induced by the RTP prior are described by a set of polytopes. We derive a sequential Monte Carlo (SMC) algorithm [8] for RTP inference which takes advantage of the hierarchical structure of the generating process for the polytope tree. We also propose a random forest version of RTPs, which we refer to as Random Tessellation Forests (RTFs). We apply our proposed model to simulated data and several gene expression datasets, and demonstrate its effectiveness compared to other modern machine learning methods. Suppose we observe a dataset (v1, z1), . . . , (vn, zn), for a classification task in which vi Rd are predictors and zi {1, . . . , K} are labels (with K levels, K N>1). Bayesian nonparametric models based on partitioning the predictors proceed by placing a prior on aspects of the partition, and associating likelihood parameters with the blocks of the partition. Inference is then done on the joint posterior of the parameters and the structure of the partition. In this section, we develop the RTP: a unifying framework that covers and extends such Bayesian nonparametric models through a prior on partitions of (v1, z1), . . . , (vn, zn) induced by tessellations. 2.1 The Random Tessellation Process A tessellation Y of a bounded domain W Rd is a finite collection of closed polytopes such that the union of the polytopes is all of W, and such that the polytopes have pairwise disjoint interiors [6]. We denote tessellations of W by Y(W) or the symbol . A polytope is an intersection of finitely many closed half-spaces. In this work we will assume that all polytopes are bounded and have nonempty interior. An RTP Yt(W) is a tessellation-valued right-continuous Markov jump process (MJP) defined on [0, τ] (we refer to the t-axis as time), in which events are cuts (specified by hyperplanes) of the tessellation s polytopes, and τ is a prespecified budget [2]. In this work we assume that all hyperplanes are affine (i.e., they need not pass through the origin). The initial tessellation Y0(W) contains a single polytope given by the convex hull of the observed predictors in the dataset: W = hull{v1, . . . , vn} (the operation hull A denotes the convex hull of the set A). In the MJP for the random tessellation process, each polytope has an exponentially distributed lifetime, and at the end of a polytope s lifetime, the polytope is replaced by two new polytopes. The two new polytopes are formed by drawing a hyperplane that intersects the interior of the old polytope, and then intersecting the old polytope with each of the two closed half-spaces bounded by the drawn hyperplane. We refer to this operation as cutting a polytope according to the hyperplane. These cutting events continue until the prespecified budget τ is reached. Let H be the set of hyperplanes in Rd. Every hyperplane h H can be written uniquely as the set of points {P : n, P u n = 0}, such that n Sd 1 is a normal vector of h, and u R 0 (u 0). Here Sd 1 is the unit (d 1)-sphere (i.e., Sd 1 = { n Rd : n = 1}). Thus, there is a bijection ϕ : Sd 1 R 0 7 H by ϕ( n, u) = {P : n, P u n = 0}, and therefore a measure Λ on H is induced by any measure Λ ϕ on Sd 1 R 0 through this bijection [20, 6]. In [27] Section 2.1, Nagel and Weiss describe a random tessellation associated with a measure Λ on H through a tessellation-valued MJP Yt such that the rate of the exponential distribution for the lifetime of a polytope a Yt is Λ([a]) (here and throughout this work, [a] denotes the set of hyperplanes in Rd that intersect the interior of a), and the hyperplane for the cutting event for a polytope a is sampled according to the probability measure Λ( [a])/Λ([a]). We use this construction as the prior for RTPs, and describe their generative process in Algorithm 1. This algorithm is equivalent to the first algorithm listed in [27]. Algorithm 1 Generative Process for RTPs 1: Inputs: a) Bounded domain W, b) RTP measure Λ on H, c) prespecified budget τ. 2: Outputs: A realisation of the Random Tessellation Process (Yt)0 t τ. 3: τ0 0. 4: Y0 {W}. 5: while τ0 τ do 6: Sample τ Exp P a Yτ0Λ([a]) . 7: Set Yt Yτ0 for all t (τ0, min{τ, τ0 + τ }]. 8: Set τ0 τ0 + τ . 9: if τ0 τ then 10: Sample a polytope a from the set Yτ0 with probability proportional to (w.p.p.t.) Λ([a]). 11: Sample a hyperplane h from [a] according to the probability measure Λ( [a])/Λ([a]). 12: Yτ0 (Yτ0/{a}) {a h , a h+}. (h and h+ are the h-bounded closed half planes.) 13: else 14: return the tessellation-valued right-continuous MJP sample (Yt)0 t τ. 2.1.1 Self-consistency of Random Tessellation Processes From Theorem 1 in [27], if the measure Λ is invariant with respect to translation (i.e., Λ(A) = Λ({h + x : h A}) for all measurable subsets A H and x Rd), and if a set of d hyperplanes with orthogonal normal vectors is contained in the support of Λ, then for all bounded domains W W, Yt(W ) is equal in distribution to Yt(W) W . This means that self-consistency holds for the random tessellations associated with such Λ. (Here, for a hyperplane h, h + x refers to the set {y + x : y h}, and for a tessellation Y and a domain W , Y W is the tessellation {a W : a Y }.) In [27], such tessellations are referred to as stable iterated tessellations. If we assume that Λ ϕ is the product measure λd λ+, with λd symmetric (i.e., λd(A) = λd( A) for all measurable sets A Sd 1) and further that λ+ is given by the Lebesgue measure on R 0, then Λ is translation invariant (a proof of this statement is given in Appendix A, Lemma 1 of the Supplementary Material). So, through Algorithm 1 and Theorem 1 in [27], any distribution λd on the sphere Sd 1 that is supported on a set of d hyperplanes with orthogonal normal vectors gives rise to a self-consistent random tessellation, and we refer to models based on this self-consistent prior as Random Tessellation Processes (RTPs). We refer to any product measure Λ ϕ = λd λ+ such these conditions hold (i.e., λd symmetric, Λ supported on d orthogonal hyperplanes and λ+ given by the Lebesgue measure) as an RTP measure. 2.1.2 Relationship to cutting nonparametric models Figure 2: Draws from priors Left) wu RTP, and Right) w MRTP, for the domain given by the rectangle W = [ 1, 1] [ 1/3, 1/3]. Weights are given by ωx = 14, ωy = 1, leading to horizontal (x-axis heavy) structure in the polygons. Colours are randomly assigned and indicate polygon identity, and black delineates polygon boundaries. If λd is the measure associated with a uniform distribution on the sphere (with respect to the usual Borel sets on Sd 1 [18]), then the resulting RTP is a generalisation of the Mondrian process with nonaxis aligned cuts. We refer to this RTP as the u RTP (for uniform RTP). In this case, λd is a probability measure and a normal vector n may be sampled according to λd by sampling ni N(0, 1) and then setting ni= ni/ n . A draw from the u RTP prior supported on a four dimensional hypercube is displayed in Figure 1. We consider a weighted version of the uniform RTP found by setting λd to the measure associated with the distribution on n induced by the scheme ni N(0, ω2 i ), ni= ni/ n . We refer to this RTP as the wu RTP (weighted uniform RTP), and the wu RTP is parameterised by d weights ωi R>0. Note that the isotropy of the multivariate Gaussian n implies symmetry of λd. Setting the weight ωi increases the prior probability of cuts orthogonal to the i-th predictor dimension, allowing prior information about the importance of each of the predictors to be incorporated. Figure 2(left) shows a draw from the wu RTP prior supported on a rectangle. The Mondrian process itself is an RTP with λd =P v poles(d) δv. Here, δx is the Dirac delta supported on x, and poles(d) is the set of normal vectors with zeros in all coordinates, except for one of the coordinates (the non-zero coordinate of these vectors can be either 1 or +1). We refer to this view of the Mondrian process as the MRTP. If λd =P v poles(d) ωi(v)δv, where ωi are d axis weights, and i(v) is the index of the nonzero element of v, then we arrive at a weighted version of the MRTP, which we refer to as the w MRTP. Figure 2(right) displays a draw from the w MRTP prior. The horizontal organisation of the lines in Figure 2 arise from the uneven axis weights. Other nonparametric models based on cutting polytopes may also be viewed in this way. For example, Binary Space Partitioning-Tree Processes [9] are u RTPs and wu RTPs restricted to two dimensions, and the generalization of the Binary Space Partitioning-Tree Process in [10] is an RTP for which λd is a sum of delta functions convolved with smooth measures on S1 projected onto pairs of axes. The Ostomachion process [11] does not arise from an RTP: it is not self-consistent, and so by Theorem 1 in [27], the OP cannot arise from an RTP measure. 2.1.3 Likelihoods for Random Tessellation Processes In this section, we illustrate how RTPs can be used to model categorical data. For example, in gene expression data of tissues, the predictors are the amounts of expression of each gene in the tissue (the vector vi for sample i), and our goal is to predict disease condition (labels zi). Let t be an RTP on the domain W = hull{v1, . . . , vn}. Let Jt denote the number of polytopes in t. We let h(vi) denote a mapping function, which matches the i-th data item to the polytope in the tessellation containing that data item. Hence, h(vi) takes a value in the set {1, . . . , Jt}. We will consider the likelihood arising at time t from the following generative process: t Y t(W ), φj Dirichlet(α) for 1 j Jt, (1) zi| t, φh(vi) Multinomial(φh(vi)) for 1 i n. (2) Here φj = (φj1, . . . , φj K) are parameters of the multinomial distribution with a Dirichlet prior with hyperparameters α = (αk)1 k K. The likelihood function for Z = (zi)1 i n conditioned on the tessellation t and V = (vi)1 i n and given the hyperparameter α is as follows: P(Z| t, V, α) = Z Z P(Z, φ| t, α)dφ1 dφJt = Here B( ) is the multivariate beta function, mj = (mjk)1 k K and mjk = P i:h(vi)=j δ(zi = k), and δ( ) is an indicator function with δ(zi = k) = 1 if zi = k and δ(zi = k) = 0 otherwise. We refer to Appendix A, Lemma 2 of the Supplementary Material for the derivation of (3). 2.2 Inference for Random Tessellation Processes Our objective is to infer the posterior of the tessellation at time t, denoted π( t|V, Z, α). We let π0( t) denote the prior distribution of t. Let P(Z| t, V, α) denote the likelihood (3) of the data given the tessellation t and the hyperparameter α. By Bayes rule, the posterior of the tessellation at time t is π( t|V, Z, α) = π0( t)P(Z| t, V, α)/P(Z|V, α). Here P(Z|V, α) is the marginal likelihood given data Z. This posterior distribution is intractable, and so in Section 2.2.3 we introduce an efficient SMC algorithm for conducting inference on π( t). The proposal distribution for this SMC algorithm involves draws from the RTP prior, and so in Section 2.2.1, we describe a rejection sampling scheme for drawing a hyperplane from the probability measure Λ( [a])/Λ([a]) for a polytope a. We provide some optimizations and approximations used in this SMC algorithm in Section 2.2.2. 2.2.1 Sampling cutting hyperplanes for Random Tessellation Processes Suppose that a is a polytope, and Λ ϕ is an RTP measure such that Λ(ϕ( )) = (λd λ+)( ). We wish to sample a hyperplane according to the probability measure Λ( [a])/Λ([a]). We note that if Br(x) is the smallest closed d-ball containing a (with radius r and centre x), then [a] is contained in [Br(x)]. If we can sample a normal vector n according to λd, then we can sample a hyperplane according to Λ( [a])/Λ([a]) through the following rejection sampling scheme. Step 1) Sample n according to λd. Step 2) Sample u Uniform[0, r]. Step 3) If the hyperplane h = x + {P : n, P u n } intersects a, then RETURN h. Otherwise, GOTO Step 1). Note that in Step 3, the set {P : n, P u n } is a hyperplane intersecting the ball Br(0) centred at the origin, and so translation of this hyperplane by x yields a hyperplane intersecting the ball Br(x). When this scheme is applied to the u RTP or wu RTP, in Step 1 n is sampled from the uniform distribution on the sphere or the appropriate isotropic Gaussian. And for the MRTP or w MRTP, n is sampled from the discrete distributions given in Section 2.1.2. 2.2.2 Optimizations and approximations We use three methods to decrease the computational requirements and complexity of inference based on RTP posteriors. First, we replace all polytopes with convex hulls formed by intersecting the polytopes with the dataset predictors. Second, in determining the rates of the lifetimes of polytopes, we approximate Λ([a]) with the measure Λ([ ]) applied to the smallest closed ball containing a. Third, we use a pausing condition so that no cuts are proposed for polytopes for which the labels of all predictors in that polytope are the same label. Convex hull replacement. In our posterior inference, if a Y is cut according to the hyperplane h, we consider the resulting tessellation to be Y/{a} {hull(a V h+), hull(a V h )}. Here h+ and h are the two closed half planes bounded by h, and / is the set minus operation. This requires a slight loosening of the definition of a tessellation Y of a bounded domain W to allow the union of the polytopes of a tessellation Y to be a strict subset of W such that V a Y a. In our computations, we do not need to explicitly compute these convex hulls, and instead for any polytope b, we store only b V , as this is enough to determine whether or not a hyperplane h intersects hull(b V ). This membership check is the only geometric operation required to sample hyperplanes intersecting b according to the rejection sampling scheme from Section 2.2.1. By the self-consistency of RTPs, this has the effect of marginalizing out MJP events involving cuts that do not further separate the predictors in the dataset. This also obviates the need for explicit computation of the facets of polytopes, significantly simplifying the codebase of our implementation of inference. After this convex hull replacement operation, a data item in the test dataset may not be contained in any polytope, and so to conduct posterior inference we augment the training dataset with a version of the testing dataset in which the label is missing, and then marginalise the missing label in the likelihood described in Section 2.1.3. Spherical approximation. Every hyperplane intersecting a polytope a also intersects a closed ball containing a. Therefore, for any RTP measure Λ, Λ([a]) is upper bounded by Λ([B(ra)]). Here ra is the radius of the smallest closed ball containing a. We approximate Λ([a]) Λ([B(ra)]) for use in polytope lifetime calculations in our u RTP inference and we do not compute Λ([a]) exactly. For the u RTP and w RTP, Λ([B(ra)]) = ra. A proof of this is given in Appendix A, Lemma 3 of the Supplementary Material. For the MRTP and w MRTP, Λ([a]) can be computed exactly [29]. Pausing condition. In our posterior inference, if zi = zj for all i, j such that vi, vj a, then we pause the polytope a and no further cuts are performed on this polytope. This improves computational efficiency without affecting inference, as cutting such a polytope cannot further separate labels. This was done in recent work for Mondrian processes [21] and was originally suggested in the Random Forest reference implementation [4]. 2.2.3 Sequential Monte Carlo for Random Tessellation Process inference Algorithm 2 SMC for inferring RTP posteriors 1: Inputs: a) Training dataset V , Z, b) RTP measure Λ on H, c) prespecified budget τ, d) likelihood hyperparameter α. 2: Outputs: Approximate RTP posterior PM m=1ϖmδ τ,mat time τ. (ϖm are particle weights.) 3: Set τm 0, for m = 1, . . . , M. 4: Set 0,m {hull V }, ϖm 1/M, for m = 1, . . . , M. 5: while min{τm}M m=1 < τ do 6: Resample τm,m from { τm,m}M m=1 w.p.p.t. {ϖm}M m=1, for m = 1, . . . , M. 7: Set τm,m τm,m, for m = 1, . . . , M. 8: Set ϖm 1/M, for m = 1, . . . , M. 9: for m {m : m = 1, . . . , M and τm < τ} do 10: Sample τ Exp P a τm,m ra . (ra is the radius of the smallest closed ball containing a.) 11: Set t,m τm,m, for all t (τm, min{τ, τm + τ }]. 12: if τm + τ τ then 13: Sample a from the set τm,m w.p.p.t. ra. 14: Sample h from [a] according to Λ( [a])/Λ([a]) using Section 2.2.1. 15: Set τm,m ( τm,m/{a}) {hull(V a h ), hull(V a h+)}. 16: Set ϖm ϖm P(Z| τm,m, V , α)/P(Z| τm,m, V , α) according to (3). 17: else 18: Set t,m τm,m, for t (τm, τ]. 19: Set τm τm + τ . 20: Set Z PM m=1 ϖm. 21: Set ϖm ϖm/Z, for m = 1, . . . , M. 22: return the particle approximation PM m=1 ϖmδ We provide an SMC method (Algorithm 2) with M particles, to approximate the posterior distribution π( t) of an RTP conditioned on Z, V , and given an RTP measure Λ and a hyperparameter α and a prespecified budget τ. Our algorithm iterates between three steps: resampling particles (Algorithm 2, line 7), propagation of particles (Algorithm 2, lines 13-15) and weighting of particles (Algorithm 2, line 16). At each SMC iteration, we sample the next MJP events using the spherical approximation of Λ([ ]) described in Section 2.2.2. For brevity, the pausing condition described in Section 2.2.2 is omitted from Algorithm 2. In our experiments, after using Algorithm 2 to yield a posterior estimate PM m=1 ϖmδ τ,m, we select the tessellation τ,m with the largest weight ϖm (i.e., we do not conduct resampling at the last SMC iteration). We then compute posterior probabilities of the test dataset labels using the particle τ,m. This method of not resampling after the last SMC iteration is recommended in [7] for lowering asymptotic variance in SMC estimates. The computational complexity of Algorithm 2 depends on the number of polytopes in the tessellations, and the organization of the labels within the polytopes. The more linearly separable the dataset is, the sooner the pausing conditions are met. The complexity of computing the spherical approximation in Section 2.2.2 (the radius ra) for a polytope a is O(|V a|2), where | | denotes set cardinality. 2.2.4 Prediction with Random Tessellation Forests Random forests are commonly used in machine learning for classification and regression problems [3]. A random forest is represented by an ensemble of decision trees, and predictions of test dataset labels are combined over all decision trees in the forest. To improve the performance of our methods, we consider random forest versions of RTPs (which we refer to as RTFs: u RTF, wu RTF, MRTF, w MRTF are random forest versions of the u RTP, MRTP and their weighted versions resp.). We run Algorithm 2 independently T times, and predict labels using the modes. Differing from [3], we do not use bagging. In [22], Lakshminarayanan, Roy and Teh consider an efficient Mondrian forest in which likelihoods are dropped from the SMC sampler and cutting is done independent of likelihood. This method follows recent theory for random forests [15]. We consider this method (by dropping line 16 of Algorithm 2) and refer to the implementations of this method as the u RTF.i and MRTF.i (i for likelihood independence). 3 Experiments # cuts 0 100 200 60 70 80 90 100 u RTP MRTP LR DT SVM Figure 3: Left) A view of the Mondrian cube, with cyan indicating label 1, magenta indicating label 2, and black delineating label boundaries. Right) Percent correct versus number of cuts for predicting Mondrian cube test dataset, with u RTP, MRTP and a variety of baseline methods. In Section 3.1, we explore a simulation study that shows differences among u RTP and MRTP, and some standard machine learning methods. Variations in gene expression across tissues in brain regions play an important role in disease conditions. In Section 3.2, we examine predictions of a variety of RTF models for gene expression data. For all our experiments, we set the likelihood hyperparameters for the RTPs and RTFs to the empirical estimates αk to nk/1000. Here nk = Pn i=1 δ(zi = k). In all of our experiments, for each train/test split, we allocate 60% of the data items at random to the training set. An implementation of our methods (released under the open source BSD 2-clause license) and a software manual are provided in the Supplementary Material. 3.1 Simulations on the Mondrian cube We consider a simulated three dimensional dataset designed to exemplify the difference between axis-aligned and non-axis aligned models. We refer to this dataset as the Mondrian cube, and we investigate the performance of u RTP and the MRTP on this dataset, along with some standard machine learning approaches, varying the number of cuts in the processes. The Mondrian cube dataset is simulated as follows: first, we sample 10,000 points uniformly in the cube [0, 1]3. Points falling in the cube [0, 0.25]3 or the cube [0.25, 1]3 are given label 1, and the remaining points are given label 2. Then, we centre the points and rotate all of the points by the angles π 4 about the x-axis and y-axis respectively, creating a dataset organised on diagonals. In Figure 3(left), we display a visualization of the Mondrian cube dataset, wherein points are colored by their label. We apply the SMC algorithm to the Mondrian cube data, with 50 random train/test splits. For each split, we run 10 independent copies of the u RTP and MRTP and take the mode of their results, and we also examine the accuracy of logistic regression (LR), a decision tree (DT) and a support vector machine (SVM). Figure 3(right) shows that the percent correct for the u RTP and MRTP both increase as the number of cuts increases, and plateaus when the number of cuts becomes larger (greater than 25). Even though the u RTP has lower accuracy at the first cut, it starts dominating the MRTP after the second cut. Overall, in terms of percent correct, with any number of cuts > 105, a sign test indicates that the u RTP performs significant better than all other methods at nominal significance, and the MRTP performs significant better than DT and LR for any number of cuts > 85 at nominal significance. 3.2 Experiment on gene expression data in brain tissue LR SV RF MRTF.i u RTF.i MRTF u RTF w MRTF wu RTF Figure 4: Box plot showing wu RTF and w MRTF improvements for GL85, and generally best performance for wu RTF method (with sign test p-value of 3.2 10 9 vs w MRTF). Reduced performance of SVM indicates structure in GL85 that is not linearly separable. Medians, quantiles and outliers beyond 99.3% coverage are indicated. We evaluate a variety of RTPs and some standard machine learning methods on a glioblastoma tissue dataset GSE83294 [13], which includes 22,283 gene expression profiles for 85 astrocytomas (26 diagnosed as grade III and 59 as grade IV). We also examine schizophrenia brain tissue datasets: GSE21935 [1], in which 54,675 gene expression in the superior temporal cortex is recorded for 42 subjects, with 23 cases (with schizophrenia), and 19 controls (without schizophrenia), and dataset GSE17612 [24], a collection of 54,675 gene expressions from samples in the anterior prefrontal cortex (i.e., a different brain area from GSE21935) with 28 schizophrenic subjects, and 23 controls. We refer to these datasets as GL85, SCZ42 and SCZ51, respectively. We also consider a combined version of SCZ42 and SCZ51 (in which all samples are concatenated), which we refer to as SCZ93. For GL85 the labels are the astrocytoma grade, and for SCZ42, SCZ51 and SCZ93 the labels are schizophrenia status. We use principal components analyais (PCA) in preprocessing to replace the predictors of each data item (a set of gene expressions) with its scores on a full set of principal components (PCs): i.e., 85 PCs for GL85, 42 PCs in SCZ42 and 51 PCs in SCZ51. We then scale the PCs. We consider 200 test/train splits for each dataset. These datasets were acquired from NCBI s Gene Expression Omnibus1 and were are released under the Open Data Commons Open Database License. We provide test/train splits of the PCA preprocessed datasets in the Supplementary Material. Through this preprocessing, the j-th predictor is the score vector of the j-th principal component. For the weighted RTFs (the wu RTF and w MRTF), we set the weight of the j-th predictor to be proportional to the variance explained by the j-th PC (σ2 j ): ωj = σ2 j . We set the number of trees in 1Downloaded from https://www.ncbi.nlm.nih.gov/geo/ in Spring 2019. Dataset BL LR SVM RF MRTF.i u RTF.i MRTF u RTF w MRTF wu RTF GL85 70.34 58.13 70.34 73.01 70.74 70.06 77.09 70.60 80.57 84.90 SCZ42 46.68 57.65 46.79 51.76 49.56 48.50 49.91 47.71 53.12 53.97 SCZ51 46.55 51.15 46.67 57.38 52.55 48.58 57.95 44.70 58.12 49.05 SCZ93 48.95 53.05 50.15 52.45 50.23 50.24 51.80 50.34 53.12 54.99 Table 1: Comparison of mean percent correct on gene expression datasets over 200 random train/test splits across different methods. Nominal statistical significance (p-value < 0.05) is ascertained by a sign test in which ties are broken in a conservative manner. Tests are performed between the top method and each other method. Bold values indicate the top method and all methods statistically indistinguishable from the top method according to nominal significance. Largest nominally significant improvement is seen for wu RTF on GL85, and wu RTF is significantly better than other methods for this dataset. The w MRTF and wu RTF have largest mean percent correct for SCZ51 and SCZ93 but are not statistically distinguishable from RF or LR for those datasets. all of the random forests to 100, which is the default in R s random Forest package [23]. For the all RTFs, we set the budget τ = , as is done in [21]. We compare percent correct for the wu RTF, u RTF, u RTF.i, and the Mondrian Random Tessellation Forests w MRTF, MRTF and MRTF.i, a random forest (RF), logistic regression (LR), a support vector machine (SVM) and a baseline (BL) in which the mode of the training set label is always predicted [25, 23]. Mean percent correct and sign tests for all of these experiments are reported in Table 1 and box plots for the GL85 experiment reported in Figure 4. We observe that the increase in accuracy of wu RTFs achieves nominal significance over other methods on GL85. For datasets SCZ42, SCZ51 and SCZ93, the performance of the RTFs is comparable to that of logistic regression and random forests. For all datasets we consider, RTFs have higher accuracy than SVMs (with nominal significance). Boxplots with accuracies for the datasets SCZ42, SCZ51 and SCZ93 are provided in Appendix B, Supplementary Figure 1 of the Supplementary Material. Results of a conservative pairwise sign test performed between each pair of methods on each dataset, standard deviation of the percent correct, and mean runtime across different methods for all these four datasets are reported in Supplementary Tables 1, 2, and 3 in Appendix B of the Supplementary Material. 5 Discussion The spherical approximation introduced in Section 2.2.2 can lead to inexact inference. In Supplementary Algorithm 1, we introduce a new algorithm based on Poisson thinning that recovers exact inference. This algorithm may improve accuracy and allow hierarchical likelihoods. There are many directions for future work in RTPs including improved SMC sampling for MJPs as in [17], hierarchical likelihoods and online methods as in [21], analysis of minimax convergence rates [26], and extensions using Bayesian additive regression trees [5, 22]. We could also consider applications of RTPs to data that naturally displays tessellation and cracking, such as sea ice [16]. 6 Conclusion We have described a framework for viewing Bayesian nonparametric methods based on space partitioning as Random Tessellation Processes. This framework includes the Mondrian process as a special case, and includes extensions of the Mondrian process allowing non-axis aligned cuts in high dimensional space. The processes are self-consistent, and we derive inference using sequential Monte Carlo and random forests. To our knowledge, this is the first work to provide self-consistent Bayesian nonparametric hierarchical partitioning with non-axis aligned cuts that is defined for more than two dimensions. As demonstrated by our simulation study and experiments on gene expression data, these non-axis aligned cuts can improve performance over the Mondrian process and other machine learning methods such as support vector machines, and random forests. Acknowledgments We are grateful to Kevin Sharp, Frauke Harms, Maasa Kawamura, Ruth Van Gurp, Lars Buesing, Tom Loughin and Hugh Chipman for helpful discussion, comments and inspiration. We would also like to thank Fred Popowich and Martin Siegert for help with computational resources at Simon Fraser University. YWT s research leading to these results has received funding from the European Research Council under the European Union s Seventh Framework Programme (FP7/2007-2013) ERC grant agreement no. 617071. This research was also funded by NSERC grant numbers RGPIN/05484-2019, DGECR/00118-2019 and RGPIN/06131-2019. [1] M. R. Barnes, J. Huxley-Jones, P. R Maycox, M. Lennon, A. Thornber, F. Kelly, S. Bates, A. Taylor, J. Reid, N. Jones, and J. Schroeder. Transcription and pathway analysis of the superior temporal cortex and anterior prefrontal cortex in schizophrenia. Journal of Neuroscience Research, 89(8), 2011. [2] M. A. Berger. An Introduction to Probability and Stochastic Processes. Springer Texts in Statistics, 2012. [3] L. Breiman. Random forests. Machine Learning, 45(1), 2001. [4] L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone. Classification and Regression Trees. Chapman and Hall/CRC, 1984. [5] H. A. Chipman, E. I. George, and R. E. Mc Culloch. BART: Bayesian additive regression trees. Annals of Applied Statistics, 4(1), 2010. [6] S. N. Chiu, D. Stoyan, W. S. Kendall, and J. Mecke. Stochastic Geometry and its Applications. Wiley Series in Probability and Statistics, 2013. [7] N. Chopin. Central limit theorem for sequential Monte Carlo methods and its application to Bayesian inference. The Annals of Statistics, 32(6), 2004. [8] A. Doucet, S. Godsill, and C. Andrieu. On sequential Monte Carlo sampling methods for Bayesian filtering. Statistics and Computing, 10(3), 2000. [9] X. Fan, B. Li, and S. Sisson. The binary space partitioning-tree process. In Proceedings of the 35th International Conference on Artificial Intelligence and Statistics, 2018. [10] X. Fan, B. Li, and S. Sisson. Binary space partitioning forests. ar Xiv preprint 1903.09348, 2019. [11] X. Fan, B. Li, Y. Wang, Y. Wang, and F. Chen. The Ostomachion process. In Proceedings of the Thirtieth Conference of the Association for the Advancement of Artificial Intelligence, 2016. [12] T. S. Ferguson. A Bayesian analysis of some nonparametric problems. The Annals of Statistics, 1(2), 1973. [13] W. A. Freije, F. E. Castro-Vargas, Z. Fang, S. Horvath, T. Cloughesy, L. M. Liau, P. S. Mischel, and S. F. Nelson. Gene expression profiling of gliomas strongly predicts survival. Cancer Research, 64(18), 2004. [14] E. I. George. Sampling random polygons. Journal of applied probability, 24(3):557 573, 1987. [15] P. Geurts, D. Ernst, and L. Wehenkel. Extremely randomized trees. Machine Learning, 63(1), 2006. [16] D. Godlovitch. Idealised models of sea ice thickness dynamics. Ph D thesis, University of Victoria, 2011. [17] M. Hajiaghayi, B. Kirkpatrick, L. Wang, and A. Bouchard-Côté. Efficient continuous-time Markov chain estimation. In Proceedings of the 31st International Conference on Machine Learning, 2014. [18] P. Halmos. Measure Theory. Springer, 1974. [19] C. Kemp, J. B. Tenenbaum, T. L. Griffiths, T. Yamada, and N. Ueda. Learning systems of concepts with an infinite relational model. In Proceedings of the 20th Conference on the Association for the Advancement of Artificial Intelligence, 2006. [20] J. F. Kingman. Poisson Processes. Oxford University Press, 1996. [21] B. Lakshminarayanan, D. M. Roy, and Y. W. Teh. Mondrian forests: Efficient online random forests. In Proceedings of the 28th Conference on Neural Information Processing Systems, 2014. [22] B. Lakshminarayanan, D. M. Roy, and Y. W. Teh. Particle Gibbs for Bayesian additive regression trees. In Proceedings of the 18th International Conference on Artificial Intelligence and Statistics, 2015. [23] A. Liaw and M. Wiener. Classification and Regression by random Forest. R News, 2(3), 2002. [24] P. R. Maycox, F. Kelly, A. Taylor, S. Bates, J. Reid, R. Logendra, M. R. Barnes, C. Larminie, N. Jones, M. Lennon, C. Davies, J. J. Hagan, C. A. Scorer, C. Angelinetta, M. T. Akbar, S. Hirsch, A. M. Mortimer, T. R. Barnes, and J. de Belleroche. Analysis of gene expression in two large schizophrenia cohorts identifies multiple changes associated with nerve terminal function. Molecular Psychiatry, 14(12), 2009. [25] D. Meyer, E. Dimitriadou, K. Hornik, A. Weingessel, and F. Leisch. e1071: Misc Functions of the Department of Statistics, Probability Theory Group, Technische Universität Wien, 2019. [26] Jaouad Mourtada, Stéphane Gaïffas, and Erwan Scornet. Minimax optimal rates for mondrian trees and forests. ar Xiv preprint 1803.05784, 2018. [27] W. Nagel and V. Weiss. Crack STIT tessellations: Characterization of stationary random tessellations stable with respect to iteration. Advances in Applied Probability, 37(4), 2005. [28] T. Rainforth and F. Wood. Canonical correlation forests. ar Xiv preprint 1507.05444, 2015. [29] D. M. Roy and Y. W. Teh. The Mondrian process. In Proceedings of the 22nd Conference on Neural Information Processing Systems, 2008. [30] T. M. Tomita, J. Browne, C. Shen, J. L. Patsolic, J. Yim, C. E. Priebe, R. Burns, M. Maggioni, and J. T. Vogelstein. Random projection forests. ar Xiv preprint 1506.03410, 2015.