# on_the_effectiveness_of_persistent_homology__34889c5f.pdf On the Effectiveness of Persistent Homology Renata Turkeš University of Antwerp renata.turkes@uantwerpen.be Guido Montúfar University of California, Los Angeles montufar@math.ucla.edu Nina Otter Queen Mary University of London n.otter@qmul.ac.uk Persistent homology (PH) is one of the most popular methods in Topological Data Analysis. Even though PH has been used in many different types of applications, the reasons behind its success remain elusive; in particular, it is not known for which classes of problems it is most effective, or to what extent it can detect geometric or topological features. The goal of this work is to identify some types of problems where PH performs well or even better than other methods in data analysis. We consider three fundamental shape analysis tasks: the detection of the number of holes, curvature and convexity from 2D and 3D point clouds sampled from shapes. Experiments demonstrate that PH is successful in these tasks, outperforming several baselines, including Point Net, an architecture inspired precisely by the properties of point clouds. In addition, we observe that PH remains effective for limited computational resources and limited training data, as well as out-of-distribution test data, including various data transformations and noise. For convexity detection, we provide a theoretical guarantee that PH is effective for this task in Rd, and demonstrate the detection of a convexity measure on the FLAVIA data set of plant leaf images. Due to the crucial role of shape classification in understanding mathematical and physical structures and objects, and in many applications, the findings of this work will provide some knowledge about the types of problems that are appropriate for PH, so that it can to borrow the words from Wigner 1960 remain valid in future research, and extend, to our pleasure", but to our lesser bafflement, to a variety of applications. 1 Introduction Persistent homology (PH) is an extension of homology, which gives a way to capture topological information about connectivity and holes in a geometric object. PH can be regarded as a framework to compute representations of raw data that can be used for further processing or as inputs to learning algorithms. There have been numerous successful applications of PH in the last decade, from prediction of biomolecular properties [18,19,110], face, gait and activity recognition [58,66,70,122] or digital forensics [6], to discriminating breast-cancer subtypes [97], quantifying the porosity of nanoporous materials [69], classifying fingerprints [50], or studying the morphology of leaves [71].1 At the same time, the reasons behind these successes are not yet well understood. Indeed, the data used in real-world applications is complex, so that there are numerous effects at play and one is often left unsure why PH worked, i.e., what type of topological or geometric information it captured that facilitated the good performance. 1A database of applications of persistent homology is being maintained at [51]. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). The title of our manuscript is inspired by a famous paper from 1960, The unreasonable effectiveness of mathematics in the natural sciences [112], in which Wigner discusses, with wonder, how mathematical concepts have applicability far beyond the context in which they were originally developed. The same, we believe, is true for persistent homology. While this method has been applied successfully to a wide range of application problems, we believe that for PH to remain relevant, there is a need to better understand why it is so successful. Thus, we distinguish between the usefulness of PH for applications, which has been attested in hundreds of applications and publications, and its effectiveness, namely that PH is capable of producing an intended or desired result. Thus, here we initiate an investigation into the effectiveness of PH, or in other words, we investigate what is seen by persistent homology: Given a data set, i.e., a point cloud, which underlying topological and geometric features can we detect with PH? This question is related to manifold learning and, specifically, topological and geometric inference: Given a finite point cloud X of (noisy) samples from an unknown manifold M, how can one infer properties of M [11, 12, 25, 29]? Obtaining a representation of a shape that can be used in statistical models is an important task in data analysis and numerous approaches to modeling surfaces and shapes [104]. To pursue our investigation, we set out to identify some fundamental data-analysis tasks that can be solved with PH. Since PH is inspired by homology, which provides a measure for the number of components, holes, voids, and higher-dimensional cycles of a space to which we collectively refer as topological features , we start with the obvious question of whether PH applied to a point cloud sampled from a geometric object can detect the number of (1-dimensional) holes of the underlying object. Unlike homology, however, PH registers also the persistence of topological features across scales, and can thereby capture geometric information, such as size or position of holes. We therefore also investigate how well PH can detect fundamental geometric notions of curvature and convexity. For each of the three problems, we first discuss theoretical results that provide a guarantee that PH can solve these tasks. Detection of convexity with PH has not been investigated in the literature to date, and we prove a new result. To investigate how well the PH pipeline works in practice, we compare its performance against several baselines on synthetic point-cloud data sets. As a first machine learning (ML) baseline we take an SVM trained on the distance matrices of point clouds. We further consider fully-connected neural networks (NN) with a single or multiple hidden layers, also trained on distance matrices. As a stronger baseline we consider a Point Net trained on the point clouds directly. Point Net [1,88] is designed specifically for point cloud data. Similar architectures with convolutional (and fully-connected and pooling) layers have been applied for Betti-number and curvature estimation [52,81]. For convexity detection, we also evaluate the performance of PH on real-world data. The theoretical guarantees above imply that the results for PH would generalize to new data. Finally, we note that our goal is not to claim the superiority of PH compared to other approaches in the literature, in particular, with the state-of-the-art methods for each of the problems. We do not necessarily expect that on well-specified mathematical problems PH will beat state-of-the-art algorithms that have been specifically designed for those tasks. Instead, what we think is interesting and remarkable is that PH can in fact solve tasks it is not specifically or uniquely designed for. Moreover, an advantage of PH is that it can reveal, e.g., both topology and curvature at the same time, avoiding the need to employ and combine state-of-the-art models for each of the tasks. Related work In spite of the growing interest in PH, so far there is only limited work in the direction that we pursue here. There is indeed theoretical evidence that the number of holes of the underlying space can be detected from PH (under some conditions about the target space, the sample density and closeness to the space) [32,63,77], and there is significant interest in investigating how well this works in practice [30]. However, so far there are only few available results. Some works demonstrate that PH can be used to detect the number of holes, but only on individual toy examples (e.g. [90], [25, Figure 19], [65, Figures 9-20], [29, Figures 2, 3, 6, 11, 12]) without looking into the statistical significance between different classes of data, or the accuracy of some classification algorithms on a comprehensive data set. There are also some works where PH is used to estimate the Betti numbers on a possibly larger data set, but only with the goal of using this information to e.g., study the behavior of deep neural networks [76] or ensure topologically correct dimensionality reduction [82] or image segmentation [56], so that the soundness of this estimation is not investigated, which is the focus of our work. Some insights about PH and curvature have been obtained in the literature, starting with an illustrative example in [39, Figure 12] which shows that PH on the filtered tangent complex can distinguish between letters (C and I) that have the same topology, since their curvature is different. Recently, [17] show both theoretically and experimentally that PH can predict curvature (with computational experiments replicated in [107]), which inspired us to investigate this problem in more detail. Regarding the important geometric problem of classification between convex and concave shapes, we were not able to identify any previous works investigating the applicability of PH to this task. Some further recent work investigating the topological and geometric features seen by PH are the following. Bubenik and Dłotko [16] show that using PH of points sampled from spheres one can determine the dimension of the underlying spheres. A connection has also been established between PH and the magnitude of a metric space (an isometric invariant) [79]. There have been several efforts in using PH to estimate fractal dimensions, such as [95] in which Schweinhart proves that the fractal dimension of some metric spaces can be recovered from the PH of random samples. Main contributions Our contributions can be summarized as follows. We prove that PH can detect convexity in Rd (Theorem 1). We define a new tubular filtration function (medium through which PH is extracted from data), that is crucial for the detection of convexity (Definition 1). We demonstrate experimentally that PH can detect the number of holes (Section 3), curvature (Section 4), and convexity (Section 5) from synthetic point clouds in R2 or R3, outperforming SVMs and fully-connected networks trained on distance matrices, and Point Net trained on point clouds. For convexity detection, we also show that PH obtains a good performance on a real-world data set of plant leaf images. We demonstrate experimentally that PH features allow to solve the above tasks even in the case of limited training data (Section 3), noisy (Section 3) and out-of-distribution (Section 5) test data, and limited computational resources (Section 3, Section 4, Section 5). We provide insights about the topological and geometric features that are captured with long and short persistence intervals (Section 6), and formulate guidelines for applications that are suitable for PH (Section 7). We provide data sets that can be directly used as a benchmark for our tasks or other related pointcloud-analysis or classification problems. We provide computer code to construct more data and replicate our experiments. 2 Background on persistent homology Homology is a topological concept that attempts to distinguish between topological spaces by constructing algebraic invariants that reflect their connectivity properties [90], i.e., k-dimensional cycles (components, holes, voids, ...). The number of independent k-dimensional cycles is called k-th Betti number and denoted by βk. For example, the circle has Betti numbers β0 = 1, β1 = 1, β2 = 0, and for a torus, we have β0 = 1, β1 = 2, β2 = 1. Persistent homology is an extension of this idea [123] that has found success in applications to data. To calculate PH from some data X, we must first build a filtration, i.e., a family of nested topological spaces {Kr}r R which, in a suitable sense, approximate X at different scales r R. Typically, X = {x1, x2, . . . , xn} is a point cloud in Rd, and Kr is a simplicial complex, a set of simplices σ (which we can think of as vertices, edges, triangles, ...) such that if σ Kr and τ σ, then τ Kr [39]. A common choice is Kr = V R(X, r), where V R(X, r) is the Vietoris-Rips simplicial complex, in which σ = {x1, . . . , xm} V R(X, r) when dist(xi, xj) r for all 1 i, j m. PH can then be summarized with a persistence diagram (PD), a scatter plot with the x and y axes respectively depicting the scale r R at which each cycle is born and dies or is identified (i.e., merges) with another cycle within a filtration. The length l = d b of a persistence interval (b, d) measures the lifespan the so-called persistence of the corresponding cycle in the filtration. Instead of working directly with persistence diagrams, these are often represented by other signatures that are better suited for machine learning frameworks. A common choice is a persistence image (PI) [2] (discretized sum of Gaussian kernels centered at the PD points), or a persistence landscape (PL) (functions obtained by stacking isosceles triangles" above persistence intervals, with height reflecting their lifespan) [15]. The steps for extracting PH features are visualized in Appendix B.2. For good choices of filtration and signature [103], there are theoretical results that guarantee that PH is stable under small perturbations [26,98]. After the PH signature is calculated, statistical hypothesis testing [10,15], or machine learning techniques such as SVM or k-NN [2,48,78] can be used on these features to study the differences within the data set of interest. It is important to note that PH is very flexible, as different choices can be made in every step of the pipeline, regarding the input and output of PH, as detailed in the remainder of this section. 2.1 Approximation of a space at scale r R Instead of the Vietoris-Rips complex, other types of complexes can be used to approximate the data X at the given scale r R. For instance, since Vietoris-Rips simplicial complex is large [80], one might rather choose the alpha complex [46] (for a visualization, see Appendix C.2), which is closely related to the Vietoris-Rips complex [63], but consists of significantly less simplices and is faster to construct when the dimension of the ambient space is 2 or 3 (for details, see Appendix A.4). If data X is an image rather than a point cloud, or if the point cloud can be seen as an image without losing important information, cubical complexes [59] might be a more suitable choice, where vertices, edges and triangles are replaced by vertices, edges and squares (for a visualization, see Appendix E.1). 2.2 Filtration A filtration {Kr}r R of a point cloud in Rd can be constructed from any function f : Rd R by considering Kr to be the sublevel set of f thresholded by r R: {y Rd | f(y) r}. The underlying filtration function in the common PH pipeline introduced above is the distance function δX : Rd R, where δX(y) = min{dist(y, x) | x X} is the distance to point cloud X. Indeed, the Vietoris-Rips simplicial complex V R(X, r) approximates the sublevel set Kr = δ 1 X (( , r]) = {y Rd | δX(y) r} = x XB(x, r), where B(x, r) is a ball with radius r centered around x X [27]. However, PH on such a filtration is very sensitive to outliers, since even a single outlier changes δX significantly. In the presence of outliers, it is better to replace the distance function with the Distance-to-Measure (DTM) δX,m : Rd R, where δX,m(x) is the average distance from a number of neighbors on the point cloud [5,27] (for a visualization, see Appendix C.2). However, depending on the task, there are many other filtration functions one could choose, such as rank [84], height, radial, erosion, dilation [48], and the resulting PH captures completely different information about the cycles [103]. For example, whereas PH with respect to the Vietoris-Rips filtration encodes the size of the hole, PH on the height filtration informs about the position of the hole. For a more detailed discussion about the influence of filtration, see Appendix F. 2.3 Persistence signature Next to PIs and PLs, a plethora of persistence signatures has been introduced in the literature, e.g., Betti numbers [57,106] or Euler characteristic [72] (across scales), or even scalar summaries such as amplitude [48], entropy [35,93], or algebraic functions of the birth and death values [4,62]. Some of these signatures summarize the same information, but lie in different metric spaces [105]. Others, however, such as the scalar summaries listed above, discard information compared to PDs, as it might sometimes be useful to, e.g., only capture the (total or maximum) persistence of intervals, but not all the detailed information about all birth and death values (see Appendix F). 3 Number of holes In this section, we focus on the task of (ordinal) classification of point clouds by the number of 1-dimensional holes. Research in psychology shows that global properties often dominate perception, and, in particular, that topological invariants such as number of holes, inside versus outside, and connectivity can be effective primitives for recognizing shapes [85]. Extracting such topological information can therefore prove useful for many computer vision tasks. There are theoretical results in the literature that ensure that PH with respect to the alpha simplicial complex can be successful for this problem (Appendix A.2), and the computational experiments that follow demonstrate this success in practice. Data We consider 20 different shapes in R2 and R3, with four different shapes having the same number of holes (0, 1, 2, 4 or 9). For each shape, we construct 50 point clouds each consisting of 1 000 points sampled from a uniform distribution over the shape, resulting in a balanced data set of 1 000 = 20 50 point clouds. A few examples of these point clouds are shown in Figure 1. The label of a point cloud is the number of holes in the underlying shape. number of holes 0 1 2 4 9 Figure 1: Number of holes data set. PH pipeline For each point cloud X and scale r R, we consider its alpha complex. We will look into scenarios in which data contains noise, and therefore, instead of the standard distance function, we consider Distance-to-Measure (DTM) as the filtration function. We extract 1-dimensional PDs, that are then transformed to PIs, PLs, or a simple signature consisting only of lifespans l = d b of the 10 most persisting cycles (as there are at maximum 9 holes of interest in the given data set)2, and classified with an SVM. We consider a PH simple" pipeline, which relies on the 10 lifespans, and a PH" pipeline wherein grid search is employed to choose the best out of the three aforementioned persistent signatures and the values of their parameters. For more details on the pipeline, see Appendix B.2 and Appendix C.2. Results We investigate the clean and robust test accuracy under four types of transformations (translation, rotation, stretch, shear) and two types of noise (Gaussian noise, outliers). For more details on these transformations, see Appendix C.1. We train the classifier on 80% of the original point clouds, and test on the remaining 20% of the data either in its original form or subject to transformations and noise. The results reported in Figure 2 (with detailed results across multiple runs in Appendix C.3) show that PH obtains very good test accuracy on this classification task, even in the presence of affine transformations or noise, outperforming baseline machineand deep-learning techniques.3 We reach a similar conclusion in case of limited training data and computational resources (Appendix C.4, Appendix C.6). Firstly, the evolution of the test accuracy across different amounts of training data demonstrates that PH achieves good performance for a small number of training point clouds, which is not the case for other pipelines. Secondly, although the hyperparameter tuning of the PH pipeline does take time (as we consider a wide range of parameters for the different persistence signatures), it is still less than for Point Net. Moreover, Figure 2 shows that even the simple PH pipeline, where the SVM is used directly on the lifespans of the 10 most persisting cycles (without any tuning of PH-related parameters) performs well. 4 Curvature This section considers a regression task to predict the curvature of an underlying shape based on a point cloud sample. Estimating curvature-related quantities is of prime importance in computer vision, computer graphics, computer-aided design or computational geometry, e.g., for surface segmentation, surface smoothing or denoising, surface reconstruction, and shape design [23]. For continuous 2Although a sphere has no 1-dimensional holes, its PD might consist of many short intervals which correspond to the small holes on the surface. In addition, in the presence of noise, additional small holes might appear for any point cloud. Hence, it is not a good idea to consider the cardinality |PD| of the PD as the signature. 3Interestingly, although Point Net was designed with the idea to be invariant to affine transformations, it performs poorly when the test data is translated or rotated (and this is consistent with some previous results [68,111,115,117 120]), or when it contains outliers. Traditional neural networks perform very poorly, which might not come as a big surprise, since it was recently demonstrated that they transform topologically complicated data into topologically simple one as it passes through the layers, vastly reducing the Betti numbers (nearly always even reducing to their lowest possible values: βk = 0 for k > 0, and β0 = 1) [76]. Of course, the choice of activation function and hyperparameters might have an important influence on performance [76]. Figure 2: Persistent homology can detect the number of holes. surfaces, normals and curvature are fundamental geometric notions which uniquely characterize local geometry up to rigid transformations [52]. Recently, it has been shown that, using PH, curvature can be both recovered in theory (Appendix A.3), and effectively estimated in practice [17]. We run a similar experiment, evaluating the PH pipeline against our baselines, and also taking a closer look into the importance of short intervals. Data A balanced data set is generated in the same way as in [17]: We consider unit disks Dκ on surfaces of constant curvature κ: (i) κ = 0, Euclidean plane, (ii) κ > 0, sphere with radius 1/ κ, and (iii) κ < 0, Poincaré disk model of the hyperbolic plane. Curvature κ lies in the interval [ 2, 2] so that a disk with radius one can be embedded on the upper hemisphere of a sphere with constant curvature κ (as it spherical cap). For each κ { 2, 1.96, . . . , 0.04, 0, 0.04, . . . , 1.96}, we construct 10 point clouds by sampling 500 points from the unit disk Dκ with the probability measure proportional to the surface area measure [15, Section 2.7, Section 4.1]. A few examples with κ { 2, 1, 0.1, 0, 0.1, 1, 2} are illustrated in Figure 34.These 101 10 = 1 010 point clouds are considered as the training data, whereas the test data set is built in a similar way for 100 values of κ chosen uniformly at random from [ 2, 2]. The label of a point cloud is the curvature κ of the underlying disk Dκ. Note that all these disks are homoemorphic: they are contractible, so that their homology is trivial and homology is thus unable to distinguish between them [17]. curvature -2.00 -1.00 -0.10 0.00 0.10 1.00 2.00 Figure 3: Curvature data set. PH pipeline For each point cloud X, we first calculate the suitable matrix of pairwise distances between the point-cloud points: hyperbolic, Euclidean or spherical, respectively for negative, zero and positive curvature [15, Section 2.7]. The input for PH is the filtered Vietoris-Rips simplicial complex.5 We extract 0and 1-dimensional PDs, which are then transformed into PIs, PLs or lifespans, to be fed to SVM. More details on the pipeline are provided in Appendix B.2 and Appendix D.1. Results Figure 4 shows the mean squared errors for the PH and other pipelines, together with their regression lines, with detailed results across multiple runs listed in Appendix D.2. The results show that PH indeed detects curvature, outperforming other methods.6 Next to the PH pipeline discussed 4The unit disks with negative curvature are here visualized on hyperbolic paraboloids. These saddle surfaces have non-constant curvature, but they locally resemble the hyperbolic plane. 5Alpha complex is faster to compute, but involves Delaunay triangulation, whose unique existence is guaranteed only in Euclidean spaces. To calculate PH, we rely on the Ripser software [102], which is at the time the most efficient library to compute PH with Vietoris-Rips complex [80]. 6Simple machine and deep learning techniques are able to differentiate between positive and negative curvature, but perform poorly in predicting the actual value of the curvature of the underlying surface. above, wherein a grid search is used to tune the parameters (Appendix B.2), we also consider SVM on the lists of lifespans of all persistence intervals (PH simple), and SVM only on the 10 longest lifespans (PH simple 10), in order to investigate if all persistence intervals contribute to prediction. We see that the performance drops if we only focus on the longest 10 intervals, so that the many short intervals together capture the geometry of interest for this problem. Similarly as in Section 3, the grid search across the different parameters for persistence signatures does take time (Appendix D.3), but Figure 4 shows that SVM on a simple signature of all (0-dimensional) lifespans performs well. We highlight that the data used here, as was the data in Bubenik s work [17], is sampled from surfaces with constant curvature. In future work it would be interesting to conduct similar experiments on shapes with non-constant curvature. 0-dim PH simple 0-dim PH simple 10 0-dim PH ML Point Net MSE = 0.06 MSE = 0.21 MSE = 0.08 MSE = 0.34 MSE = 578.28 1-dim PH simple 1-dim PH simple 10 1-dim PH NN shallow NN deep MSE = 0.34 MSE = 0.29 MSE = 0.18 MSE = 0.66 MSE = 0.43 Figure 4: Persistent homology can detect curvature. 5 Convexity In this section, we consider the binary classification task that consists of detecting whether a point cloud is sampled from a convex set. Convexity is a fundamental concept in geometry [40], which plays an important role in learning, optimization [9], numerical analysis, statistics, information theory, and economics [91]. Furthermore, points of convexities and concavities have been demonstrated as crucial for human perception of shapes across many experiments [94]. To the best of our knowledge, prior to our work PH has not been employed to analyze convexity, and it is a task for which PH s effectiveness might seem surprising. In the first decade after the introduction of PH, it was seen primarily as the descriptor of global topology. Recently, there have been many discussions and greater understanding that PH also captures local geometry [3]. However, it is still suggested that the long persistence intervals capture topology (as was the case with the detection of holes in Section 3), and many even too many for the human eye to count short persistence intervals capture geometrical properties (as was the case with curvature prediction in Section 4). However, as we show in Theorem 1 (proof in Appendix A.1) and as our experiments suggest, it is a single, and the second-longest persistence interval that enables us to detect concavity. A crucial ingredient in our result is the introduction of tubular filtrations (Definition 1), which, to the best of our knowledge, are a novel contribution to the TDA literature (details in Appendix A.1). Definition 1. Given a line α Rd, we define the tubular function with respect to α as follows: τα : Rd R x 7 dist(x, α) , where dist(x, α) is the distance of the point x from the line α. Given X Rd and a line α, we are interested in studying the sublevel sets of τα, i.e., the subsets of X consisting of points at a specific distance from the line. We define Xτα,r = {x X | τα(x) r} = {x X | dist(x, α) r} . We call {Xτα,r}r R 0 the tubular filtration with respect to α. Theorem 1. Let X Rd be triangulizable. We have that X is convex if and only if for every line α in Rd the persistence diagram in degree 0 with respect to the tubular filtration {Xτα,r}r R 0 contains exactly one interval. Data We construct a balanced data set by sampling 5 000 points from convex and concave (nonconvex) shapes in R2. First, we consider the regular" convex shapes of triangle, square, pentagon and circle, and their concave variants, sampling 60 point clouds of each of the eight shapes, 480 point clouds in total. Next, we build 480 random" convex and concave shapes, in order to be able to investigate if an algorithm is actually detecting convexity, or only the different basic shapes. A few examples are shown in Figure 5. To construct a random convex shape, we generate 10 points at random, and then build their convex hull using the quickhull algorithm [108]. We construct random concave shapes in a similar way, but instead of the convex hull, we build the alpha shape [45,47] with the optimized alpha parameter, which gives a finer approximation of a shape from a given set of points. If the alpha shape is convex (i.e., if the alpha shape and its convex hull are the same), we reconstruct the concave shape from scratch. A point cloud has label 1 if it is sampled from a convex shape, and 0 otherwise. convexity 1 0 Figure 5: Convexity data set. PH pipeline To build a filtration, we consider cubical complexes filtered by tubular functions that measure the distance of points from a certain line, see Definition 1 for a precise definition. For a good choice of line, multiple components would be seen in the filtration of a point cloud sampled from a concave shape, at least for some values r R (see also the illustrations in Appendix A.1). For this reason we consider the cubical complex, rather than the standard Vietoris-Rips simplicial complex wherein these separate components could be connected with an edge (for details, see Appendix E.1). To build an image from the point cloud, we construct a 20 20 grid and define a pixel as black if it contains any point-cloud points, and white otherwise. Since sources of concavity can lie anywhere on the point cloud, we consider nine different lines for the tubular filtration function (for a visualization of the pipeline, see Appendix E.1). For each of the nine lines, we extract 0-dimensional PD, as it captures information about the components. If the point cloud is thus sampled from a convex shape, its PD will only see a single component for any line, whereas there will be multiple components at least for some lines for points clouds sampled from concave shapes. For this reason, for each of the nine lines, we focus our attention only on the lifespan of the second most persisting cycle. We can consider this 9-dimensional vector as our PH signature, but in our experiment choose an even simpler summary: the maximum of these lifespans, since we only care if there are multiple components for at least one line. This scalar could even be used as some measure of a level of concavity of a shape. Results As already indicated, to gain some insights into how well the different approaches discriminate convexity from concavity rather than differentiating between the different basic shapes, we look at the classification accuracies under different conditions (Figure 6, with detailed results across multiple runs in Appendix E.2). We start with the easiest case, where both the train and test data consist of the simple regular convex and concave shapes (Figure 5, first row), and then proceed to the scenario where both train and test data are random shapes (Figure 5, second row). Next we proceed to out-of-distribution test data, where we train on the regular and test on random shapes, or vice versa. In every case, we train on 400 and test on 80 point clouds. The results show that PH is able to detect convexity, surpassing other methods significantly in all scenarios except for Point Net in the scenario on the data set of regular shapes which performs on par. Results reported in Appendix E.3 show PH is also computationally efficient. Figure 6: Persistent homology can detect convexity. The PH pipeline above makes a wrong prediction when concavity is barely pronounced, or if it is missed by the selected tubular filtration lines (for details see Appendix E.4). However, the accuracy of PH can easily be improved simply by considering a finer resolution for the cubical complexes and/or additional tubular filtration lines. The particular PH pipeline summarized in this section would also make a wrong prediction if the data set would include shapes that have small or non-central holes, e.g., a square with a hole in the top left corner. In this case, the accuracy could also be improved by considering a finer cubical complex resolution and by considering additional non-central tubular filtration lines within shapes, or by adding (the maximum lifespan of the) 1-dimensional PH which captures holes. The pipeline is not limited to polygons, or connected shapes, and it can be generalized to surfaces in higher dimensions (Theorem 1). In Appendix G, we also consider the real-world data set FLAVIA for which we demonstrate that a PH pipeline is effective in detecting a continuous measure of convexity. 6 Implications for PH interpretation: topology vs. geometry Here we discuss how our results contribute to the important and ongoing discussion about the interpretation of long versus short persistence intervals. When PH was first introduced in the literature, the long intervals were commonly considered as important or signal , and short intervals as irrelevant or noise [38]. Subsequently the discussion has refined when it was shown that short and medium-length persistence intervals have the most distinguishing power for specific types of applications [8,99]. The current understanding is roughly that long intervals reflect the topological signal, and (many) short intervals can help in detecting geometric features [3]. We believe that our work brings new insight into this discussion. We give a summary of the implications of our work in this section, and we provide a more detailed discussion in Appendix F. Topology and long persistence Stability results guarantee that a number of longest persistence intervals reflect the topological signal, i.e., the number of cycles [32]. These theorems give information about the threshold that differentiates between long and short persistence intervals. In Section 3, where we focus on the topology of underlying shapes, the experiments demonstrate that this threshold can be learned with simple machine learning techniques. However, it is important to highlight that the distinction between long and short persistence is vague in practice. Indeed, seemingly short persistence intervals capture the topology in Section 3, but the second-longest interval is topological noise in Section 5, since every shape in the data set has only a single component (although this second-longest interval captures important geometric information, what enabled us to discriminate between convex and concave shapes). These two problems also clearly indicate how the long intervals that encode topology might or might not be irrelevant, depending on the signal of the particular application domain. Geometry and short persistence The current understanding is that (many) short persistence intervals detect geometry. Section 4 confirms that this indeed can be the case. However, we highlight that all cycles can encode geometric information, such as the information about their size (with respect to the Vietoris-Rips and related filtrations, as in Sections 3 and 4) or their position (with respect to the height or tubular filtration, as in Section 5). This further implies that, depending on the application, any number of short or intervals of any persistence can be important, which was clearly demonstrated in Section 5, where we show that a single interval detects convexity. 7 Conclusions Main contribution The goal of this work is to gain a better understanding of the topological and geometric features that can be captured with persistent homology. We focus on the detection of number of holes (Section 3), curvature (Section 4), and convexity (Section 5). Theoretical evidence for the first two classes of problems has been established in the literature, and we prove a new result that guarantees that PH can detect convexity (Theorem 1). We also experimentally demonstrate that PH can solve all three problems for synthetic point clouds in R2 and R3, outperforming a few baselines. This is true even when there is limited training data and computational resources, and for noisy or out of distribution test data. For convexity detection, we also show the effectiveness of PH in a real world plant morphology application. Relevance Firstly, the findings point the way to further advances in utilizing the potential of PH in applications: we can expect PH to be successful for classification or regression problems where the data classes differ with respect to number of holes, curvature and/or convexity. Detailed guidelines are discussed in Appendix F. Due to the crucial role of shape classification in understanding and recognizing physical structures and objects, image processing and computer vision [73], our results demonstrate that PH can to borrow the words from Wigner [112] remain valid in future research, and extend, to our pleasure", and lesser bafflement, to a variety of applications. Secondly, the results advance the discussion about the importance of long and short persistence intervals, and their relationship to topology and geometry (Section 6). Topology is captured by the long intervals, geometry is encoded in all persistence intervals, and any interval can encode the signal in the particular application domain. Limitations The results focus on three selected problems and data sets, and it would therefore be interesting to consider other tasks. In addition, we do not have an extensive comparison of the state-of-the-art for the given problems. Our work seeks to understand if PH is successful for a selected set of tasks by benchmarking it against some well-performing methods. Future research An in-depth analysis of the hypothetical applications discussed in this paper (Appendix F) and selected success stories of PH from the literature could further improve our understanding of the topological and geometric information encoded in PH, and the interpretation of persistence intervals of different lengths. Alternative approaches for detection of convexity with PH (relying on higher homological dimensions, or multiparameter persistence) are particularly interesting avenues for further work. Furthermore, even though our results imply that PH features are recommended over baseline models for the three selected classes of problems, they also provide inspiration on how to improve existing learning architectures. Further work could investigate deep learning models on PH (and standard) features or kernels [14, 55, 89, 116], an additional network layer for topological signatures, or PH-based priors, regularization or loss functions [14,34,36,37,56, 113,121]. Potential negative societal impact While we recognize that the applications of shape analysis can take many different directions, we do not foresee a direct path of this research to negative societal impacts. Funding transparency statement This research was conducted while RT was a Fulbright Visiting Researcher at UCLA. GM acknowledges support from ERC grant 757983, DFG grant 464109215, and NSF-CAREER grant DMS-2145630. NO acknowledges support from the Royal Society, under grant RGS\R2\212169. [1] Point cloud classification with Point Net. https://keras.io/examples/vision/ pointnet/. Accessed: 2022-02-01. [2] Henry Adams, Tegan Emerson, Michael Kirby, Rachel Neville, Chris Peterson, Patrick Shipman, Sofya Chepushtanova, Eric Hanson, Francis Motta, and Lori Ziegelmeier. Persistence images: A stable vector representation of persistent homology. The Journal of Machine Learning Research, 18(1):218 252, 2017. [3] Henry Adams and Michael Moy. Topology applied to machine learning: From global to local. Frontiers in Artificial Intelligence, 4:54, 2021. [4] Aaron Adcock, Erik Carlsson, and Gunnar Carlsson. The ring of algebraic functions on persistence bar codes. ar Xiv preprint ar Xiv:1304.0530, 2013. [5] Hirokazu Anai, Frédéric Chazal, Marc Glisse, Yuichi Ike, Hiroya Inakoshi, Raphaël Tinarrage, and Yuhei Umeda. DTM-based filtrations. In Topological Data Analysis, pages 33 66. Springer, 2020. [6] Aras Asaad and Sabah Jassim. Topological data analysis for image tampering detection. In International Workshop on Digital Watermarking, pages 136 146. Springer, 2017. [7] Ulrich Bauer. Ripser: efficient computation of Vietoris-Rips persistence barcodes. Journal of Applied and Computational Topology, 2021. [8] Paul Bendich, James S Marron, Ezra Miller, Alex Pieloch, and Sean Skwerer. Persistent homology analysis of brain artery trees. Annals of Applied Statistics, 10(1):198, 2016. [9] Piotr Berman, Meiram Murzabulatov, and Sofya Raskhodnikova. Testing convexity of figures under the uniform distribution. Random Structures & Algorithms, 54(3):413 443, 2019. [10] Eric Berry, Yen-Chi Chen, Jessi Cisewski-Kehe, and Brittany Terese Fasy. Functional summaries of persistence diagrams. ar Xiv preprint ar Xiv:1804.01618, 2018. [11] Omer Bobrowski and Sayan Mukherjee. The topology of probability distributions on manifolds. Probability Theory and Related Fields, 161(3):651 686, 2015. [12] Jean-Daniel Boissonnat, Frédéric Chazal, and Mariette Yvinec. Geometric and Topological Inference, volume 57. Cambridge University Press, 2018. [13] Doug M Boyer, Jesus Puente, Justin T Gladman, Chris Glynn, Sayan Mukherjee, Gabriel S Yapuncich, and Ingrid Daubechies. A new fully automated approach for aligning and comparing shapes. The Anatomical Record, 298(1):249 276, 2015. [14] Rickard Brüel-Gabrielsson, Bradley J Nelson, Anjan Dwaraknath, Primoz Skraba, Leonidas J Guibas, and Gunnar Carlsson. A topology layer for machine learning. ar Xiv preprint ar Xiv:1905.12200, 2019. [15] Peter Bubenik. Statistical topological data analysis using persistence landscapes. The Journal of Machine Learning Research, 16(1):77 102, 2015. [16] Peter Bubenik and Paweł Dłotko. A persistence landscapes toolbox for topological statistics. Journal of Symbolic Computation, 78:91 114, 2017. [17] Peter Bubenik, Michael Hull, Dhruv Patel, and Benjamin Whittle. Persistent homology detects curvature. Inverse Problems, 36(2):025008, 2020. [18] Zixuan Cang, Lin Mu, and Guo-Wei Wei. Representability of algebraic topology for biomolecules in machine learning based scoring and virtual screening. PLOS Computational Biology, 14(1):e1005929, 2018. [19] Zixuan Cang and Guo-Wei Wei. Topology Net: Topology based deep convolutional and multitask neural networks for biomolecular property predictions. PLOS Computational Biology, 13(7):e1005690, 2017. [20] Gunnar Carlsson. Topological pattern recognition for point cloud data. Acta Numerica, 23:289 368, 2014. [21] Mathieu Carrière, Frédéric Chazal, Yuichi Ike, Théo Lacombe, Martin Royer, and Yuhei Umeda. Pers Lay: A neural network layer for persistence diagrams and new graph topological signatures. In International Conference on Artificial Intelligence and Statistics, pages 2786 2796. PMLR, 2020. [22] Mathieu Carrière, Steve Y Oudot, and Maks Ovsjanikov. Stable topological signatures for points on 3D shapes. In Computer Graphics Forum, volume 34, pages 1 12. Wiley Online Library, 2015. [23] Frédéric Cazals and Marc Pouget. Estimating differential quantities using polynomial fitting of osculating jets. Computer Aided Geometric Design, 22(2):121 146, 2005. [24] Wojciech Chachólski and Henri Riihimaki. Metrics and stabilization in one parameter persistence. SIAM Journal on Applied Algebra and Geometry, 4(1):69 98, 2020. [25] Frédéric Chazal and David Cohen-Steiner. Geometric inference, 2013. [26] Frédéric Chazal, David Cohen-Steiner, André Lieutier, and Boris Thibert. Stability of curvature measures. In Computer Graphics Forum, volume 28, pages 1485 1496. Wiley Online Library, 2009. [27] Frédéric Chazal, David Cohen-Steiner, and Quentin Mérigot. Geometric inference for probability measures. Foundations of Computational Mathematics, 11(6):733 751, 2011. [28] Frédéric Chazal, Vin De Silva, Marc Glisse, and Steve Oudot. The structure and stability of persistence modules, volume 10. Springer, 2016. [29] Frédéric Chazal, Brittany Fasy, Fabrizio Lecci, Bertrand Michel, Alessandro Rinaldo, Alessandro Rinaldo, and Larry Wasserman. Robust topological inference: Distance to a measure and kernel distance. The Journal of Machine Learning Research, 18(1):5845 5884, 2017. [30] Frédéric Chazal, Brittany Terese Fasy, Fabrizio Lecci, Alessandro Rinaldo, Aarti Singh, and Larry Wasserman. On the bootstrap for persistence diagrams and landscapes. ar Xiv preprint ar Xiv:1311.0376, 2013. [31] Frédéric Chazal, Leonidas J Guibas, Steve Y Oudot, and Primoz Skraba. Scalar field analysis over point cloud data. Discrete & Computational Geometry, 46(4):743 775, 2011. [32] Frédéric Chazal and Steve Yann Oudot. Towards persistence-based reconstruction in Euclidean spaces. In Proceedings of the twenty-fourth annual symposium on Computational geometry, pages 232 241, 2008. [33] Chao Chen and Michael Kerber. Persistent homology computation with a twist. In Proceedings 27th European workshop on computational geometry, volume 11, pages 197 200, 2011. [34] Chao Chen, Xiuyan Ni, Qinxun Bai, and Yusu Wang. A topological regularizer for classifiers via persistent homology. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 2573 2582. PMLR, 2019. [35] Harish Chintakunta, Thanos Gentimis, Rocio Gonzalez-Diaz, Maria-Jose Jimenez, and Hamid Krim. An entropy-based persistence barcode. Pattern Recognition, 48(2):391 401, 2015. [36] James Clough, Nicholas Byrne, Ilkay Oksuz, Veronika A Zimmer, Julia A Schnabel, and Andrew King. A topological loss function for deep-learning based image segmentation using persistent homology. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020. [37] James R Clough, Ilkay Oksuz, Nicholas Byrne, Julia A Schnabel, and Andrew P King. Explicit topological priors for deep-learning based image segmentation using persistent homology. In International Conference on Information Processing in Medical Imaging, pages 16 28. Springer, 2019. [38] David Cohen-Steiner, Herbert Edelsbrunner, and John Harer. Stability of persistence diagrams. Discrete & Computational Geometry, 37(1):103 120, 2007. [39] Anne Collins, Afra Zomorodian, Gunnar Carlsson, and Leonidas J Guibas. A barcode shape descriptor for curve point cloud data. Computers & Graphics, 28(6):881 894, 2004. [40] Loïc Crombez, Guilherme D da Fonseca, and Yan Gérard. Efficient algorithms to test digital convexity. In International Conference on Discrete Geometry for Computer Imagery, pages 409 419. Springer, 2019. [41] Justin Curry, Sayan Mukherjee, and Katharine Turner. How many directions determine a shape and other sufficiency results for two topological transforms. ar Xiv preprint ar Xiv:1805.09782, 2018. [42] Thibault de Surrel, Felix Hensel, Mathieu Carrière, Théo Lacombe, Yuichi Ike, Hiroaki Kurihara, Marc Glisse, and Frédéric Chazal. Rips Net: a general architecture for fast and robust estimation of the persistent homology of point clouds. ar Xiv preprint ar Xiv:2202.01725, 2022. [43] Paweł Dłotko and Hubert Wagner. Simplification of complexes for persistent homology computations. Homology, Homotopy and Applications, 16(1):49 63, 2014. [44] Herbert Edelsbrunner and John Harer. Computational topology: An introduction. American Mathematical Society, 2010. [45] Herbert Edelsbrunner, David Kirkpatrick, and Raimund Seidel. On the shape of a set of points in the plane. IEEE Transactions on information theory, 29(4):551 559, 1983. [46] Herbert Edelsbrunner, David Letscher, and Afra Zomorodian. Topological persistence and simplification. In Proceedings 41st Annual Symposium on Foundations of Computer Science, pages 454 463. IEEE, 2000. [47] Herbert Edelsbrunner and Ernst P Mücke. Three-dimensional alpha shapes. ACM Transactions on Graphics (TOG), 13(1):43 72, 1994. [48] Adélie Garin and Guillaume Tauzin. A topological reading lesson: Classification of MNIST using TDA. In 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), pages 1551 1556. IEEE, 2019. [49] Robert Ghrist, Rachel Levanger, and Huy Mai. Persistent homology and Euler integral transforms. Journal of Applied and Computational Topology, 2(1):55 60, 2018. [50] Noah Giansiracusa, Robert Giansiracusa, and Chul Moon. Persistent homology machine learning for fingerprint classification. ar Xiv preprint ar Xiv:1711.09158, 2017. [51] Barbara Giunti. TDA-applications. https://www.zotero.org/groups/2425412/tda-applications. Accessed: 10/12/2022. [52] Paul Guerrero, Yanir Kleiman, Maks Ovsjanikov, and Niloy J Mitra. PCPNET learning local shape properties from raw point clouds. In Computer Graphics Forum, volume 37, pages 75 85. Wiley Online Library, 2018. [53] Allen Hatcher. Algebraic topology. Cambridge University Press, 2002. [54] Tong He, Haibin Huang, Li Yi, Yuqian Zhou, Chihao Wu, Jue Wang, and Stefano Soatto. Geo Net: Deep geodesic networks for point cloud analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6888 6897, 2019. [55] Christoph Hofer, Roland Kwitt, Marc Niethammer, and Andreas Uhl. Deep learning with topological signatures. ar Xiv preprint ar Xiv:1707.04041, 2017. [56] Xiaoling Hu, Fuxin Li, Dimitris Samaras, and Chao Chen. Topology-preserving deep image segmentation. Advances in Neural Information Processing Systems, 32, 2019. [57] Umar Islambekov, Monisha Yuvaraj, and Yulia R Gel. Harnessing the power of topological data analysis to detect change points in time series. ar Xiv preprint ar Xiv:1910.12939, 2019. [58] Maria-Jose Jimenez, Belen Medrano, David Monaghan, and Noel E O Connor. Designing a topological algorithm for 3D activity recognition. In International Workshop on Computational Topology in Image Context, pages 193 203. Springer, 2016. [59] Tomasz Kaczynski, Konstantin Mischaikow, and Marian Mrozek. Computational homology, volume 3. Springer, 2004. [60] Tomasz Kaczynski, Konstantin Mischaikow, and Marian Mrozek. Homology of Topological Polyhedra, pages 377 393. Springer New York, New York, NY, 2004. [61] Jules Raymond Kala, Serestina Viriri, Deshendran Moodley, and Jules Raymond Tapamo. Leaf classification using convexity measure of polygons. In International Conference on Image and Signal Processing, pages 51 60. Springer, 2016. [62] Sara Kališnik. Tropical coordinates on the space of persistence barcodes. Foundations of Computational Mathematics, 19(1):101 129, 2019. [63] Jisu Kim, Jaehyeok Shin, Frédéric Chazal, Alessandro Rinaldo, and Larry Wasserman. Homotopy reconstruction via the Cech complex and the Vietoris-Rips complex. ar Xiv preprint ar Xiv:1903.06955, 2019. [64] Ron Kimmel and Xue-Cheng Tai. Processing, Analyzing and Learning of Images, Shapes, and Forms: Part 2. Elsevier, 2019. [65] Vitaliy Kurlin. A fast and robust algorithm to count topologically persistent holes in noisy clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1458 1463, 2014. [66] Javier Lamar-León, Edel B Garcia-Reyes, and Rocio Gonzalez-Diaz. Human gait identification using persistent homology. In Iberoamerican Congress on Pattern Recognition, pages 244 251. Springer, 2012. [67] Longin Jan Latecki, Rolf Lakamper, and T Eckhardt. Shape descriptors for non-rigid shapes with a single closed contour. In Proceedings IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000 (Cat. No. PR00662), volume 1, pages 424 429. IEEE, 2000. [68] Hoanh Le. Geometric invariance of Point Net. Science and Engineering, 2021. [69] Yongjin Lee, Senja D Barthel, Paweł Dłotko, S Mohamad Moosavi, Kathryn Hess, and Berend Smit. Quantifying similarity of pore-geometry in nanoporous materials. Nature Communications, 8(1):1 8, 2017. [70] Javier Lamar Leon, Raúl Alonso, Edel Garcia Reyes, and Rocio Gonzalez Diaz. Topological features for monitoring human activities at distance. In International Workshop on Activity Monitoring by Multiple Distributed Sensing, pages 40 51. Springer, 2014. [71] Mao Li, Hong An, Ruthie Angelovici, Clement Bagaza, Albert Batushansky, Lynn Clark, Viktoriya Coneva, Michael J Donoghue, Erika Edwards, Diego Fajardo, et al. Topological data analysis as a morphometric method: using persistent homology to demarcate a leaf morphospace. Frontiers in Plant Science, 9:553, 2018. [72] Mao Li, Margaret H Frank, Viktoriya Coneva, Washington Mio, Daniel H Chitwood, and Christopher N Topp. The persistent homology mathematical framework provides enhanced genotype-to-phenotype associations for plant morphology. Plant Physiology, 177(4):1382 1395, 2018. [73] Kart-Leong Lim and Hamed Kiani Galoogahi. Shape classification using local and global features. In 2010 Fourth Pacific-Rim Symposium on Image and Video Technology, pages 115 120. IEEE, 2010. [74] Joseph SB Mitchell, David M Mount, and Christos H Papadimitriou. The discrete geodesic problem. SIAM Journal on Computing, 16(4):647 668, 1987. [75] Guido Montúfar, Nina Otter, and Yuguang Wang. Can neural networks learn persistent homology features? ar Xiv preprint ar Xiv:2011.14688, 2020. [76] Gregory Naitzat, Andrey Zhitnikov, and Lek-Heng Lim. Topology of deep neural networks. The Journal of Machine Learning Research, 21(184):1 40, 2020. [77] Partha Niyogi, Stephen Smale, and Shmuel Weinberger. Finding the homology of submanifolds with high confidence from random samples. Discrete & Computational Geometry, 39(1-3):419 441, 2008. [78] Ippei Obayashi, Yasuaki Hiraoka, and Masao Kimura. Persistence diagrams with linear machine learning models. Journal of Applied and Computational Topology, 1(3-4):421 449, 2018. [79] Nina Otter. Magnitude meets persistence. Homology theories for filtered simplicial sets. ar Xiv preprint ar Xiv:1807.01540, 2018. to appear in Homology, Homotopy and Applications. [80] Nina Otter, Mason A Porter, Ulrike Tillmann, Peter Grindrod, and Heather A Harrington. A roadmap for the computation of persistent homology. EPJ Data Science, 6(1):17, 2017. [81] Rahul Paul and Stephan Chalup. Estimating Betti numbers using deep learning. In 2019 International Joint Conference on Neural Networks (IJCNN), pages 1 7. IEEE, 2019. [82] Rahul Paul and Stephan K Chalup. A study on validating non-linear dimensionality reduction using persistent homology. Pattern Recognition Letters, 100:160 166, 2017. [83] Jose A Perea, Anastasia Deckard, Steve B Haase, and John Harer. Sw1pers: Sliding windows and 1-persistence scoring; discovering periodicity in gene expression time series data. BMC Bioinformatics, 16(1):257, 2015. [84] Giovanni Petri, Martina Scolamiero, Irene Donato, and Francesco Vaccarino. Topological strata of weighted complex networks. PLOS ONE, 8(6):e66506, 2013. [85] James R Pomerantz. Wholes, holes, and basic features in vision. Trends in Cognitive Sciences, 7(11):471 473, 2003. [86] Rolandos Alexandros Potamias, Alexandros Neofytou, Kyriaki Margarita Bintsi, and Stefanos Zafeiriou. Graph Walks: Efficient shape agnostic geodesic shortest path estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2968 2977, 2022. [87] Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. Point Net: Deep learning on point sets for 3D classification and segmentation. https://github.com/charlesq34/pointnet. Accessed: 2022-02-01. [88] Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. Point Net: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 652 660, 2017. [89] Archit Rathore, Sourabh Palande, Jeffrey S Anderson, Brandon A Zielinski, P Thomas Fletcher, and Bei Wang. Autism classification using topological features and deep learning: a cautionary tale. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 736 744. Springer, 2019. [90] Vanessa Robins. Computational topology for point data: Betti numbers of α-shapes. In Morphology of Condensed Matter, pages 261 274. Springer, 2002. [91] Christian Ronse. A bibliography on digital and computational convexity (1961 1988). IEEE Transactions on Pattern Analysis and Machine Intelligence, 11(2):181 190, 1989. [92] Martin Royer, Frédéric Chazal, Clément Levrard, Yuichi Ike, and Yuhei Umeda. ATOL: Measure vectorisation for automatic topologically-oriented learning. ar Xiv preprint ar Xiv:1909.13472, 2019. [93] Matteo Rucco, Filippo Castiglione, Emanuela Merelli, and Marco Pettini. Characterisation of the idiotypic immune network through persistent entropy. In Proceedings of ECCS 2014, pages 117 128. Springer, 2016. [94] Gunnar Schmidtmann, Ben J Jennings, and Frederick AA Kingdom. Shape recognition: convexities, concavities and things in between. Scientific Reports, 5(1):1 11, 2015. [95] Benjamin Schweinhart. Fractal dimension and the persistent homology of random geometric complexes. Advances in Mathematics, 372:107291, 2020. [96] Thomas Sikora. The MPEG-7 visual standard for content description An overview. IEEE Transactions on Circuits and Systems for Video Technology, 11(6):696 702, 2001. [97] Nikhil Singh, Heather D Couture, JS Marron, Charles Perou, and Marc Niethammer. Topological descriptors of histology images. In International Workshop on Machine Learning in Medical Imaging, pages 231 239. Springer, 2014. [98] Primoz Skraba and Katharine Turner. Wasserstein stability for persistence diagrams. ar Xiv preprint ar Xiv:2006.16824, 2020. [99] Bernadette J Stolz, Heather A Harrington, and Mason A Porter. Persistent homology of time-dependent functional networks constructed from coupled time series. Chaos: An Interdisciplinary Journal of Nonlinear Science, 27(4):047410, 2017. [100] The GUDHI Project. GUDHI User and Reference Manual. GUDHI Editorial Board, 3.3.0 edition, 2020. [101] The RIVET Developers. Rivet, 2020. [102] Christopher Tralie, Nathaniel Saul, and Rann Bar-On. Ripser.py: A lean persistent homology library for python. The Journal of Open Source Software, 3(29):925, Sep 2018. [103] Renata Turkeš, Jannes Nys, Tim Verdonck, and Steven Latré. Noise robustness of persistent homology on greyscale images, across filtrations and signatures. PLOS ONE, 16(9):e0257215, 2021. [104] Katharine Turner, Sayan Mukherjee, and Doug M Boyer. Persistent homology transform for modeling shapes and surfaces. Information and Inference: A Journal of the IMA, 3(4):310 344, 2014. [105] Katharine Turner and Gard Spreemann. Same but different: Distance correlations between topological summaries. In Topological Data Analysis, pages 459 490. Springer, 2020. [106] Yuhei Umeda. Time series classification via topological data analysis. Information and Media Technologies, 12:228 239, 2017. [107] Oliver Vipond. Multiparameter persistence landscapes. The Journal of Machine Learning Research, 21(61):1 38, 2020. [108] Pauli Virtanen, Ralf Gommers, Travis E Oliphant, Matt Haberland, Tyler Reddy, David Cournapeau, Evgeni Burovski, Pearu Peterson, Warren Weckesser, Jonathan Bright, et al. Sci Py 1.0: fundamental algorithms for scientific computing in Python. Nature methods, 17(3):261 272, 2020. [109] Hubert Wagner, Chao Chen, and Erald Vuçini. Efficient computation of persistent homology for cubical data. In Topological methods in data analysis and visualization II, pages 91 106. Springer, 2012. [110] Menglun Wang, Zixuan Cang, and Guo-Wei Wei. A topology-based network tree for the prediction of protein protein binding affinity changes following mutation. Nature Machine Intelligence, 2(2):116 123, 2020. [111] Yan Wang, Yining Zhao, Shihui Ying, Shaoyi Du, and Yue Gao. Rotation-invariant point cloud representation for 3-D model recognition. IEEE Transactions on Cybernetics, 2022. [112] Eugene P Wigner. The unreasonable effectiveness of mathematics in the natural sciences. Communications on Pure and Applied Mathematics, 13:1 14, 1960. [113] Chi-Chong Wong and Chi-Man Vong. Persistent homology based graph convolution network for fine-grained 3D shape segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7098 7107, 2021. [114] Stephen Gang Wu, Forrest Sheng Bao, Eric You Xu, Yu-Xuan Wang, Yi-Fan Chang, and Qiao-Liang Xiang. A leaf recognition algorithm for plant classification using probabilistic neural network. In 2007 IEEE International Symposium on Signal Processing and Information Technology, pages 11 16. IEEE, 2007. [115] Chenxi Xiao and Juan Wachs. Triangle-Net: Towards robustness in point cloud learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 826 835, 2021. [116] Zuoyu Yan, Tengfei Ma, Liangcai Gao, Zhi Tang, and Chao Chen. Link prediction with persistent homology: An interactive view. In International Conference on Machine Learning, pages 11659 11669. PMLR, 2021. [117] Junming Zhang, Ming-Yuan Yu, Ram Vasudevan, and Matthew Johnson-Roberson. Learning rotation-invariant representations of point clouds using aligned edge convolutional neural networks. In 2020 International Conference on 3D Vision (3DV), pages 200 209. IEEE, 2020. [118] Zhiyuan Zhang, Binh-Son Hua, David W Rosen, and Sai-Kit Yeung. Rotation invariant convolutions for 3D point clouds deep learning. In 2019 International Conference on 3D Vision (3DV), pages 204 213. IEEE, 2019. [119] Ziwei Zhang, Xin Wang, Zeyang Zhang, Peng Cui, and Wenwu Zhu. Revisiting transformation invariant geometric deep learning: Are initial representations all you need? ar Xiv preprint ar Xiv:2112.12345, 2021. [120] Chen Zhao, Jiaqi Yang, Xin Xiong, Angfan Zhu, Zhiguo Cao, and Xin Li. Rotation invariant point cloud analysis: Where local geometry meets global topology. Pattern Recognition, page 108626, 2022. [121] Qi Zhao, Ze Ye, Chao Chen, and Yusu Wang. Persistence enhanced graph neural network. In International Conference on Artificial Intelligence and Statistics, pages 2896 2906. PMLR, 2020. [122] Zhen Zhou, Yongzhen Huang, Liang Wang, and Tieniu Tan. Exploring generalized shape analysis by topological representations. Pattern Recognition Letters, 87:177 185, 2017. [123] Afra Zomorodian and Gunnar Carlsson. Computing persistent homology. Discrete & Computational Geometry, 33(2):249 274, 2005. [124] Jovisa Zunic and Paul L Rosin. A new convexity measure for polygons. IEEE Transactions on Pattern Analysis and Machine Intelligence, 26(7):923 934, 2004. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] In both the introductory and concluding section, we clearly point to the relevant sections for each of the contributions. (b) Did you describe the limitations of your work? [Yes] The limitations are highlighted in the concluding Section 7, and in the end of curvature and convexity Sections 4 and 5 we provide more details about the interesting types of point clouds that could be considered. (c) Did you discuss any potential negative societal impacts of your work? [Yes] There is a paragraph in Section 7 addressing these impacts. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] See Section 1 and Appendix B. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Appendix B. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] See Appendix C, Appendix D, Appendix E. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See Appendix B. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] We use libraries described and acknowledged in Appendix B. (b) Did you mention the license of the assets? [Yes] See Appendix B. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] The data and code are provided in the Supplemental and referenced in Section 1 and Appendix B. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] We do not use any personal data. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] We construct data sets of point clouds that do not contain any personal information. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]