Topological deep learning

Last updated

Topological deep learning (TDL) [1] [2] [3] [4] [5] [6] is a research field that extends deep learning to handle complex, non-Euclidean data structures. Traditional deep learning models, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), excel in processing data on regular grids and sequences. However, scientific and real-world data often exhibit more intricate data domains encountered in scientific computations , including point clouds, meshes, time series, scalar fields graphs, or general topological spaces like simplicial complexes and CW complexes. [7] TDL addresses this by incorporating topological concepts to process data with higher-order relationships, such as interactions among multiple entities and complex hierarchies. This approach leverages structures like simplicial complexes and hypergraphs to capture global dependencies and qualitative spatial properties, offering a more nuanced representation of data. TDL also encompasses methods from computational and algebraic topology that permit studying properties of neural networks and their training process, such as their predictive performance or generalization properties.,. [8] [9] [10] [11] [12] [13] [14] The mathematical foundations of TDL are algebraic topology, differential topology, and geometric topology. Therefore, TDL can be generalized for data on differentiable manifolds, knots, links, tangles, curves, etc.

Contents

History and motivation

The term ``topological deep learning``, including multichannel TDL and multitask TDL, was first introduced in 2017. [15] Traditional techniques from deep learning often operate under the assumption that a dataset is residing in a highly-structured space (like images, where convolutional neural networks exhibit outstanding performance over alternative methods) or a Euclidean space. The prevalence of new types of data, in particular graphs, meshes, and molecules, resulted in the development of new techniques, culminating in the field of geometric deep learning, which originally proposed a signal-processing perspective for treating such data types. [16] While originally confined to graphs, where connectivity is defined based on nodes and edges, follow-up work extended concepts to a larger variety of data types, including simplicial complexes [17] [3] and CW complexes, [8] [18] with recent work proposing a unified perspective of message-passing on general combinatorial complexes. [1]

An independent perspective on different types of data originated from topological data analysis, which proposed a new framework for describing structural information of data, i.e., their "shape," that is inherently aware of multiple scales in data, ranging from local information to global information. [19] While at first restricted to smaller datasets, subsequent work developed new descriptors that efficiently summarized topological information of datasets to make them available for traditional machine-learning techniques, such as support vector machines or random forests. [20] Such descriptors ranged from new techniques for feature engineering over new ways of providing suitable coordinates for topological descriptors, [21] [22] [23] or the creation of more efficient dissimilarity measures. [24] [25] [26] [27]

Contemporary research in this field is largely concerned with either integrating information about the underlying data topology into existing deep-learning models or obtaining novel ways of training on topological domains.

Learning on topological spaces

Learning Tasks on topological domains can be broadly classified into three categories : cell classification, cell prediction and complex classification. Learning tasks on topological spaces.jpg
Learning Tasks on topological domains can be broadly classified into three categories : cell classification, cell prediction and complex classification.

Focusing on topology in the sense of point set topology, an active branch of TDL is concerned with learning on topological spaces, that is, on different topological domains.

An introduction to topological domains

One of the core concepts in topological deep learning is the domain upon which this data is defined and supported. In case of Euclidean data, such as images, this domain is a grid, upon which the pixel value of the image is supported. In a more general setting this domain might be a topological domain. Next, we introduce the most common topological domains that are encountered in a deep learning setting. These domains include, but not limited to, graphs, simplicial complexes, cell complexes, combinatorial complexes and hypergraphs.

Given a finite set S of abstract entities, a neighborhood function on S is an assignment that attach to every point in S a subset of S or a relation. Such a function can be induced by equipping S with an auxiliary structure. Edges provide one way of defining relations among the entities of S. More specifically, edges in a graph allow one to define the notion of neighborhood using, for instance, the one hop neighborhood notion. Edges however, limited in their modeling capacity as they can only be used to model binary relations among entities of S since every edge is connected typically to two entities. In many applications, it is desirable to permit relations that incorporate more than two entities. The idea of using relations that involve more than two entities is central to topological domains. Such higher-order relations allow for a broader range of neighborhood functions to be defined on S to capture multi-way interactions among entities of S.

Next we review the main properties, advantages, and disadvantages of some commonly studied topological domains in the context of deep learning, including (abstract) simplicial complexes, regular cell complexes, hypergraphs, and combinatorial complexes.

(a): A group S is made up of basic parts (vertices) without any connections.(b): A graph represents simple connections between its parts (vertices) that are elements of S.(c): A simplicial complex shows a way parts (relations) are connected to each other, but with strict rules about how they're connected.(d): Like simplicial complexes, a cell complex shows how parts (relations) are connected, but it's more flexible in how they're shaped (like 'cells').(f): A hypergraph shows any kind of connections between parts of S, but these connections aren't organized in any particular order.(e): A CC mixes elements from cell complexes (connections with order) and hypergraphs (varied connections), covering both kinds of setups. Higher order networks.png
(a): A group S is made up of basic parts (vertices) without any connections.(b): A graph represents simple connections between its parts (vertices) that are elements of S.(c): A simplicial complex shows a way parts (relations) are connected to each other, but with strict rules about how they're connected.(d): Like simplicial complexes, a cell complex shows how parts (relations) are connected, but it's more flexible in how they're shaped (like 'cells').(f): A hypergraph shows any kind of connections between parts of S, but these connections aren't organized in any particular order.(e): A CC mixes elements from cell complexes (connections with order) and hypergraphs (varied connections), covering both kinds of setups.

Comparisons among topological domains

Each of the enumerated topological domains has its own characteristics, advantages, and limitations:

  • Simplicial complexes
    • Simplest form of higher-order domains.
    • Extensions of graph-based models.
    • Admit hierarchical structures, making them suitable for various applications.
    • Hodge theory can be naturally defined on simplicial complexes.
    • Require relations to be subsets of larger relations, imposing constraints on the structure.
  • Cell Complexes
    • Generalize simplicial complexes.
    • Provide more flexibility in defining higher-order relations.
    • Each cell in a cell complex is homeomorphic to an open ball, attached together via attaching maps.
    • Boundary cells of each cell in a cell complex are also cells in the complex.
    • Represented combinatorially via incidence matrices.
  • Hypergraphs
    • Allow arbitrary set-type relations among entities.
    • Relations are not imposed by other relations, providing more flexibility.
    • Do not explicitly encode the dimension of cells or relations.
    • Useful when relations in the data do not adhere to constraints imposed by other models like simplicial and cell complexes.
  • Combinatorial Complexes [1] :
    • Generalize and bridge the gaps between simplicial complexes, cell complexes, and hypergraphs.
    • Allow for hierarchical structures and set-type relations.
    • Combine features of other complexes while providing more flexibility in modeling relations.
    • Can be represented combinatorially, similar to cell complexes.

Hierarchical structure and set-type relations

The properties of simplicial complexes, cell complexes, and hypergraphs give rise to two main features of relations on higher-order domains, namely hierarchies of relations and set-type relations. [1]

Rank function

A rank function on a higher-order domain X is an order-preserving function rk: XZ, where rk(x) attaches a non-negative integer value to each relation x in X, preserving set inclusion in X. Cell and simplicial complexes are common examples of higher-order domains equipped with rank functions and therefore with hierarchies of relations. [1]

Set-type relations

Relations in a higher-order domain are called set-type relations if the existence of a relation is not implied by another relation in the domain. Hypergraphs constitute examples of higher-order domains equipped with set-type relations. Given the modeling limitations of simplicial complexes, cell complexes, and hypergraphs, we develop the combinatorial complex, a higher-order domain that features both hierarchies of relations and set-type relations. [1]

The learning tasks in TDL can be broadly classified into three categories: [1]

  • Cell classification: Predict targets for each cell in a complex. Examples include triangular mesh segmentation, where the task is to predict the class of each face or edge in a given mesh.
  • Complex classification: Predict targets for an entire complex. For example, predict the class of each input mesh.
  • Cell prediction: Predict properties of cell-cell interactions in a complex, and in some cases, predict whether a cell exists in the complex. An example is the prediction of linkages among entities in hyperedges of a hypergraph.

In practice, to perform the aforementioned tasks, deep learning models designed for specific topological spaces must be constructed and implemented. These models, known as topological neural networks, are tailored to operate effectively within these spaces.

Topological neural networks

Central to TDL are topological neural networks (TNNs), [15] specialized architectures designed to operate on data structured in topological domains. [2] [1] Unlike traditional neural networks tailored for grid-like structures, TNNs are adept at handling more intricate data representations, such as graphs, simplicial complexes, and cell complexes. By harnessing the inherent topology of the data, TNNs can capture both local and global relationships, enabling nuanced analysis and interpretation.

Message passing topological neural networks

In a general topological domain, higher-order message passing involves exchanging messages among entities and cells using a set of neighborhood functions.

Definition: Higher-Order Message Passing on a General Topological Domain

Higher order message passing is a deep learning model defined on a topological domain and relies on message passing information among entities in the underlying domain in order to perform a learning task. Higher order message passing.png
Higher order message passing is a deep learning model defined on a topological domain and relies on message passing information among entities in the underlying domain in order to perform a learning task.

Let be a topological domain. We define a set of neighborhood functions on . Consider a cell and let for some . A message between cells and is a computation dependent on these two cells or the data supported on them. Denote as the multi-set , and let represent some data supported on cell at layer . Higher-order message passing on , [1] [8] induced by , is defined by the following four update rules:

  1. , where is the intra-neighborhood aggregation function.
  2. , where is the inter-neighborhood aggregation function.
  3. , where are differentiable functions.

Some remarks on Definition above are as follows.

First, Equation 1 describes how messages are computed between cells and . The message is influenced by both the data and associated with cells and , respectively. Additionally, it incorporates characteristics specific to the cells themselves, such as orientation in the case of cell complexes. This allows for a richer representation of spatial relationships compared to traditional graph-based message passing frameworks.

Second, Equation 2 defines how messages from neighboring cells are aggregated within each neighborhood. The function aggregates these messages, allowing information to be exchanged effectively between adjacent cells within the same neighborhood.

Third, Equation 3 outlines the process of combining messages from different neighborhoods. The function aggregates messages across various neighborhoods, facilitating communication between cells that may not be directly connected but share common neighborhood relationships.

Fourth, Equation 4 specifies how the aggregated messages influence the state of a cell in the next layer. Here, the function updates the state of cell based on its current state and the aggregated message obtained from neighboring cells.

Non-message passing topological neural networks

While the majority of TNNs follow the message passing paradigm from graph learning, several models have been suggested that do not follow this approach. For instance, Maggs et al. [28] leverage geometric information from embedded simplicial complexes, i.e., simplicial complexes with high-dimensional features attached to their vertices.This offers interpretability and geometric consistency without relying on message passing. Furthermore, in [29] a contrastive loss-based method was suggested to learn the simplicial representation.

Learning on topological descriptors

Motivated by the modular nature of deep neural networks, initial work in TDL drew inspiration from topological data analysis, and aimed to make the resulting descriptors amenable to integration into deep-learning models. This led to work defining new layers for deep neural networks. Pioneering work by Hofer et al., [30] for instance, introduced a layer that permitted topological descriptors like persistence diagrams or persistence barcodes to be integrated into a deep neural network. This was achieved by means of end-to-end-trainable projection functions, permitting topological features to be used to solve shape classification tasks, for instance. Follow-up work expanded more on the theoretical properties of such descriptors and integrated them into the field of representation learning. [31] Other such topological layers include layers based on extended persistent homology descriptors, [32] persistence landscapes, [33] or coordinate functions. [34] In parallel, persistent homology also found applications in graph-learning tasks. Noteworthy examples include new algorithms for learning task-specific filtration functions for graph classification or node classification tasks. [35] [36] [37]

Learning through alternative formulations

Most existing TDL techniques are rooted in homology. However, alternative mathematical approaches, such as topological Laplacians and topological Dirac operators, also provide valuable insights into the topological properties of TDL. Topological Laplacians, including Hodge Laplacians on differentiable manifolds, [38] combinatorial Laplacians for point clouds, [39] [40] and Khovanov Laplacians for knots and links, [41] serve as powerful tools for extracting topological features from their respective data formats. Despite being defined in distinct contexts, these Laplacians share a common algebraic foundation. They are constructed using the (co-)boundary operator and its adjoint, with their kernels isomorphic to homology groups. Consequently, the number of zero eigenvalues corresponds to the Betti numbers of the associated (co-)homology groups. Moreover, the nonzero eigenvalues provide richer insights into the data structure, particularly when analyzed through the perspective of spectral theory.

Hodge Laplacians, combinatorial Laplacians, and Khovanov Laplacians draw upon the mathematical fields of differential geometry, graph theory, and geometric topology, respectively, to extend the classical theory of homology beyond the domain of algebraic topology. Each functions as a bridge, linking algebraic topology to its associated mathematical discipline.

Persistent Hodge Laplacians were first introduced in 2019 to analyze data on differentiable manifolds with boundary. [42] Additionally, persistent combinatorial Laplacians, also known as persistent Laplacians, were developed for point cloud data. [43] These approaches extend classical persistent homology and have stimulated research interest, fueling advancements in both theory [44] [45] and applications. [46] [47] [48] Persistent Laplacians outperform persistent homology in extensive protein engineering tasks [46] and the prediction of mutation induced protein-protein binding affinity changes. [47] Persistent topological Laplacians have been constructed on various mathematical objects, including simplicial complex, directed flag complex, path complex, cellular sheaves, hypergraph, hyperdigraph, and differentiable manifolds. [49]

Persistent Dirac was constructed on various topological spaces, including simplicial complex, path complex, and hypergraph. [50] [51] [52] These new approaches extend the scope of TDL to manifold topological learning and curve data learning. [49]

Applications

TDL is rapidly finding new applications across different domains, including data compression, [53] enhancing the expressivity and predictive performance of graph neural networks, [17] [18] [35] action recognition, [54] and trajectory prediction. [55]

Topology inherently simplifies data, which implies the irreversible loss of certain information. Therefore, competitive performance from TDL mostly involves intrinsically complex data, such as those arising in biological sciences. [15] [56] Perhaps some of the most compelling examples of applications in which TDL consistently demonstrates its advantages over other competing methods are the victories of TDL in the D3R Grand Challenges, [57] [58] the discovery of SARS-CoV-2 evolution mechanisms, [59] [60] and the successful forecasting of SARS-CoV-2 variants BA.2, [61] BA.4 and BA.5, [47] about two months in advance.

See also

Related Research Articles

In mathematics, the term homology, originally introduced in algebraic topology, has three primary, closely-related usages. The most direct usage of the term is to take the homology of a chain complex, resulting in a sequence of abelian groups called homology groups. This operation, in turn, allows one to associate various named homologies or homology theories to various other types of mathematical objects. Lastly, since there are many homology theories for topological spaces that produce the same answer, one also often speaks of the homology of a topological space. There is also a related notion of the cohomology of a cochain complex, giving rise to various cohomology theories, in addition to the notion of the cohomology of a topological space.

In mathematics, and more specifically in algebraic topology and polyhedral combinatorics, the Euler characteristic is a topological invariant, a number that describes a topological space's shape or structure regardless of the way it is bent. It is commonly denoted by .

<span class="mw-page-title-main">Simplicial complex</span> Mathematical set

In mathematics, a simplicial complex is a structured set composed of points, line segments, triangles, and their n-dimensional counterparts, called simplices, such that all the faces and intersections of the elements are also included in the set. Simplicial complexes should not be confused with the more abstract notion of a simplicial set appearing in modern simplicial homotopy theory. The purely combinatorial counterpart to a simplicial complex is an abstract simplicial complex. To distinguish a simplicial complex from an abstract simplicial complex, the former is often called a geometric simplicial complex.

<span class="mw-page-title-main">Barycentric subdivision</span>

In mathematics, the barycentric subdivision is a standard way to subdivide a given simplex into smaller ones. Its extension on simplicial complexes is a canonical method to refine them. Therefore, the barycentric subdivision is an important tool in algebraic topology.

In algebraic topology, a homology sphere is an n-manifold X having the homology groups of an n-sphere, for some integer . That is,

<span class="mw-page-title-main">Abstract simplicial complex</span> Mathematical object

In combinatorics, an abstract simplicial complex (ASC), often called an abstract complex or just a complex, is a family of sets that is closed under taking subsets, i.e., every subset of a set in the family is also in the family. It is a purely combinatorial description of the geometric notion of a simplicial complex. For example, in a 2-dimensional simplicial complex, the sets in the family are the triangles, their edges, and their vertices.

In algebraic topology, simplicial homology is the sequence of homology groups of a simplicial complex. It formalizes the idea of the number of holes of a given dimension in the complex. This generalizes the number of connected components.

<span class="mw-page-title-main">Triangulation (topology)</span> Representation of mathematical space

In mathematics, triangulation describes the replacement of topological spaces with simplicial complexes by the choice of an appropriate homeomorphism. A space that admits such a homeomorphism is called a triangulable space. Triangulations can also be used to define a piecewise linear structure for a space, if one exists. Triangulation has various applications both in and outside of mathematics, for instance in algebraic topology, in complex analysis, and in modeling.

<span class="mw-page-title-main">Vietoris–Rips complex</span> Topological space formed from distances

In topology, the Vietoris–Rips complex, also called the Vietoris complex or Rips complex, is a way of forming a topological space from distances in a set of points. It is an abstract simplicial complex that can be defined from any metric space M and distance δ by forming a simplex for every finite set of points that has diameter at most δ. That is, it is a family of finite subsets of M, in which we think of a subset of k points as forming a (k − 1)-dimensional simplex (an edge for two points, a triangle for three points, a tetrahedron for four points, etc.); if a finite set S has the property that the distance between every pair of points in S is at most δ, then we include S as a simplex in the complex.

Discrete Morse theory is a combinatorial adaptation of Morse theory developed by Robin Forman. The theory has various practical applications in diverse fields of applied mathematics and computer science, such as configuration spaces, homology computation, denoising, mesh compression, and topological data analysis.

In applied mathematics, topological data analysis (TDA) is an approach to the analysis of datasets using techniques from topology. Extraction of information from datasets that are high-dimensional, incomplete and noisy is generally challenging. TDA provides a general framework to analyze such data in a manner that is insensitive to the particular metric chosen and provides dimensionality reduction and robustness to noise. Beyond this, it inherits functoriality, a fundamental concept of modern mathematics, from its topological nature, which allows it to adapt to new mathematical tools.

<span class="mw-page-title-main">Clique complex</span> Abstract simplicial complex describing a graphs cliques

Clique complexes, independence complexes, flag complexes, Whitney complexes and conformal hypergraphs are closely related mathematical objects in graph theory and geometric topology that each describe the cliques of an undirected graph.

Persistent homology is a method for computing topological features of a space at different spatial resolutions. More persistent features are detected over a wide range of spatial scales and are deemed more likely to represent true features of the underlying space rather than artifacts of sampling, noise, or particular choice of parameters.

In mathematics, the graph Fourier transform is a mathematical transform which eigendecomposes the Laplacian matrix of a graph into eigenvalues and eigenvectors. Analogously to the classical Fourier transform, the eigenvalues represent frequencies and eigenvectors form what is known as a graph Fourier basis.

In algebraic topology, homological connectivity is a property describing a topological space based on its homology groups.

Graph neural networks (GNN) are specialized artificial neural networks that are designed for tasks whose inputs are graphs.

The degree-Rips bifiltration is a simplicial filtration used in topological data analysis for analyzing the shape of point cloud data. It is a multiparameter extension of the Vietoris–Rips filtration that possesses greater stability to data outliers than single-parameter filtrations, and which is more amenable to practical computation than other multiparameter constructions. Introduced in 2015 by Lesnick and Wright, the degree-Rips bifiltration is a parameter-free and density-sensitive vehicle for performing persistent homology computations on point cloud data.

<span class="mw-page-title-main">Offset filtration</span>

The offset filtration is a growing sequence of metric balls used to detect the size and scale of topological features of a data set. The offset filtration commonly arises in persistent homology and the field of topological data analysis. Utilizing a union of balls to approximate the shape of geometric objects was first suggested by Frosini in 1992 in the context of submanifolds of Euclidean space. The construction was independently explored by Robins in 1998, and expanded to considering the collection of offsets indexed over a series of increasing scale parameters, in order to observe the stability of topological features with respect to attractors. Homological persistence as introduced in these papers by Frosini and Robins was subsequently formalized by Edelsbrunner et al. in their seminal 2002 paper Topological Persistence and Simplification. Since then, the offset filtration has become a primary example in the study of computational topology and data analysis.

In topological data analysis, a subdivision bifiltration is a collection of filtered simplicial complexes, typically built upon a set of data points in a metric space, that captures shape and density information about the underlying data set. The subdivision bifiltration relies on a natural filtration of the barycentric subdivision of a simplicial complex by flags of minimum dimension, which encodes density information about the metric space upon which the complex is built. The subdivision bifiltration was first introduced by Donald Sheehy in 2011 as part of his doctoral thesis as a discrete model of the multicover bifiltration, a continuous construction whose underlying framework dates back to the 1970s. In particular, Sheehy applied the construction to both the Vietoris-Rips and Čech filtrations, two common objects in the field of topological data analysis. Whereas single parameter filtrations are not robust with respect to outliers in the data, the subdivision-Rips and -Cech bifiltrations satisfy several desirable stability properties.

In persistent homology, a persistent Betti number is a multiscale analog of a Betti number that tracks the number of topological features that persist over multiple scale parameters in a filtration. Whereas the classical Betti number equals the rank of the homology group, the persistent Betti number is the rank of the persistent homology group. The concept of a persistent Betti number was introduced by Herbert Edelsbrunner, David Letscher, and Afra Zomorodian in the 2002 paper Topological Persistence and Simplification, one of the seminal papers in the field of persistent homology and topological data analysis. Applications of the persistent Betti number appear in a variety of fields including data analysis, machine learning, and physics.

References

  1. 1 2 3 4 5 6 7 8 9 10 11 12 Hajij, M.; Zamzmi, G.; Papamarkou, T.; Miolane, N.; Guzmán-Sáenz, A.; Ramamurthy, K. N.; Schaub, M. T. (2022), Topological deep learning: Going beyond graph data, arXiv: 2206.00606
  2. 1 2 Papillon, M.; Sanborn, S.; Hajij, M.; Miolane, N. (2023). "Architectures of topological deep learning: A survey on topological neural networks". arXiv: 2304.10031 [cs.LG].
  3. 1 2 Ebli, S.; Defferrard, M.; Spreemann, G. (2020), Simplicial neural networks, arXiv: 2010.03633
  4. Battiloro, C.; Testa, L.; Giusti, L.; Sardellitti, S.; Di Lorenzo, P.; Barbarossa, S. (2023), Generalized simplicial attention neural networks, arXiv: 2309.02138
  5. Yang, M.; Isufi, E. (2023), Convolutional learning on simplicial complexes, arXiv: 2301.11163
  6. Chen, Y.; Gel, Y. R.; Poor, H. V. (2022), "BScNets: Block Simplicial Complex Neural Networks", Proceedings of the AAAI Conference on Artificial Intelligence, 36 (6): 6333–6341, arXiv: 2112.06826 , doi:10.1609/aaai.v36i6.20583
  7. Uray, Martin; Giunti, Barbara; Kerber, Michael; Huber, Stefan (2024-10-01). "Topological Data Analysis in smart manufacturing: State of the art and future directions". Journal of Manufacturing Systems. 76: 75–91. arXiv: 2310.09319 . doi:10.1016/j.jmsy.2024.07.006. ISSN   0278-6125.
  8. 1 2 3 Hajij, M.; Istvan, K.; Zamzmi, G. (2020), Cell complex neural networks, arXiv: 2010.00743
  9. Bianchini, Monica; Scarselli, Franco (2014). "On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures". IEEE Transactions on Neural Networks and Learning Systems. 25 (8): 1553–1565. doi:10.1109/TNNLS.2013.2293637. ISSN   2162-237X.
  10. Naitzat, Gregory; Zhitnikov, Andrey; Lim, Lek-Heng (2020). "Topology of Deep Neural Networks" (PDF). Journal of Machine Learning Research. 21 (1): 184:7503–184:7542. ISSN   1532-4435.
  11. Birdal, Tolga; Lou, Aaron; Guibas, Leonidas J; Simsekli, Umut (2021). "Intrinsic Dimension, Persistent Homology and Generalization in Neural Networks". Advances in Neural Information Processing Systems. 34. Curran Associates, Inc.: 6776–6789. arXiv: 2111.13171 .
  12. Ballester, Rubén; Clemente, Xavier Arnal; Casacuberta, Carles; Madadi, Meysam; Corneanu, Ciprian A.; Escalera, Sergio (2024). "Predicting the generalization gap in neural networks using topological data analysis". Neurocomputing. 596: 127787. arXiv: 2203.12330 . doi:10.1016/j.neucom.2024.127787.
  13. Rieck, Bastian; Togninalli, Matteo; Bock, Christian; Moor, Michael; Horn, Max; Gumbsch, Thomas; Borgwardt, Karsten (2018-09-27). "Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology". International Conference on Learning Representations. arXiv: 1812.09764 . doi:10.3929/ethz-b-000327207.
  14. Dupuis, Benjamin; Deligiannidis, George; Simsekli, Umut (2023-07-03). "Generalization Bounds using Data-Dependent Fractal Dimensions". Proceedings of the 40th International Conference on Machine Learning. PMLR: 8922–8968.
  15. 1 2 3 Cang, Zixuan; Wei, Guo-Wei (2017-03-31). "TopologyNet: Topology based deep convolutional and multi-task neural networks for biomolecular property predictions". PLOS Computational Biology. 13 (7): e1005690. arXiv: 1704.00063 . Bibcode:2017PLSCB..13E5690C. doi: 10.1371/journal.pcbi.1005690 . PMID   28749969.{{cite journal}}: CS1 maint: date and year (link)
  16. Bronstein, Michael M.; Bruna, Joan; LeCun, Yann; Szlam, Arthur; Vandergheynst, Pierre (2017). "Geometric Deep Learning: Going beyond Euclidean data". IEEE Signal Processing Magazine. 34 (4): 18–42. arXiv: 1611.08097 . Bibcode:2017ISPM...34...18B. doi:10.1109/MSP.2017.2693418. ISSN   1053-5888.
  17. 1 2 Bodnar, Cristian; Frasca, Fabrizio; Wang, Yuguang; Otter, Nina; Montufar, Guido F.; Lió, Pietro; Bronstein, Michael (2021-07-01). "Weisfeiler and Lehman Go Topological: Message Passing Simplicial Networks". Proceedings of the 38th International Conference on Machine Learning. PMLR: 1026–1037. arXiv: 2103.03212 .
  18. 1 2 Bodnar, Cristian; Frasca, Fabrizio; Otter, Nina; Wang, Yuguang; Liò, Pietro; Montufar, Guido F; Bronstein, Michael (2021). "Weisfeiler and Lehman Go Cellular: CW Networks". Advances in Neural Information Processing Systems. 34. Curran Associates, Inc.: 2625–2640. arXiv: 2106.12575 .
  19. Carlsson, Gunnar (2009-01-29). "Topology and data". Bulletin of the American Mathematical Society. 46 (2): 255–308. doi: 10.1090/S0273-0979-09-01249-X . ISSN   0273-0979.
  20. Cang, Zixuan; Mu, Lin; Wu, Kedi; Opron, Kristopher; Xia, Kelin; Wei, Guo-Wei (2015). "A topological approach for protein classification". Computational and Mathematical Biophysics. 3 (1): 140–152. arXiv: 1510.00953 . doi: 10.1515/mlbmb-2015-0009 .{{cite journal}}: CS1 maint: date and year (link)
  21. Adcock, Aaron; Carlsson, Erik; Carlsson, Gunnar (2016). "The ring of algebraic functions on persistence bar codes". Homology, Homotopy and Applications. 18 (1): 381–402. arXiv: 1304.0530 . doi:10.4310/HHA.2016.v18.n1.a21.
  22. Adams, Henry; Emerson, Tegan; Kirby, Michael; Neville, Rachel; Peterson, Chris; Shipman, Patrick; Chepushtanova, Sofya; Hanson, Eric; Motta, Francis; Ziegelmeier, Lori (2017). "Persistence Images: A Stable Vector Representation of Persistent Homology". Journal of Machine Learning Research. 18 (8): 1–35. ISSN   1533-7928.
  23. Bubenik, Peter (2015). "Statistical Topological Data Analysis using Persistence Landscapes". Journal of Machine Learning Research. 16 (3): 77–102. ISSN   1533-7928.
  24. Kwitt, Roland; Huber, Stefan; Niethammer, Marc; Lin, Weili; Bauer, Ulrich (2015). "Statistical Topological Data Analysis - A Kernel Perspective". Advances in Neural Information Processing Systems. 28. Curran Associates, Inc.
  25. Carrière, Mathieu; Cuturi, Marco; Oudot, Steve (2017-07-17). "Sliced Wasserstein Kernel for Persistence Diagrams". Proceedings of the 34th International Conference on Machine Learning. PMLR: 664–673. arXiv: 1706.03358 .
  26. Kusano, Genki; Fukumizu, Kenji; Hiraoka, Yasuaki (2018). "Kernel Method for Persistence Diagrams via Kernel Embedding and Weight Factor". Journal of Machine Learning Research. 18 (189): 1–41. arXiv: 1706.03472 . ISSN   1533-7928.
  27. Le, Tam; Yamada, Makoto (2018). "Persistence Fisher Kernel: A Riemannian Manifold Kernel for Persistence Diagrams". Advances in Neural Information Processing Systems. 31. Curran Associates, Inc. arXiv: 1802.03569 .
  28. Maggs, Kelly; Hacker, Celia; Rieck, Bastian (2023-10-13). "Simplicial Representation Learning with Neural k-Forms". International Conference on Learning Representations. arXiv: 2312.08515 .
  29. Ramamurthy, K. N.; Guzmán-Sáenz, A.; Hajij, M. (2023), Topo-mlp: A simplicial network without message passing, pp. 1–5
  30. Hofer, Christoph; Kwitt, Roland; Niethammer, Marc; Uhl, Andreas (2017). "Deep Learning with Topological Signatures". Advances in Neural Information Processing Systems. 30. Curran Associates, Inc. arXiv: 1707.04041 .
  31. Hofer, Christoph D.; Kwitt, Roland; Niethammer, Marc (2019). "Learning Representations of Persistence Barcodes". Journal of Machine Learning Research. 20 (126): 1–45. ISSN   1533-7928.
  32. Carriere, Mathieu; Chazal, Frederic; Ike, Yuichi; Lacombe, Theo; Royer, Martin; Umeda, Yuhei (2020-06-03). "PersLay: A Neural Network Layer for Persistence Diagrams and New Graph Topological Signatures". Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics. PMLR: 2786–2796. arXiv: 1904.09378 .
  33. Kim, Kwangho; Kim, Jisu; Zaheer, Manzil; Kim, Joon; Chazal, Frederic; Wasserman, Larry (2020). "PLLay: Efficient Topological Layer based on Persistent Landscapes". Advances in Neural Information Processing Systems. 33. Curran Associates, Inc.: 15965–15977. arXiv: 2002.02778 .
  34. Gabrielsson, Rickard Brüel; Nelson, Bradley J.; Dwaraknath, Anjan; Skraba, Primoz (2020-06-03). "A Topology Layer for Machine Learning". Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics. PMLR: 1553–1563.
  35. 1 2 Horn, Max; Brouwer, Edward De; Moor, Michael; Moreau, Yves; Rieck, Bastian; Borgwardt, Karsten (2021-10-06). "Topological Graph Neural Networks". International Conference on Learning Representations.
  36. Hofer, Christoph; Graf, Florian; Rieck, Bastian; Niethammer, Marc; Kwitt, Roland (2020-11-21). "Graph Filtration Learning". Proceedings of the 37th International Conference on Machine Learning. PMLR: 4314–4323. arXiv: 1905.10996 .
  37. Immonen, Johanna; Souza, Amauri; Garg, Vikas (2023-12-15). "Going beyond persistent homology using persistent homology". Advances in Neural Information Processing Systems. 36: 63150–63173. arXiv: 2311.06152 .
  38. Chanillo, Sagun; Treves, Francois (1997). "On the lowest eigenvalue of the Hodge Laplacian". Journal of Differential Geometry. 45 (2): 273--287.{{cite journal}}: CS1 maint: date and year (link)
  39. Eckmann, Beno (1944). "Harmonische funktionen und randwertaufgaben in einem komplex". Commentarii Mathematici Helvetici. 17 (1): 240--255.{{cite journal}}: CS1 maint: date and year (link)
  40. Horak, Danijela; Jost, Jurgen (2013). "Spectra of combinatorial Laplace operators on simplicial complexes". Advances in Mathematics. 244: 303--336. doi:10.1016/j.aim.2013.05.007.{{cite journal}}: CS1 maint: date and year (link)
  41. Jones, Benjamin; Wei, Guo-Wei (2024). "Khovanov Laplacian and Khovanov Dirac for Knots and Links". ArXiv. arXiv: 2411.18841 . doi: 10.48550/arXiv.:2411.18841 .{{cite journal}}: CS1 maint: date and year (link)
  42. Chen, Jiahui; Zhao, Rundong; Tong, Yiying; Wei, Guo-Wei (2021). "Evolutionary de Rham-Hodge method". Discrete and Continuous Dynamical Systems. Series B. 26 (7): 3785–38215. arXiv: 1912.12388 . doi:10.3934/dcdsb.2020257.{{cite journal}}: CS1 maint: date and year (link)
  43. Wang, Rui; Nguyen, Duc; Wei, Guo-Wei (2020). "Persistent spectral graph". International Journal for Numerical Methods in Biomedical Engineering. 36 (9): e3376. arXiv: 1912.04135 . doi:10.1002/cnm.3376.{{cite journal}}: CS1 maint: date and year (link)
  44. Mémoli, Facundo; Wan, Zhengchao; Wang, Yusu (2022). "Persistent Laplacians: Properties, Algorithms and Implications". SIAM Journal on Mathematics of Data Science. 4 (2): 858--884. arXiv: 2012.02808 . doi:10.1137/21M1435471.{{cite journal}}: CS1 maint: date and year (link)
  45. Liu, Jian; Li, Jingyan; Wu, Jie (2023). "The algebraic stability for persistent Laplacians". ArXiv: 858--884. arXiv: 2302.03902 . doi:10.48550/arXiv.2302.03902.{{cite journal}}: CS1 maint: date and year (link)
  46. 1 2 Qiu, Yuchi; Wei, Guo-Wei (2023). "Persistent spectral theory-guided protein engineering". Nature Computational Science. 3 (2): 149–163. doi:10.1038/s43588-022-00394-y. PMID   37637776.{{cite journal}}: CS1 maint: date and year (link)
  47. 1 2 3 Chen, Jiahui; Qiu, Yuchi; Wang, Rui; Wei, Guo-Wei (2022). "Persistent Laplacian projected Omicron BA. 4 and BA. 5 to become new dominating variants". Computers in Biology and Medicine. 151: 106262. arXiv: 2205.00532 . doi:10.1016/j.compbiomed.2022.106262. PMID   36379191.{{cite journal}}: CS1 maint: date and year (link)
  48. Zhenyu, Meng; Kelin, Xia (2021). "Persistent spectral–based machine learning (PerSpect ML) for protein-ligand binding affinity prediction". Science Advances. 7 (19): eabc5329. doi:10.1126/sciadv.abc5329.{{cite journal}}: CS1 maint: date and year (link)
  49. 1 2 Wei, Xiaoqi; Wei, Guo-Wei (2025). "Persistent Topological Laplacians -- a Survey". Mathematics. 13 (2). arXiv: 2312.07563 . doi: 10.3390/math13020208 .{{cite journal}}: CS1 maint: date and year (link)
  50. Ameneyro, Bernardo; Maroulas, Vasileios; Siopsis, George (2024). "Quantum persistent homology". Journal of Applied and Computational Topology. 8: 1961–1980. arXiv: 2202.12965 . doi:10.1007/s41468-023-00160-7.{{cite journal}}: CS1 maint: date and year (link)
  51. Zhenyu, Meng; Kelin, Xia (2023). "Persistent Dirac for molecular representation". Scientific Report. 13: 11183. doi:10.1038/s41598-023-37853-z.{{cite journal}}: CS1 maint: date and year (link)
  52. Suwayyid, Faisal; Wei, Guo-Wei (2024). "Persistent Dirac of paths on digraphs and hypergraphs". Foundations of Data Science. 6 (2): 124--152. arXiv: 2202.12965 . doi:10.3934/fods.2024001.{{cite journal}}: CS1 maint: date and year (link)
  53. Battiloro, C.; Di Lorenzo, P.; Ribeiro, A. (September 2023), Parametric dictionary learning for topological signal representation, IEEE, pp. 1958–1962
  54. Wang, C.; Ma, N.; Wu, Z.; Zhang, J.; Yao, Y. (August 2022), Survey of Hypergraph Neural Networks and Its Application to Action Recognition, Springer Nature Switzerland, pp. 387–398
  55. Roddenberry, T. M.; Glaze, N.; Segarra, S. (July 2021), Principled simplicial neural networks for trajectory prediction, PMLR, pp. 9020–9029, arXiv: 2102.10058
  56. Cang, Zixuan; Mu, Lin; Wei, Guo-Wei (2018-01-08). "Representability of algebraic topology for biomolecules in machine learning based scoring and virtual screening". PLOS Computational Biology. 14 (1): e1005929. arXiv: 1708.08135 . doi: 10.1371/journal.pcbi.1005929 . PMID   28749969.{{cite journal}}: CS1 maint: date and year (link)
  57. Nguyen, Duc; Cang, Zixuan; Wu, Kedi; Wang, Menglun; Cao, Yin; Wei, Guo-Wei (2019-01-15). "Mathematical deep learning for pose and binding affinity prediction and ranking in D3R Grand Challenges". Journal of Computer-aided Molecular Design. 33: 71–82. arXiv: 1804.10647 . doi:10.1007/s10822-018-0146-6. PMID   30116918.{{cite journal}}: CS1 maint: date and year (link)
  58. Nguyen, Duc; Gao, Kaifu; Wang, Menglun; Wei, Guo-Wei (2020). "MathDL: mathematical deep learning for D3R Grand Challenge 4". Journal of Computer-aided Molecular Design. 34: 131–147. arXiv: 1909.07784 . doi:10.1007/s10822-019-00237-5. PMID   31734815.{{cite journal}}: CS1 maint: date and year (link)
  59. Chen, Jiahui; Wang, Rui; Wang, Menglun; Wei, Guo-Wei (2020). "Mutations strengthened SARS-CoV-2 infectivity". Journal of Molecular Biology. 432 (19): 5212–5226. arXiv: 2005.14669 . doi:10.1016/j.jmb.2020.07.009. PMID   32710986.{{cite journal}}: CS1 maint: date and year (link)
  60. Wang, Rui; Chen, Jiahui; Wei, Guo-Wei (2021). "Mechanisms of SARS-CoV-2 evolution revealing vaccine-resistant mutations in Europe and America". The Journal of Physical Chemistry Letters. 12 (49): 11850–11857. arXiv: 2110.04626 . doi:10.1021/acs.jpclett.1c03380. PMID   34873910.{{cite journal}}: CS1 maint: date and year (link)
  61. Chen, Jiahui; Wei, Guo-Wei (2022). "Omicron BA.2 (B.1.1.529.2): High Potential for Becoming the Next Dominant Variant". The Journal of Physical Chemistry Letters. 13 (17): 3840–3849. arXiv: 2202.05031 . doi:10.1021/acs.jpclett.2c00469. PMID   35467344.{{cite journal}}: CS1 maint: date and year (link)