Physics-informed neural networks

Last updated
Physics-informed neural networks for solving Navier-Stokes equations Physics-informed nerural networks.png
Physics-informed neural networks for solving Navier–Stokes equations

Physics-informed neural networks (PINNs), [1] also referred to as Theory-Trained Neural Networks (TTNs), [2] are a type of universal function approximators that can embed the knowledge of any physical laws that govern a given data-set in the learning process, and can be described by partial differential equations (PDEs). They overcome the low data availability of some biological and engineering systems that makes most state-of-the-art machine learning techniques lack robustness, rendering them ineffective in these scenarios. [1] The prior knowledge of general physical laws acts in the training of neural networks (NNs) as a regularization agent that limits the space of admissible solutions, increasing the correctness of the function approximation. This way, embedding this prior information into a neural network results in enhancing the information content of the available data, facilitating the learning algorithm to capture the right solution and to generalize well even with a low amount of training examples.

Contents

Function approximation

Most of the physical laws that govern the dynamics of a system can be described by partial differential equations. For example, the Navier–Stokes equations [3] are a set of partial differential equations derived from the conservation laws (i.e., conservation of mass, momentum, and energy) that govern fluid mechanics. The solution of the Navier–Stokes equations with appropriate initial and boundary conditions allows the quantification of flow dynamics in a precisely defined geometry. However, these equations cannot be solved exactly and therefore numerical methods must be used (such as finite differences, finite elements and finite volumes). In this setting, these governing equations must be solved while accounting for prior assumptions, linearization, and adequate time and space discretization.

Recently, solving the governing partial differential equations of physical phenomena using deep learning has emerged as a new field of scientific machine learning (SciML), leveraging the universal approximation theorem [4] and high expressivity of neural networks. In general, deep neural networks could approximate any high-dimensional function given that sufficient training data are supplied. [5] However, such networks do not consider the physical characteristics underlying the problem, and the level of approximation accuracy provided by them is still heavily dependent on careful specifications of the problem geometry as well as the initial and boundary conditions. Without this preliminary information, the solution is not unique and may lose physical correctness. On the other hand, physics-informed neural networks (PINNs) leverage governing physical equations in neural network training. Namely, PINNs are designed to be trained to satisfy the given training data as well as the imposed governing equations. In this fashion, a neural network can be guided with training data that do not necessarily need to be large and complete. [5] Potentially, an accurate solution of partial differential equations can be found without knowing the boundary conditions. [6] Therefore, with some knowledge about the physical characteristics of the problem and some form of training data (even sparse and incomplete), PINN may be used for finding an optimal solution with high fidelity.

PINNs allow for addressing a wide range of problems in computational science and represent a pioneering technology leading to the development of new classes of numerical solvers for PDEs. PINNs can be thought of as a meshfree alternative to traditional approaches (e.g., CFD for fluid dynamics), and new data-driven approaches for model inversion and system identification. [7] Notably, the trained PINN network can be used for predicting the values on simulation grids of different resolutions without the need to be retrained. [8] In addition, they allow for exploiting automatic differentiation (AD) [9] to compute the required derivatives in the partial differential equations, a new class of differentiation techniques widely used to derive neural networks assessed to be superior to numerical or symbolic differentiation.

Modeling and computation

A general nonlinear partial differential equation can be:

where denotes the solution, is a nonlinear operator parameterized by , and is a subset of . This general form of governing equations summarizes a wide range of problems in mathematical physics, such as conservative laws, diffusion process, advection-diffusion systems, and kinetic equations. Given noisy measurements of a generic dynamic system described by the equation above, PINNs can be designed to solve two classes of problems:

Data-driven solution of partial differential equations

The data-driven solution of PDE [1] computes the hidden state of the system given boundary data and/or measurements , and fixed model parameters . We solve:

.

By defining the residual as

,

and approximating by a deep neural network. This network can be differentiated using automatic differentiation. The parameters of and can be then learned by minimizing the following loss function :

.

Where is the error between the PINN and the set of boundary conditions and measured data on the set of points where the boundary conditions and data are defined, and is the mean-squared error of the residual function. This second term encourages the PINN to learn the structural information expressed by the partial differential equation during the training process.

This approach has been used to yield computationally efficient physics-informed surrogate models with applications in the forecasting of physical processes, model predictive control, multi-physics and multi-scale modeling, and simulation. [10] It has been shown to converge to the solution of the PDE. [11]

Data-driven discovery of partial differential equations

Given noisy and incomplete measurements of the state of the system, the data-driven discovery of PDE [7] results in computing the unknown state and learning model parameters that best describe the observed data and it reads as follows:

.

By defining as

,

and approximating by a deep neural network, results in a PINN. This network can be derived using automatic differentiation. The parameters of and , together with the parameter of the differential operator can be then learned by minimizing the following loss function :

.

Where , with and state solutions and measurements at sparse location , respectively and residual function. This second term requires the structured information represented by the partial differential equations to be satisfied in the training process.

This strategy allows for discovering dynamic models described by nonlinear PDEs assembling computationally efficient and fully differentiable surrogate models that may find application in predictive forecasting, control, and data assimilation. [12] [13] [14] [15]

Physics-informed neural networks for piece-wise function approximation

PINN is unable to approximate PDEs that have strong non-linearity or sharp gradients that commonly occur in practical fluid flow problems. Piece-wise approximation has been an old practice in the field of numerical approximation. With the capability of approximating strong non-linearity extremely light weight PINNs are used to solve PDEs in much larger discrete subdomains that increases accuracy substantially and decreases computational load as well. [16] [17] DPINN (Distributed physics-informed neural networks) and DPIELM (Distributed physics-informed extreme learning machines) are generalizable space-time domain discretization for better approximation. [16] DPIELM is an extremely fast and lightweight approximator with competitive accuracy. Domain scaling on the top has a special effect. [17] Another school of thought is discretization for parallel computation to leverage usage of available computational resources.

XPINNs [18] is a generalized space-time domain decomposition approach for the physics-informed neural networks (PINNs) to solve nonlinear partial differential equations on arbitrary complex-geometry domains. The XPINNs further pushes the boundaries of both PINNs as well as Conservative PINNs (cPINNs), [19] which is a spatial domain decomposition approach in the PINN framework tailored to conservation laws. Compared to PINN, the XPINN method has large representation and parallelization capacity due to the inherent property of deployment of multiple neural networks in the smaller subdomains. Unlike cPINN, XPINN can be extended to any type of PDEs. Moreover, the domain can be decomposed in any arbitrary way (in space and time), which is not possible in cPINN. Thus, XPINN offers both space and time parallelization, thereby reducing the training cost more effectively. The XPINN is particularly effective for the large-scale problems (involving large data set) as well as for the high-dimensional problems where single network based PINN is not adequate. The rigorous bounds on the errors resulting from the approximation of the nonlinear PDEs (incompressible Navier–Stokes equations) with PINNs and XPINNs are proved. [15] However, DPINN debunks the use of residual (flux) matching at the domain interfaces as they hardly seem to improve the optimization. [17]

Physics-informed neural networks and functional interpolation

X-TFC framework scheme for PDE solution learning Xtfc scheme.png
X-TFC framework scheme for PDE solution learning

In the PINN framework, initial and boundary conditions are not analytically satisfied, thus they need to be included in the loss function of the network to be simultaneously learned with the differential equation (DE) unknown functions. Having competing objectives during the network's training can lead to unbalanced gradients while using gradient-based techniques, which causes PINNs to often struggle to accurately learn the underlying DE solution. This drawback is overcome by using functional interpolation techniques such as the Theory of Functional Connections (TFC)'s constrained expression, in the Deep-TFC framework, which reduces the solution search space of constrained problems to the subspace of neural network that analytically satisfies the constraints. [20] A further improvement of PINN and functional interpolation approach is given by the Extreme Theory of Functional Connections (X-TFC) framework, where a single-layer Neural Network and the extreme learning machine training algorithm are employed. [21] X-TFC allows to improve the accuracy and performance of regular PINNs, and its robustness and reliability are proved for stiff problems, optimal control, aerospace, and rarefied gas dynamics applications. [22] [23] [24]

Physics-informed PointNet (PIPN) for multiple sets of irregular geometries

Regular PINNs are only able to obtain the solution of a forward or inverse problem on a single geometry. It means that for any new geometry (computational domain), one must retrain a PINN. This limitation of regular PINNs imposes high computational costs, specifically for a comprehensive investigation of geometric parameters in industrial designs. Physics-informed PointNet (PIPN) [25] is fundamentally the result of a combination of PINN's loss function with PointNet. [26] In fact, instead of using a simple fully connected neural network, PIPN uses PointNet as the core of its neural network. PointNet has been primarily designed for deep learning of 3D object classification and segmentation by the research group of Leonidas J. Guibas. PointNet extracts geometric features of input computational domains in PIPN. Thus, PIPN is able to solve governing equations on multiple computational domains (rather than only a single domain) with irregular geometries, simultaneously. The effectiveness of PIPN has been shown for incompressible flow, heat transfer and linear elasticity. [25] [27]

Physics-informed neural networks (PINNs) for inverse computations

Physics-informed neural networks (PINNs) have proven particularly effective in solving inverse problems within differential equations, [28] demonstrating their applicability across science, engineering, and economics. They have shown useful for solving inverse problems in a variety of fields, including nano-optics, [29] topology optimization/characterization, [30] multiphase flow in porous media, [31] [32] and high-speed fluid flow. [33] [13] PINNs have demonstrated flexibility when dealing with noisy and uncertain observation datasets. They also demonstrated clear advantages in the inverse calculation of parameters for multi-fidelity datasets, meaning datasets with different quality, quantity, and types of observations. Uncertainties in calculations can be evaluated using ensemble-based or Bayesian-based calculations. [34]

Physics-informed neural networks (PINNs) with backward stochastic differential equation

Deep backward stochastic differential equation method is a numerical method that combines deep learning with Backward stochastic differential equation (BSDE) to solve high-dimensional problems in financial mathematics. By leveraging the powerful function approximation capabilities of deep neural networks, deep BSDE addresses the computational challenges faced by traditional numerical methods like finite difference methods or Monte Carlo simulations, which struggle with the curse of dimensionality. Deep BSDE methods use neural networks to approximate solutions of high-dimensional partial differential equations (PDEs), effectively reducing the computational burden. Additionally, integrating Physics-informed neural networks (PINNs) into the deep BSDE framework enhances its capability by embedding the underlying physical laws into the neural network architecture, ensuring solutions adhere to governing stochastic differential equations, resulting in more accurate and reliable solutions. [35]

Physics-informed neural networks for biology

An extension or adaptation of PINNs are Biologically-informed neural networks (BINNs). BINNs introduce two key adaptations to the typical PINN framework: (i) the mechanistic terms of the governing PDE are replaced by neural networks, and (ii) the loss function is modified to include , a term used to incorporate domain-specific knowledge that helps enforce biological applicability. For (i), this adaptation has the advantage of relaxing the need to specify the governing differential equation a priori, either explicitly or by using a library of candidate terms. Additionally, this approach circumvents the potential issue of misspecifying regularization terms in stricter theory-informed cases. [36] [37]

A natural example of BINNs can be found in cell dynamics, where the cell density is governed by a reaction-diffusion equation with diffusion and growth functions and , respectively:

In this case, a component of could be for , which penalizes values of that fall outside a biologically relevant diffusion range defined by . Furthermore, the BINN architecture, when utilizing multilayer-perceptrons (MLPs), would function as follows: an MLP is used to construct from model inputs , serving as a surrogate model for the cell density . This surrogate is then fed into the two additional MLPs, and , which model the diffusion and growth functions. Automatic differentiation can then be applied to compute the necessary derivatives of , and to form the governing reaction-diffusion equation. [36]

Note that since is a surrogate for the cell density, it may contain errors, particularly in regions where the PDE is not fully satisfied. Therefore, the reaction-diffusion equation may be solved numerically, for instance using a method-of-lines approach approach.

Limitations

Translation and discontinuous behavior are hard to approximate using PINNs. [17] They fail when solving differential equations with slight advective dominance and hence asymptotic behaviour causes the method to fail. Such PDEs could be solved by scaling variables. [17] This difficulty in training of PINNs in advection-dominated PDEs can be explained by the Kolmogorov n–width of the solution. [38] They also fail to solve a system of dynamical systems and hence have not been a success in solving chaotic equations. [39] One of the reasons behind the failure of regular PINNs is soft-constraining of Dirichlet and Neumann boundary conditions which pose a multi-objective optimization problem which requires manually weighing the loss terms to be able to optimize. [17] More generally, posing the solution of a PDE as an optimization problem brings with it all the problems that are faced in the world of optimization, the major one being getting stuck in local optima. [17] [40]

Related Research Articles

<span class="mw-page-title-main">Partial differential equation</span> Type of differential equation

In mathematics, a partial differential equation (PDE) is an equation which computes a function between various partial derivatives of a multivariable function.

The Hamilton-Jacobi-Bellman (HJB) equation is a nonlinear partial differential equation that provides necessary and sufficient conditions for optimality of a control with respect to a loss function. Its solution is the value function of the optimal control problem which, once known, can be used to obtain the optimal control by taking the maximizer of the Hamiltonian involved in the HJB equation.

In mathematics, the Helmholtz equation is the eigenvalue problem for the Laplace operator. It corresponds to the elliptic partial differential equation: where 2 is the Laplace operator, k2 is the eigenvalue, and f is the (eigen)function. When the equation is applied to waves, k is known as the wave number. The Helmholtz equation has a variety of applications in physics and other sciences, including the wave equation, the diffusion equation, and the Schrödinger equation for a free particle.

In mathematics, a differential equation is an equation that relates one or more unknown functions and their derivatives. In applications, the functions generally represent physical quantities, the derivatives represent their rates of change, and the differential equation defines a relationship between the two. Such relations are common; therefore, differential equations play a prominent role in many disciplines including engineering, physics, economics, and biology.

<span class="mw-page-title-main">Kadomtsev–Petviashvili equation</span>

In mathematics and physics, the Kadomtsev–Petviashvili equation is a partial differential equation to describe nonlinear wave motion. Named after Boris Borisovich Kadomtsev and Vladimir Iosifovich Petviashvili, the KP equation is usually written as where . The above form shows that the KP equation is a generalization to two spatial dimensions, x and y, of the one-dimensional Korteweg–de Vries (KdV) equation. To be physically meaningful, the wave propagation direction has to be not-too-far from the x direction, i.e. with only slow variations of solutions in the y direction.

Moving least squares is a method of reconstructing continuous functions from a set of unorganized point samples via the calculation of a weighted least squares measure biased towards the region around the point at which the reconstructed value is requested.

A parabolic partial differential equation is a type of partial differential equation (PDE). Parabolic PDEs are used to describe a wide variety of time-dependent phenomena in, i.a., engineering science, quantum mechanics and financial mathematics. Examples include the heat equation, time-dependent Schrödinger equation and the Black–Scholes equation.

Dispersionless limits of integrable partial differential equations (PDE) arise in various problems of mathematics and physics and have been intensively studied in recent literature. They typically arise when considering slowly modulated long waves of an integrable dispersive PDE system.

<span class="mw-page-title-main">Finite element method</span> Numerical method for solving physical or engineering problems

The finite element method (FEM) is a popular method for numerically solving differential equations arising in engineering and mathematical modeling. Typical problem areas of interest include the traditional fields of structural analysis, heat transfer, fluid flow, mass transport, and electromagnetic potential.

Fluid motion is governed by the Navier–Stokes equations, a set of coupled and nonlinear partial differential equations derived from the basic laws of conservation of mass, momentum and energy. The unknowns are usually the flow velocity, the pressure and density and temperature. The analytical solution of this equation is impossible hence scientists resort to laboratory experiments in such situations. The answers delivered are, however, usually qualitatively different since dynamical and geometric similitude are difficult to enforce simultaneously between the lab experiment and the prototype. Furthermore, the design and construction of these experiments can be difficult, particularly for stratified rotating flows. Computational fluid dynamics (CFD) is an additional tool in the arsenal of scientists. In its early days CFD was often controversial, as it involved additional approximation to the governing equations and raised additional (legitimate) issues. Nowadays CFD is an established discipline alongside theoretical and experimental methods. This position is in large part due to the exponential growth of computer power which has allowed us to tackle ever larger and more complex problems.

The closest point method (CPM) is an embedding method for solving partial differential equations on surfaces. The closest point method uses standard numerical approaches such as finite differences, finite element or spectral methods in order to solve the embedding partial differential equation (PDE) which is equal to the original PDE on the surface. The solution is computed in a band surrounding the surface in order to be computationally efficient. In order to extend the data off the surface, the closest point method uses a closest point representation. This representation extends function values to be constant along directions normal to the surface.

The Harrow–Hassidim–Lloyd algorithm or HHL algorithm is a quantum algorithm for numerically solving a system of linear equations, designed by Aram Harrow, Avinatan Hassidim, and Seth Lloyd. The algorithm estimates the result of a scalar measurement on the solution vector to a given linear system of equations.

In mathematics, the walk-on-spheres method (WoS) is a numerical probabilistic algorithm, or Monte-Carlo method, used mainly in order to approximate the solutions of some specific boundary value problem for partial differential equations (PDEs). The WoS method was first introduced by Mervin E. Muller in 1956 to solve Laplace's equation, and was since then generalized to other problems.

Applying classical methods of machine learning to the study of quantum systems is the focus of an emergent area of physics research. A basic example of this is quantum state tomography, where a quantum state is learned from measurement. Other examples include learning Hamiltonians, learning quantum phase transitions, and automatically generating new quantum experiments. Classical machine learning is effective at processing large amounts of experimental or calculated data in order to characterize an unknown quantum system, making its application useful in contexts including quantum information theory, quantum technologies development, and computational materials design. In this context, it can be used for example as a tool to interpolate pre-calculated interatomic potentials or directly solving the Schrödinger equation with a variational method.

PDE-constrained optimization is a subset of mathematical optimization where at least one of the constraints may be expressed as a partial differential equation. Typical domains where these problems arise include aerodynamics, computational fluid dynamics, image segmentation, and inverse problems. A standard formulation of PDE-constrained optimization encountered in a number of disciplines is given by:where is the control variable and is the squared Euclidean norm and is not a norm itself. Closed-form solutions are generally unavailable for PDE-constrained optimization problems, necessitating the development of numerical methods.

Probabilistic numerics is an active field of study at the intersection of applied mathematics, statistics, and machine learning centering on the concept of uncertainty in computation. In probabilistic numerics, tasks in numerical analysis such as finding numerical solutions for integration, linear algebra, optimization and simulation and differential equations are seen as problems of statistical, probabilistic, or Bayesian inference.

The frequency principle/spectral bias is a phenomenon observed in the study of Artificial Neural Networks(ANNs), specifically deep neural networks(DNNs). It describes the tendency of deep neural networks to fit target functions from low to high frequencies during the training process.

Neural operators are a class of deep learning architectures designed to learn maps between infinite-dimensional function spaces. Neural operators represent an extension of traditional artificial neural networks, marking a departure from the typical focus on learning mappings between finite-dimensional Euclidean spaces or finite sets. Neural operators directly learn operators between function spaces; they can receive input functions, and the output function can be evaluated at any discretization.

<span class="mw-page-title-main">Deep backward stochastic differential equation method</span>

Deep backward stochastic differential equation method is a numerical method that combines deep learning with Backward stochastic differential equation (BSDE). This method is particularly useful for solving high-dimensional problems in financial derivatives pricing and risk management. By leveraging the powerful function approximation capabilities of deep neural networks, deep BSDE addresses the computational challenges faced by traditional numerical methods in high-dimensional settings.

In machine learning, a neural differential equation is a differential equation whose right-hand side is parametrized by the weights θ of a neural network. In particular, a neural ordinary differential equation (neural ODE) is an ordinary differential equation of the form

References

  1. 1 2 3 Raissi, Maziar; Perdikaris, Paris; Karniadakis, George Em (2017-11-28). "Physics Informed Deep Learning (Part I): Data-driven Solutions of Nonlinear Partial Differential Equations". arXiv: 1711.10561 [cs.AI].
  2. Torabi Rad, M.; Viardin, A.; Schmitz, G.J.; Apel, M. (2020-03-01). "Theory-training deep neural networks for an alloy solidification benchmark problem". Computational Materials Science. 18. arXiv: 1912.09800 . doi:10.1016/j.commatsci.2020.109687. ISSN   0893-6080.
  3. Batchelor, G. K. (2000). An introduction to fluid dynamics (2nd pbk. ed.). Cambridge, U.K.: Cambridge University Press. ISBN   978-0-521-66396-0.
  4. Hornik, Kurt; Tinchcombe, Maxwell; White, Halbert (1989-01-01). "Multilayer feedforward networks are universal approximators". Neural Networks. 2 (5): 359–366. doi:10.1016/0893-6080(89)90020-8. ISSN   0893-6080. S2CID   2757547.
  5. 1 2 Arzani, Amirhossein; Dawson, Scott T. M. (2021). "Data-driven cardiovascular flow modelling: examples and opportunities". Journal of the Royal Society Interface. 18 (175): 20200802. arXiv: 2010.00131 . doi:10.1098/rsif.2020.0802. PMC   8086862 . PMID   33561376.
  6. Arzani, Amirhossein; Wang, Jian-Xun; D'Souza, Roshan M. (2021-06-07). "Uncovering near-wall blood flow from sparse data with physics-informed neural networks". Physics of Fluids. 33 (7): 071905. arXiv: 2104.08249 . Bibcode:2021PhFl...33g1905A. doi:10.1063/5.0055600. S2CID   233289904.
  7. 1 2 Raissi, Maziar; Perdikaris, Paris; Karniadakis, George Em (2019-02-01). "Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations". Journal of Computational Physics. 378: 686–707. Bibcode:2019JCoPh.378..686R. doi: 10.1016/j.jcp.2018.10.045 . ISSN   0021-9991. OSTI   1595805. S2CID   57379996.
  8. Markidis, Stefano (2021-03-11). "Physics-Informed Deep-Learning for Scientific Computing". arXiv: 2103.09655 [math.NA].
  9. Baydin, Atilim Gunes; Pearlmutter, Barak A.; Radul, Alexey Andreyevich; Siskind, Jeffrey Mark (2018-02-05). "Automatic differentiation in machine learning: a survey". arXiv: 1502.05767 [cs.SC].
  10. Raissi, Maziar; Yazdani, Alireza; Karniadakis, George Em (2018-08-13). "Hidden Fluid Mechanics: A Navier–Stokes Informed Deep Learning Framework for Assimilating Flow Visualization Data". arXiv: 1808.04327 [cs.CE].
  11. Doumèche, Nathan; Biau, Gérard; Boyer, Claire (2023-05-02). "Convergence and error analysis of PINNs". arXiv: 2305.01240 [math.ST].
  12. Fu, Jinlong; Xiao, Dunhui; Fu, Rui; Li, Chenfeng; Zhu, Chuanhua; Arcucci, Rossella; Navon, Ionel M. (February 2023). "Physics-data combined machine learning for parametric reduced-order modelling of nonlinear dynamical systems in small-data regimes". Computer Methods in Applied Mechanics and Engineering. 404: 115771. Bibcode:2023CMAME.404k5771F. doi:10.1016/j.cma.2022.115771. S2CID   254397985.
  13. 1 2 Raissi, Maziar; Yazdani, Alireza; Karniadakis, George Em (2020-02-28). "Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations". Science. 367 (6481): 1026–1030. Bibcode:2020Sci...367.1026R. doi:10.1126/science.aaw4741. ISSN   0036-8075. PMC   7219083 . PMID   32001523.
  14. Mishra, Siddhartha; Molinaro, Roberto (2021-04-01). "Estimates on the generalization error of Physics Informed Neural Networks (PINNs) for approximating a class of inverse problems for PDEs". arXiv: 2007.01138 [math.NA].
  15. 1 2 Ryck, Tim De; Jagtap, Ameya D.; Mishra, Siddhartha (2022). "Error estimates for physics informed neural networks approximating the Navier–Stokes equations". arXiv: 2203.09346 [math.NA].
  16. 1 2 Dwivedi, Vikas; Parashar, Nishant; Srinivasan, Balaji (2021-01-08). "Distributed learning machines for solving forward and inverse problems in partial differential equations". Neurocomputing. 420: 299–316. arXiv: 1907.08967 . doi:10.1016/j.neucom.2020.09.006. ISSN   0925-2312. S2CID   225014697.
  17. 1 2 3 4 5 6 7 Rout, Siddharth (2019). "Numerical Approximation in CFD Problems Using Physics Informed Machine Learning". arXiv: 2111.02987 [cs.LG]. Master's Thesis, Indian Institute of Technology Madras, July 2019.
  18. Jagtap, Ameya D.; Karniadakis, George Em (2020). "Extended physics-informed neural networks (xpinns): A generalized space-time domain decomposition based deep learning framework for nonlinear partial differential equations". Communications in Computational Physics. 28 (5): 2002–2041. Bibcode:2020CCoPh..28.2002K. doi:10.4208/cicp.OA-2020-0164. OSTI   2282003. S2CID   229083388.
  19. Jagtap, Ameya D.; Kharazmi, Ehsan; Karniadakis, George Em (2020). "Conservative Physics-Informed Neural Networks on Discrete Domains for Conservation Laws: Applications to forward and inverse problems". Computer Methods in Applied Mechanics and Engineering. 365: 113028. Bibcode:2020CMAME.365k3028J. doi: 10.1016/j.cma.2020.113028 . S2CID   216333271.
  20. Leake, Carl; Mortari, Daniele (12 March 2020). "Deep Theory of Functional Connections: A New Method for Estimating the Solutions of Partial Differential Equations". Machine Learning and Knowledge Extraction. 2 (1): 37–55. doi: 10.3390/make2010004 . PMC   7259480 . PMID   32478283.
  21. Schiassi, Enrico; Furfaro, Roberto; Leake, Carl; De Florio, Mario; Johnston, Hunter; Mortari, Daniele (October 2021). "Extreme theory of functional connections: A fast physics-informed neural network method for solving ordinary and partial differential equations". Neurocomputing. 457: 334–356. doi: 10.1016/j.neucom.2021.06.015 . S2CID   236290147.
  22. Schiassi, Enrico; De Florio, Mario; Ganapol, Barry D.; Picca, Paolo; Furfaro, Roberto (March 2022). "Physics-informed neural networks for the point kinetics equations for nuclear reactor dynamics". Annals of Nuclear Energy. 167: 108833. Bibcode:2022AnNuE.16708833S. doi:10.1016/j.anucene.2021.108833. S2CID   244913655.
  23. Schiassi, Enrico; D’Ambrosio, Andrea; Drozd, Kristofer; Curti, Fabio; Furfaro, Roberto (4 January 2022). "Physics-Informed Neural Networks for Optimal Planar Orbit Transfers". Journal of Spacecraft and Rockets. 59 (3): 834–849. Bibcode:2022JSpRo..59..834S. doi:10.2514/1.A35138. S2CID   245725265.
  24. De Florio, Mario; Schiassi, Enrico; Ganapol, Barry D.; Furfaro, Roberto (April 2021). "Physics-informed neural networks for rarefied-gas dynamics: Thermal creep flow in the Bhatnagar–Gross–Krook approximation". Physics of Fluids. 33 (4): 047110. Bibcode:2021PhFl...33d7110D. doi:10.1063/5.0046181. S2CID   234852360.
  25. 1 2 Kashefi, Ali; Mukerji, Tapan (2022). "Physics-informed PointNet: A deep learning solver for steady-state incompressible flows and thermal fields on multiple sets of irregular geometries". Journal of Computational Physics. 468: 111510. arXiv: 2202.05476 . Bibcode:2022JCoPh.46811510K. doi:10.1016/j.jcp.2022.111510. S2CID   246823515.
  26. Qi, Charles; Su, Hao; Mo, Kaichun; Guibas, Leonidas (2017). "Pointnet: Deep learning on point sets for 3d classification and segmentation" (PDF). Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition: 652–660. arXiv: 1612.00593 .
  27. Kashefi, Ali; Mukerji, Tapan (2023). "Physics-informed PointNet: On how many irregular geometries can it solve an inverse problem simultaneously? Application to linear elasticity". arXiv: 2303.13634 [cs.LG].
  28. Thanasutives, Pongpisit; Morita, Takashi; Numao, Masayuki; Fukui, Ken-ichi (2023-02-01). "Noise-aware physics-informed machine learning for robust PDE discovery". Machine Learning: Science and Technology. 4 (1): 015009. arXiv: 2206.12901 . doi:10.1088/2632-2153/acb1f0. ISSN   2632-2153.
  29. "Optica Publishing Group". opg.optica.org. Retrieved 2024-03-18.
  30. Zhang, Enrui; Dao, Ming; Karniadakis, George Em; Suresh, Subra (2022-02-18). "Analyses of internal structures and defects in materials using physics-informed neural networks". Science Advances. 8 (7): eabk0644. Bibcode:2022SciA....8..644Z. doi:10.1126/sciadv.abk0644. ISSN   2375-2548. PMC   8849303 . PMID   35171670.
  31. Serebrennikova, Alexandra; Teubler, Raimund; Hoffellner, Lisa; Leitner, Erich; Hirn, Ulrich; Zojer, Karin (2022-12-01). "Transport of Organic Volatiles through Paper: Physics-Informed Neural Networks for Solving Inverse and Forward Problems". Transport in Porous Media. 145 (3): 589–612. Bibcode:2022TPMed.145..589S. doi: 10.1007/s11242-022-01864-7 . ISSN   1573-1634.
  32. Abbasi, Jassem; Andersen, Pål Østebø (2024-01-01). "Application of Physics-Informed Neural Networks for Estimation of Saturation Functions from Countercurrent Spontaneous Imbibition Tests". SPE Journal. 29 (4): 1710–1729. doi:10.2118/218402-PA. ISSN   1086-055X.
  33. Jagtap, Ameya D.; Mao, Zhiping; Adams, Nikolaus; Karniadakis, George Em (October 2022). "Physics-informed neural networks for inverse problems in supersonic flows". Journal of Computational Physics. 466: 111402. arXiv: 2202.11821 . Bibcode:2022JCoPh.46611402J. doi:10.1016/j.jcp.2022.111402. ISSN   0021-9991.
  34. Yang, Liu; Meng, Xuhui; Karniadakis, George Em (January 2021). "B-PINNs: Bayesian physics-informed neural networks for forward and inverse PDE problems with noisy data". Journal of Computational Physics. 425: 109913. arXiv: 2003.06097 . Bibcode:2021JCoPh.42509913Y. doi:10.1016/j.jcp.2020.109913. ISSN   0021-9991.
  35. Han, J.; Jentzen, A.; E, W. (2018). "Solving high-dimensional partial differential equations using deep learning". Proceedings of the National Academy of Sciences. 115 (34): 8505–8510.
  36. 1 2 Lagergren JH, Nardini JT, Baker RE, Simpson MJ, Flores KB (2020). "Biologically-informed neural networks guide mechanistic modeling from sparse experimental data." PLoS Comput Biol 16(12): e1008462. https://doi.org/10.1371/journal.pcbi.1008462
  37. Nardini JT (2024). "Forecasting and Predicting Stochastic Agent-Based Model Data with Biologically-Informed Neural Networks." Bull Math Biol 86, 130. https://doi.org/10.1007/s11538-024-01357-2
  38. Mojgani, Rambod; Balajewicz, Maciej; Hassanzadeh, Pedram (February 2023). "Kolmogorov n–width and Lagrangian physics-informed neural networks: A causality-conforming manifold for convection-dominated PDEs". Computer Methods in Applied Mechanics and Engineering. 404: 115810. arXiv: 2205.02902 . doi:10.1016/j.cma.2022.115810.
  39. Eric Aislan Antonelo; Camponogara, Eduardo; Laio Oriel Seman; Eduardo Rehbein de Souza; Jordanou, Jean P.; Hubner, Jomi F. (2024). "Physics-informed neural nets for control of dynamical systems". Neurocomputing. 579. arXiv: 2104.02556 . doi:10.1016/j.neucom.2024.127419.
  40. Wang, Sifan; Teng, Yujun; Perdikaris, Paris (2020-01-13). "Understanding and mitigating gradient pathologies in physics-informed neural networks". arXiv: 2001.04536 [cs.LG].