Draft:Physics-Enhanced Machine Learning
Physics-Enhanced Machine Learning
Physics-Enhanced Machine Learning (PEML) is a subfield of machine learning that combines physics-based modelling and data-driven algorithms, and aims to combine the complimentary strengths of physical knowledge and machine learning algorithms to overcome the limitations of using either approach alone.[1] PEML incorporates known physical laws and domain knowledge into the machine learning process so that models not only fit observed data patterns, but also respect the governing physics equations of the model.
This concept is closely related to scientific machine learning (SciML)[2], Physics-Informed Machine Learning (PIML), Physics-Enhanced Artificial Intelligence (PEAI) and Physics-Guided Machine Learning, but is distinct in its emphasis on enhancing predictive capability through the use of physics-based components in hybrid or constrained models.
Background and Motivation
The motivation for PEML arose from challenges often encountered in engineering and real-world problems: The volume of useful data is generally limited; predictions obtained from modelling can make accurate, but physically implausible predictions that violate known physics; uncertainties could not easily be dealt with; and traditional machine learning models are not very explainable or interpretable.[3] By infusing physics into machine learning, it ensures better generalisation to unseen conditions, improved physical consistency of predictions, and a greater explainability of the learnt models. This approach has gained traction since 2019, where techniques have been used to encompass strategies where "prior physics knowledge is embedded to the learner" [1], particularly in dealing with dynamical systems in engineering.
Early examples of the need for PEML arose from fields such as structural mechanics[1][4] and environmental science[5], where purely data-driven methods struggled with limited data or lacked reliability. For instance, in structural engineering, traditional physics-based simulations can be very accurate but often require costly modelling and still face uncertainty in loads or material properties; on the other hand, data-driven models may fit experimental data but fail to generalise outside those conditions. PEML approaches were developed to bridge this gap, effectively creating a "spectrum" between the extremes of purely physics-based (white-box modelling) and purely data-driven (black-box modelling), known as grey-box or hybrid modelling.[6] In practice, this means a PEML model can leverage governing equations or simulation data to inform the learning process, thus requiring less training data and yielding outputs that obey physical laws.
Methods and Techniques
PEML encompasses a range of methods that integrate domain-specific physical knowledge into the machine learning process.[3] These techniques differ in how physics is incorporated, whether through loss functions, model structures, feature design, or data generation, and can be summarised into three different categories: Physics-Informed, Physics-Guided, and Physics-Encoded Machine Learning.
Physics-Informed Learning
Physics-Informed learning techniques integrate physical laws directly into the machine learning process to simulate complex systems using partial differential equations (PDEs) and embedding physical constraints directly into a machine learning algorithm, such as a neural network. An example of physics-informed learning is through the use of Physics-Informed Neural Networks (PINNs), which implement composite loss functions that balance errors with PDE residuals, effectively blending sparse observations with physical constraints.[7] They are most suitable for irregular geometries due to their ability to operate in a meshless paradigm by sampling random collocation points. Physics-informed learning excels in multi-physics scenarios such as electroconvection[8], molecular dynamics[9], and real-time 4D flow reconstruction from MRI observations.[7]
Physics-Guided Learning
In many PEML approaches, physical knowledge is introduced into the learning process not by altering the model itself, but through data preprocessing and feature engineering. This strategy enables conventional machine learning algorithms to work with inputs that already encode important physical structure, enhancing both accuracy and interpretability. Common techniques include:
- Physics-based feature extraction. Raw data is transformed into features with physical meaning, such as dimensionless numbers (e.g. Reynolds number or Mach number), wavelet coefficients, or energy spectra. For example, Mohan et al.[10] used a wavelet transform to extract turbulence-related features from velocity fields, embedding known physics of turbulent cascades into the model inputs.
- Simulation or theory-driven feature augmentation. Outputs of simplified physical models (or residuals between observed and predicted behaviour) are used as additional features, reducing the learning burden of the machine learning model by letting it focus only on the discrepancy. This technique has been used in chemical kinetics applications, where delta learning was applied to graph neural networks (GNNs) to enhance activation energy predictions in chemical reactions,[11] or in Quantitative Structure-Activity Relationships (QSARs), where molecular descriptors derived from quantum chemistry calculations or physical models are used as features to predict chemical properties.[12]
- Physical domain transformations. Data is mapped into domains where physics-relevant patterns are more easily captured. For example, signal processing often employs Fourier transforms to reveal frequency content, allowing oscillatory features to be revealed. This enables machine learning algorithms, such as convolutional neural networks (CNNs) to apply standard vision models, yielding better generalisation and efficiency by learning from spectrograms instead of raw waveforms.[13]
These preprocessing methods are especially useful when physical insight is available, but the system is too complex for fully mechanistic modelling. By encoding physics into the data, standard machine learning architectures such as multi-layer perceptrons (MLPs) can be trained without needing architecture-specific changes. This class of approach enables physics-guided learning, where training data already obeys physical laws. As a result, the learned mapping is inherently constrained by the input features, and the model does not need to discover fundamental physical relationships from scratch, since key patterns are already embedded in the data.[14]
Physics-Encoded Learning
Physics-encoded learning, otherwise known as hybrid modelling, combines physics-based components with data-driven components in a singular framework. This approach is useful when the underlying physical laws are partially understood but insufficient to describe the full system behaviour, and are computationally expensive to simulate. In such methods, the final model integrates the physics-based model and the data-driven correction term , along with additional biases to narrow the solution space to physically plausible outputs such that the system formulates:Common examples of physics-encoded learning include Gaussian Process (GP) latent force models[15][16] and Physics-Informed Sparse Identification of Nonlinear Dynamics (PhI-SINDy)[17], which have been used to model multiple degree-of-freedom (MDOF) oscillator with multiple Coulomb friction contacts under harmonic load using both synthetic and experimental noisy experiments with multiple sources of discontinuous nonlinearities.[18]
Applications of Physics-Enhanced Machine Learning
Challenges and Limitations
- ^ a b c Haywood-Alexander, Marcus; Liu, Wei; Bacsa, Kiran; Lai, Zhilu; Chatzi, Eleni (2024-01). "Discussing the spectrum of physics-enhanced machine learning: a survey on structural mechanics applications". Data-Centric Engineering. 5: e31. doi:10.1017/dce.2024.33. ISSN 2632-6736.
{{cite journal}}
: Check date values in:|date=
(help) - ^ Noordijk, Ben; Garcia Gomez, Monica L.; ten Tusscher, Kirsten H. W. J.; de Ridder, Dick; van Dijk, Aalt D. J.; Smith, Robert W. (2024-08-02). "The rise of scientific machine learning: a perspective on combining mechanistic modelling with machine learning for systems biology". Frontiers in Systems Biology. 4. doi:10.3389/fsysb.2024.1407994. ISSN 2674-0702.
{{cite journal}}
: CS1 maint: unflagged free DOI (link) - ^ a b Cicirello, Alice (2024-12-01). "Physics-Enhanced Machine Learning: a position paper for dynamical systems investigations". Journal of Physics: Conference Series. 2909 (1): 012034. doi:10.1088/1742-6596/2909/1/012034. ISSN 1742-6588.
- ^ Cross, Elizabeth J.; Gibson, S. J.; Jones, M. R.; Pitchforth, D. J.; Zhang, S.; Rogers, T. J. (2022), Cury, Alexandre; Ribeiro, Diogo; Ubertini, Filippo; Todd, Michael D. (eds.), "Physics-Informed Machine Learning for Structural Health Monitoring", Structural Health Monitoring Based on Data Science Techniques, Cham: Springer International Publishing, pp. 347–367, doi:10.1007/978-3-030-81716-9_17, ISBN 978-3-030-81716-9, retrieved 2025-06-03
- ^ Zhao, Ying; Chadha, Mayank; Barthlow, Dakota; Yeates, Elissa; Mcknight, Charles J.; Memarsadeghi, Natalie P.; Gugaratshan, Guga; Todd, Michael D.; Hu, Zhen (2024-09-20). "Physics-enhanced machine learning models for streamflow discharge forecasting". Journal of Hydroinformatics. 26 (10): 2506–2537. doi:10.2166/hydro.2024.061. ISSN 1464-7141.
- ^ Schweidtmann, Artur M.; Zhang, Dongda; von Stosch, Moritz (2024-03-01). "A review and perspective on hybrid modeling methodologies". Digital Chemical Engineering. 10: 100136. doi:10.1016/j.dche.2023.100136. ISSN 2772-5081.
- ^ a b Karniadakis, George Em; Kevrekidis, Ioannis G.; Lu, Lu; Perdikaris, Paris; Wang, Sifan; Yang, Liu (2021-06). "Physics-informed machine learning". Nature Reviews Physics. 3 (6): 422–440. doi:10.1038/s42254-021-00314-5. ISSN 2522-5820.
{{cite journal}}
: Check date values in:|date=
(help) - ^ Cai, Shengze; Wang, Zhicheng; Lu, Lu; Zaki, Tamer A.; Karniadakis, George Em (2021-07). "DeepM&Mnet: Inferring the electroconvection multiphysics fields based on operator approximation by neural networks". Journal of Computational Physics. 436: 110296. doi:10.1016/j.jcp.2021.110296. ISSN 0021-9991.
{{cite journal}}
: Check date values in:|date=
(help) - ^ Jia, Weile; Wang, Han; Chen, Mohan; Lu, Denghui; Lin, Lin; Car, Roberto; E, Weinan; Zhang, Linfeng (2020-11-09). "Pushing the limit of molecular dynamics with ab initio accuracy to 100 million atoms with machine learning". Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis. SC '20. Atlanta, Georgia: IEEE Press: 1–14. doi:10.5555/3433701.3433707. ISBN 978-1-7281-9998-6.
{{cite journal}}
: Check|doi=
value (help) - ^ Mohan, Livescu, Chertkov, Arvind T, Daniel, Michael (April 2020). "Wavelet-powered neural networks for turbulence".
{{cite web}}
: CS1 maint: multiple names: authors list (link) CS1 maint: url-status (link) - ^ Chang, Han-Chung; Tsai, Ming-Hsuan; Li, Yi-Pei (2025-02-10). "Enhancing Activation Energy Predictions under Data Constraints Using Graph Neural Networks". Journal of Chemical Information and Modeling. 65 (3): 1367–1377. doi:10.1021/acs.jcim.4c02319. ISSN 1549-960X. PMC 11815826. PMID 39862160.
- ^ Cherkasov, Artem; Muratov, Eugene N.; Fourches, Denis; Varnek, Alexandre; Baskin, Igor I.; Cronin, Mark; Dearden, John; Gramatica, Paola; Martin, Yvonne C.; Todeschini, Roberto; Consonni, Viviana; Kuz'min, Victor E.; Cramer, Richard; Benigni, Romualdo; Yang, Chihae (2014-06-26). "QSAR modeling: where have you been? Where are you going to?". Journal of Medicinal Chemistry. 57 (12): 4977–5010. doi:10.1021/jm4004285. ISSN 1520-4804. PMC 4074254. PMID 24351051.
- ^ Rudolph, Maja; Kurz, Stefan; Rakitsch, Barbara (2024-03-19). "Hybrid modeling design patterns". Journal of Mathematics in Industry. 14 (1): 3. doi:10.1186/s13362-024-00141-0. ISSN 2190-5983.
{{cite journal}}
: CS1 maint: unflagged free DOI (link) - ^ Faroughi, Salah A.; Pawar, Nikhil M.; Fernandes, Célio; Raissi, Maziar; Das, Subasish; Kalantari, Nima K.; Kourosh Mahjour, Seyed (2024-01-29). "Physics-Guided, Physics-Informed, and Physics-Encoded Neural Networks and Operators in Scientific Computing: Fluid and Solid Mechanics". Journal of Computing and Information Science in Engineering. 24 (040802). doi:10.1115/1.4064449. ISSN 1530-9827.
- ^ Zou, Joanna; Lourens, Eliz-Mari; Cicirello, Alice (2023-10). "Virtual sensing of subsoil strain response in monopile-based offshore wind turbines via Gaussian process latent force models". Mechanical Systems and Signal Processing. 200: 110488. doi:10.1016/j.ymssp.2023.110488. ISSN 0888-3270.
{{cite journal}}
: Check date values in:|date=
(help) - ^ Marino, Luca; Cicirello, Alice (2023-01). "A switching Gaussian process latent force model for the identification of mechanical systems with a discontinuous nonlinearity". Data-Centric Engineering. 4: e18. doi:10.1017/dce.2023.12. ISSN 2632-6736.
{{cite journal}}
: Check date values in:|date=
(help) - ^ Lathourakis, Christos (2025-03-20), xristosl0610/PhI-SINDy, retrieved 2025-06-03
- ^ Lathourakis, Christos; Cicirello, Alice (2024-07-01). "Physics enhanced sparse identification of dynamical systems with discontinuous nonlinearities". Nonlinear Dynamics. 112 (13): 11237–11264. doi:10.1007/s11071-024-09652-2. ISSN 1573-269X.