Jump to content

Draft:Physics-Enhanced Machine Learning

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Ryan-907316 (talk | contribs) at 00:27, 4 June 2025. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Physics-Enhanced Machine Learning

Physics-Enhanced Machine Learning (PEML) is a subfield of machine learning that combines physics-based modelling and data-driven algorithms, and aims to combine the complimentary strengths of physical knowledge and machine learning algorithms to overcome the limitations of using either approach alone.[1] PEML incorporates known physical laws and domain knowledge into the machine learning process so that models not only fit observed data patterns, but also respect the governing physics equations of the model.

This concept is closely related to scientific machine learning (SciML)[2], Physics-Informed Machine Learning (PIML), Physics-Enhanced Artificial Intelligence (PEAI) and Physics-Guided Machine Learning, but is distinct in its emphasis on enhancing predictive capability through the use of physics-based components in hybrid or constrained models.

Background and Motivation

The motivation for PEML arose from challenges often encountered in engineering and real-world problems: The volume of useful data is generally limited; predictions obtained from modelling can make accurate, but physically implausible predictions that violate known physics; uncertainties could not easily be dealt with; and traditional machine learning models are not very explainable or interpretable.[3] By infusing physics into machine learning, it ensures better generalisation to unseen conditions, improved physical consistency of predictions, and a greater explainability of the learnt models. This approach has gained traction since 2019, where techniques have been used to encompass strategies where "prior physics knowledge is embedded to the learner" [1], particularly in dealing with dynamical systems in engineering.

Early examples of the need for PEML arose from fields such as structural mechanics[1][4] and environmental science[5], where purely data-driven methods struggled with limited data or lacked reliability. For instance, in structural engineering, traditional physics-based simulations can be very accurate but often require costly modelling and still face uncertainty in loads or material properties; on the other hand, data-driven models may fit experimental data but fail to generalise outside those conditions. PEML approaches were developed to bridge this gap, effectively creating a "spectrum" between the extremes of purely physics-based (white-box modelling) and purely data-driven (black-box modelling), known as grey-box or hybrid modelling.[6] In practice, this means a PEML model can leverage governing equations or simulation data to inform the learning process, thus requiring less training data and yielding outputs that obey physical laws.

Methods and Techniques

PEML encompasses a range of methods that integrate domain-specific physical knowledge into the machine learning process.[3] These techniques differ in how physics is incorporated, whether through loss functions, model structures, feature design, or data generation, and can be summarised into three different categories: Physics-Informed, Physics-Guided, and Physics-Encoded Machine Learning.

Physics-Informed Learning

Physics-Informed learning techniques integrate physical laws directly into the machine learning process to simulate complex systems using partial differential equations (PDEs) and embedding physical constraints directly into a machine learning algorithm, such as a neural network. An example of physics-informed learning is through the use of Physics-Informed Neural Networks (PINNs), which implement composite loss functions that balance errors with PDE residuals, effectively blending sparse observations with physical constraints.[7] They are most suitable for irregular geometries due to their ability to operate in a meshless paradigm by sampling random collocation points. Physics-informed learning excels in multi-physics scenarios such as electroconvection[8], molecular dynamics[9], and real-time 4D flow reconstruction from MRI observations.[7]

Physics-Guided Learning

In many PEML approaches, physical knowledge is introduced into the learning process not by altering the model itself, but through data preprocessing and feature engineering. This strategy enables conventional machine learning algorithms to work with inputs that already encode important physical structure, enhancing both accuracy and interpretability. Common techniques include:

  1. Physics-based feature extraction. Raw data is transformed into features with physical meaning, such as dimensionless numbers (e.g. Reynolds number or Mach number), wavelet coefficients, or energy spectra. For example, Mohan et al.[10] used a wavelet transform to extract turbulence-related features from velocity fields, embedding known physics of turbulent cascades into the model inputs.
  2. Simulation or theory-driven feature augmentation. Outputs of simplified physical models (or residuals between observed and predicted behaviour) are used as additional features, reducing the learning burden of the machine learning model by letting it focus only on the discrepancy. This technique has been used in chemical kinetics applications, where delta learning was applied to graph neural networks (GNNs) to enhance activation energy predictions in chemical reactions,[11] or in Quantitative Structure-Activity Relationships (QSARs), where molecular descriptors derived from quantum chemistry calculations or physical models are used as features to predict chemical properties.[12]
  3. Physical domain transformations. Data is mapped into domains where physics-relevant patterns are more easily captured. For example, signal processing often employs Fourier transforms to reveal frequency content, allowing oscillatory features to be revealed. This enables machine learning algorithms, such as convolutional neural networks (CNNs) to apply standard vision models, yielding better generalisation and efficiency by learning from spectrograms instead of raw waveforms.[13]

These preprocessing methods are especially useful when physical insight is available, but the system is too complex for fully mechanistic modelling. By encoding physics into the data, standard machine learning architectures such as multi-layer perceptrons (MLPs) can be trained without needing architecture-specific changes. This class of approach enables physics-guided learning, where training data already obeys physical laws. As a result, the learned mapping is inherently constrained by the input features, and the model does not need to discover fundamental physical relationships from scratch, since key patterns are already embedded in the data.[14]

Physics-Encoded Learning

Physics-encoded learning, otherwise known as hybrid modelling, combines physics-based components with data-driven components in a singular framework. This approach is useful when the underlying physical laws are partially understood but insufficient to describe the full system behaviour, and are computationally expensive to simulate. In such methods, the final model integrates the physics-based model and the data-driven correction term , along with additional biases to narrow the solution space to only contain physically plausible outputs such that the system is in the form:Common examples of physics-encoded learning include Gaussian Process (GP) latent force models[15][16] and Physics-Informed Sparse Identification of Nonlinear Dynamics (PhI-SINDy)[17], which have been used to model multiple degree-of-freedom (MDOF) oscillators with multiple Coulomb friction contacts under harmonic load using both synthetic and experimental noisy experiments with multiple sources of discontinuous nonlinearities.[18]

Applications of Physics-Enhanced Machine Learning

PEML methods have moved beyond theoretical development and are now actively deployed in real-world systems across engineering, biology, chemistry, physics, scientific discovery, and computer science, to name a few applications. These applications are especially valuable in high-stakes or data-scarce environments where traditional machine learning or purely physics-based models may fall short.

Wind Turbine Structural Monitoring

PEML has been applied to predict fatigue loads in wind turbine blades under wake steering control (WSC), a strategy that improves wind farms efficiency by intentionally misaligning turbine yaw angles to reduce wake interference.[19] While WSC can enhance power output, it also introduces additional fatigue loads on downstream turbines, complicating structural health monitoring. Traditional methods, such as look-up-tables (LUTs), often fail to capture the nonlinear dynamics of wake-induced loading. A recent approach addressed this by using Gaussian process (GP) models trained on physics-informed features, including damage-equivalent loads (DELs) derived from Rainflow Counting and the Palmgren-Miner rule. These GPs provided probabilistic fatigue predictions with improved accuracy. Compared to LUTs, the PEML model reduced the root mean square error (RMSE) by 13.99% for edgewise moments and by 51.87% for flapwise moments, highlighting the value of incorporating fatigue physics into machine learning-based predictive maintenance.

Tuned Mass Damper Optimisation

Tuned Mass Dampers (TMDs) are widely used to mitigate structural vibrations in tall buildings during seismic events. Traditional physics-based design methods, such as the Den Hartog approach, assume linear structural behaviour and do not fully capture the effects of nonlinear dynamics or variable seismic loads. Conversely, purely data-driven optimisation techniques may lack physical constraints, resulting in unrealistic or inefficient damping configurations. To address this, researchers developed a PEML framework based on a generative adversarial network (GAN) architecture.

Challenges and Limitations

  1. ^ a b c Haywood-Alexander, Marcus; Liu, Wei; Bacsa, Kiran; Lai, Zhilu; Chatzi, Eleni (2024-01). "Discussing the spectrum of physics-enhanced machine learning: a survey on structural mechanics applications". Data-Centric Engineering. 5: e31. doi:10.1017/dce.2024.33. ISSN 2632-6736. {{cite journal}}: Check date values in: |date= (help)
  2. ^ Noordijk, Ben; Garcia Gomez, Monica L.; ten Tusscher, Kirsten H. W. J.; de Ridder, Dick; van Dijk, Aalt D. J.; Smith, Robert W. (2024-08-02). "The rise of scientific machine learning: a perspective on combining mechanistic modelling with machine learning for systems biology". Frontiers in Systems Biology. 4. doi:10.3389/fsysb.2024.1407994. ISSN 2674-0702.{{cite journal}}: CS1 maint: unflagged free DOI (link)
  3. ^ a b Cicirello, Alice (2024-12-01). "Physics-Enhanced Machine Learning: a position paper for dynamical systems investigations". Journal of Physics: Conference Series. 2909 (1): 012034. doi:10.1088/1742-6596/2909/1/012034. ISSN 1742-6588.
  4. ^ Cross, Elizabeth J.; Gibson, S. J.; Jones, M. R.; Pitchforth, D. J.; Zhang, S.; Rogers, T. J. (2022), Cury, Alexandre; Ribeiro, Diogo; Ubertini, Filippo; Todd, Michael D. (eds.), "Physics-Informed Machine Learning for Structural Health Monitoring", Structural Health Monitoring Based on Data Science Techniques, Cham: Springer International Publishing, pp. 347–367, doi:10.1007/978-3-030-81716-9_17, ISBN 978-3-030-81716-9, retrieved 2025-06-03
  5. ^ Zhao, Ying; Chadha, Mayank; Barthlow, Dakota; Yeates, Elissa; Mcknight, Charles J.; Memarsadeghi, Natalie P.; Gugaratshan, Guga; Todd, Michael D.; Hu, Zhen (2024-09-20). "Physics-enhanced machine learning models for streamflow discharge forecasting". Journal of Hydroinformatics. 26 (10): 2506–2537. doi:10.2166/hydro.2024.061. ISSN 1464-7141.
  6. ^ Schweidtmann, Artur M.; Zhang, Dongda; von Stosch, Moritz (2024-03-01). "A review and perspective on hybrid modeling methodologies". Digital Chemical Engineering. 10: 100136. doi:10.1016/j.dche.2023.100136. ISSN 2772-5081.
  7. ^ a b Karniadakis, George Em; Kevrekidis, Ioannis G.; Lu, Lu; Perdikaris, Paris; Wang, Sifan; Yang, Liu (2021-06). "Physics-informed machine learning". Nature Reviews Physics. 3 (6): 422–440. doi:10.1038/s42254-021-00314-5. ISSN 2522-5820. {{cite journal}}: Check date values in: |date= (help)
  8. ^ Cai, Shengze; Wang, Zhicheng; Lu, Lu; Zaki, Tamer A.; Karniadakis, George Em (2021-07). "DeepM&Mnet: Inferring the electroconvection multiphysics fields based on operator approximation by neural networks". Journal of Computational Physics. 436: 110296. doi:10.1016/j.jcp.2021.110296. ISSN 0021-9991. {{cite journal}}: Check date values in: |date= (help)
  9. ^ Jia, Weile; Wang, Han; Chen, Mohan; Lu, Denghui; Lin, Lin; Car, Roberto; E, Weinan; Zhang, Linfeng (2020-11-09). "Pushing the limit of molecular dynamics with ab initio accuracy to 100 million atoms with machine learning". Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis. SC '20. Atlanta, Georgia: IEEE Press: 1–14. doi:10.5555/3433701.3433707. ISBN 978-1-7281-9998-6. {{cite journal}}: Check |doi= value (help)
  10. ^ Mohan, Livescu, Chertkov, Arvind T, Daniel, Michael (April 2020). "Wavelet-powered neural networks for turbulence".{{cite web}}: CS1 maint: multiple names: authors list (link) CS1 maint: url-status (link)
  11. ^ Chang, Han-Chung; Tsai, Ming-Hsuan; Li, Yi-Pei (2025-02-10). "Enhancing Activation Energy Predictions under Data Constraints Using Graph Neural Networks". Journal of Chemical Information and Modeling. 65 (3): 1367–1377. doi:10.1021/acs.jcim.4c02319. ISSN 1549-960X. PMC 11815826. PMID 39862160.
  12. ^ Cherkasov, Artem; Muratov, Eugene N.; Fourches, Denis; Varnek, Alexandre; Baskin, Igor I.; Cronin, Mark; Dearden, John; Gramatica, Paola; Martin, Yvonne C.; Todeschini, Roberto; Consonni, Viviana; Kuz'min, Victor E.; Cramer, Richard; Benigni, Romualdo; Yang, Chihae (2014-06-26). "QSAR modeling: where have you been? Where are you going to?". Journal of Medicinal Chemistry. 57 (12): 4977–5010. doi:10.1021/jm4004285. ISSN 1520-4804. PMC 4074254. PMID 24351051.
  13. ^ Rudolph, Maja; Kurz, Stefan; Rakitsch, Barbara (2024-03-19). "Hybrid modeling design patterns". Journal of Mathematics in Industry. 14 (1): 3. doi:10.1186/s13362-024-00141-0. ISSN 2190-5983.{{cite journal}}: CS1 maint: unflagged free DOI (link)
  14. ^ Faroughi, Salah A.; Pawar, Nikhil M.; Fernandes, Célio; Raissi, Maziar; Das, Subasish; Kalantari, Nima K.; Kourosh Mahjour, Seyed (2024-01-29). "Physics-Guided, Physics-Informed, and Physics-Encoded Neural Networks and Operators in Scientific Computing: Fluid and Solid Mechanics". Journal of Computing and Information Science in Engineering. 24 (040802). doi:10.1115/1.4064449. ISSN 1530-9827.
  15. ^ Zou, Joanna; Lourens, Eliz-Mari; Cicirello, Alice (2023-10). "Virtual sensing of subsoil strain response in monopile-based offshore wind turbines via Gaussian process latent force models". Mechanical Systems and Signal Processing. 200: 110488. doi:10.1016/j.ymssp.2023.110488. ISSN 0888-3270. {{cite journal}}: Check date values in: |date= (help)
  16. ^ Marino, Luca; Cicirello, Alice (2023-01). "A switching Gaussian process latent force model for the identification of mechanical systems with a discontinuous nonlinearity". Data-Centric Engineering. 4: e18. doi:10.1017/dce.2023.12. ISSN 2632-6736. {{cite journal}}: Check date values in: |date= (help)
  17. ^ Lathourakis, Christos (2025-03-20), xristosl0610/PhI-SINDy, retrieved 2025-06-03
  18. ^ Lathourakis, Christos; Cicirello, Alice (2024-07-01). "Physics enhanced sparse identification of dynamical systems with discontinuous nonlinearities". Nonlinear Dynamics. 112 (13): 11237–11264. doi:10.1007/s11071-024-09652-2. ISSN 1573-269X.
  19. ^ Miao, Yizhi; Soltani, Mohsen N.; Hajizadeh, Amin (2022-07-22). "A Machine Learning Method for Modeling Wind Farm Fatigue Load". Applied Sciences. 12 (15): 7392. doi:10.3390/app12157392. ISSN 2076-3417. Archived from the original on 2024-07-04.{{cite journal}}: CS1 maint: unflagged free DOI (link)