Abstract
We use a Gaussian Process Regression (GPR) strategy to analyze different types of curves that are commonly encountered in parametric eigenvalue problems. We employ an offline-online decomposition method. In the offline phase, we generate the basis of the reduced space by applying the proper orthogonal decomposition (POD) method on a collection of pre-computed, full-order snapshots at a chosen set of parameters. Then we generate our GPR model using four different Matérn covariance functions. In the online phase, we use this model to predict both eigenvalues and eigenvectors at new parameters. We then illustrate how the choice of each covariance function influences the performance of GPR. Furthermore, we discuss the connection between Gaussian Process Regression and spline methods and compare the performance of the GPR method against linear and cubic spline methods. We show that GPR outperforms other methods for functions with a certain regularity.
Original language | English (US) |
---|---|
Pages (from-to) | 533-555 |
Number of pages | 23 |
Journal | Computational Methods in Applied Mathematics |
Volume | 24 |
Issue number | 3 |
DOIs | |
State | Published - Jul 1 2024 |
Bibliographical note
Publisher Copyright:© 2024 Walter de Gruyter GmbH, Berlin/Boston.
Keywords
- Covariance Function
- Gaussian Process Regression
- Machine Learning
- PDE Eigenvalue Problems
- Reduced Order Modeling
- Splines
ASJC Scopus subject areas
- Numerical Analysis
- Computational Mathematics
- Applied Mathematics