This chapter deals with kernel methods as a special class of techniques for surrogate modeling. Kernel methods have proven to be efficient in machine learning, pattern recognition and signal analysis due to their flexibility, excellent experimental performance and elegant functional analytic background.These data-based techniques provide so called kernel-expansions, i.e., linear combinations of kernel functions which are generated from given input–output point samples that may be arbitrarily scattered. In particular, these techniques are meshless, do not require or depend on a grid, hence are less prone to the curse of dimensionality, even for high-dimensional problems. In contrast to projection-based model reduction, we do not necessarily assume a high-dimensional model, but a general function that models input–output behavior within some simulation context. This could be some micro-model in a multiscale simulation, some submodel in a coupled system, some initialization function for solvers, coefficient function in Partial Differential Equations(PDEs),etc. First, kernel surrogates can be useful if the input–output function is expensive to evaluate, e.g. as a result of a finite element simulation. Here, acceleration can be obtained by sparse kernel expansions. Second, if a function is available only via measurements or a few function evaluation samples, kernel approximation techniques can provide function surrogates that allow for global evaluation. We present some important kernel approximation techniques, which are kernel interpolation, greedy kernel approximation and support vector regression. Pseudo-code is provided for ease of reproducibility. In order toillustrate the main features,commonalities and differences,wec ompare these techniques on a real-world application.The experiments clearly indicate the enormous acceleration potential.

9 Kernel methods for surrogate modeling

Santin, Gabriele
;
2021

Abstract

This chapter deals with kernel methods as a special class of techniques for surrogate modeling. Kernel methods have proven to be efficient in machine learning, pattern recognition and signal analysis due to their flexibility, excellent experimental performance and elegant functional analytic background.These data-based techniques provide so called kernel-expansions, i.e., linear combinations of kernel functions which are generated from given input–output point samples that may be arbitrarily scattered. In particular, these techniques are meshless, do not require or depend on a grid, hence are less prone to the curse of dimensionality, even for high-dimensional problems. In contrast to projection-based model reduction, we do not necessarily assume a high-dimensional model, but a general function that models input–output behavior within some simulation context. This could be some micro-model in a multiscale simulation, some submodel in a coupled system, some initialization function for solvers, coefficient function in Partial Differential Equations(PDEs),etc. First, kernel surrogates can be useful if the input–output function is expensive to evaluate, e.g. as a result of a finite element simulation. Here, acceleration can be obtained by sparse kernel expansions. Second, if a function is available only via measurements or a few function evaluation samples, kernel approximation techniques can provide function surrogates that allow for global evaluation. We present some important kernel approximation techniques, which are kernel interpolation, greedy kernel approximation and support vector regression. Pseudo-code is provided for ease of reproducibility. In order toillustrate the main features,commonalities and differences,wec ompare these techniques on a real-world application.The experiments clearly indicate the enormous acceleration potential.
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: http://hdl.handle.net/11582/329152
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
social impact