On transfer learning of neural networks using bi-fidelity data for uncertainty propagation

02/11/2020
by   Subhayan De, et al.
21

Due to their high degree of expressiveness, neural networks have recently been used as surrogate models for mapping inputs of an engineering system to outputs of interest. Once trained, neural networks are computationally inexpensive to evaluate and remove the need for repeated evaluations of computationally expensive models in uncertainty quantification applications. However, given the highly parameterized construction of neural networks, especially deep neural networks, accurate training often requires large amounts of simulation data that may not be available in the case of computationally expensive systems. In this paper, to alleviate this issue for uncertainty propagation, we explore the application of transfer learning techniques using training data generated from both high- and low-fidelity models. We explore two strategies for coupling these two datasets during the training procedure, namely, the standard transfer learning and the bi-fidelity weighted learning. In the former approach, a neural network model mapping the inputs to the outputs of interest is trained based on the low-fidelity data. The high-fidelity data is then used to adapt the parameters of the upper layer(s) of the low-fidelity network, or train a simpler neural network to map the output of the low-fidelity network to that of the high-fidelity model. In the latter approach, the entire low-fidelity network parameters are updated using data generated via a Gaussian process model trained with a small high-fidelity dataset. The parameter updates are performed via a variant of stochastic gradient descent with learning rates given by the Gaussian process model. Using three numerical examples, we illustrate the utility of these bi-fidelity transfer learning methods where we focus on accuracy improvement achieved by transfer learning over standard training approaches.

READ FULL TEXT

page 9

page 12

page 17

page 21

research
05/27/2021

Neural Network Training Using ℓ_1-Regularization and Bi-fidelity Data

With the capability of accurately representing a functional relationship...
research
09/13/2022

Quadrature Sampling of Parametric Models with Bi-fidelity Boosting

Least squares regression is a ubiquitous tool for building emulators (a....
research
10/05/2020

Using Bayesian deep learning approaches for uncertainty-aware building energy surrogate models

Fast machine learning-based surrogate models are trained to emulate slow...
research
05/10/2022

Multifidelity data fusion in convolutional encoder/decoder networks

We analyze the regression accuracy of convolutional neural networks asse...
research
03/30/2023

Surrogate Neural Networks for Efficient Simulation-based Trajectory Planning Optimization

This paper presents a novel methodology that uses surrogate models in th...
research
10/03/2020

MFPC-Net: Multi-fidelity Physics-Constrained Neural Process

In this work, we propose a network which can utilize computational cheap...
research
01/16/2022

Efficient Training of Transfer Mapping in Physics-Infused Machine Learning Models of UAV Acoustic Field

Physics-Infused Machine Learning (PIML) architectures aim at integrating...

Please sign up or login with your details

Forgot password? Click here to reset