Residual multi-fidelity neural network computing

Owen Davis*, Mohammad Motamed, Raúl Tempone

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

In this work, we consider the general problem of constructing a neural network surrogate model using multi-fidelity information. Motivated by error-complexity estimates for ReLU neural networks, we formulate the correlation between an inexpensive low-fidelity model and an expensive high-fidelity model as a possibly non-linear residual function. This function defines a mapping between (1) the shared input space of the models along with the low-fidelity model output, and (2) the discrepancy between the outputs of the two models. The computational framework proceeds by training two neural networks to work in concert. The first network learns the residual function on a small set of high- and low-fidelity data. Once trained, this network is used to generate additional synthetic high-fidelity data, which is used in the training of the second network. The trained second network then acts as our surrogate for the high-fidelity quantity of interest. We present four numerical examples to demonstrate the power of the proposed framework, showing that significant savings in computational cost may be achieved when the output predictions are desired to be accurate within small tolerances.

Original languageEnglish (US)
Article number15
JournalBIT Numerical Mathematics
Volume65
Issue number2
DOIs
StatePublished - Jun 2025

Keywords

  • Deep neural networks
  • Multi-fidelity computing
  • Residual modeling
  • Uncertainty quantification

ASJC Scopus subject areas

  • Software
  • Computer Networks and Communications
  • Computational Mathematics
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'Residual multi-fidelity neural network computing'. Together they form a unique fingerprint.

Cite this