Reset filters

Search publications


Search by keyword
List by department / centre / faculty

No publications found.

 

Near-optimal learning of Banach-valued, high-dimensional functions via deep neural networks

Authors: Adcock BBrugiapaglia SDexter NMoraga S


Affiliations

1 Department of Mathematics, Simon Fraser University, 8888 University Drive, Burnaby BC, Canada, V5A 1S6. Electronic address: ben_adcock@sfu.ca.
2 Department of Mathematics and Statistics, Concordia University, J.W. McConnell Building, 1400 De Maisonneuve Blvd. W., Montréal, QC, Canada, H3G 1M8. Electronic address: simone.brugiapaglia@concordia.ca.
3 Department of Scientific Computing, Florida State University, 400 Dirac Science Library, Tallahassee, FL, 32306-4120, USA. Electronic address: nick.dexter@fsu.edu.
4 Department of Mathematics, Simon Fraser University, 8888 University Drive, Burnaby BC, Canada, V5A 1S6. Electronic address: smoragas@sfu.ca.

Description

The past decade has seen increasing interest in applying Deep Learning (DL) to Computational Science and Engineering (CSE). Driven by impressive results in applications such as computer vision, Uncertainty Quantification (UQ), genetics, simulations and image processing, DL is increasingly supplanting classical algorithms, and seems poised to revolutionize scientific computing. However, DL is not yet well-understood from the standpoint of numerical analysis. Little is known about the efficiency and reliability of DL from the perspectives of stability, robustness, accuracy, and, crucially, sample complexity. For example, approximating solutions to parametric PDEs is a key task in UQ for CSE. Yet, training data for such problems is often scarce and corrupted by errors. Moreover, the target function, while often smooth, is a potentially infinite-dimensional function taking values in the PDE solution space, which is generally an infinite-dimensional Banach space. This paper provides arguments for Deep Neural Network (DNN) approximation of such functions, with both known and unknown parametric dependence, that overcome the curse of dimensionality. We establish practical existence theorems that describe classes of DNNs with dimension-independent architecture widths and depths, and training procedures based on minimizing a (regularized) l2-loss which achieve near-optimal algebraic rates of convergence in terms of the amount of training data m. These results involve key extensions of compressed sensing for recovering Banach-valued vectors and polynomial emulation with DNNs. When approximating solutions of parametric PDEs, our results account for all sources of error, i.e., sampling, optimization, approximation and physical discretization, and allow for training high-fidelity DNN approximations from coarse-grained sample data. Our theoretical results fall into the category of non-intrusive methods, providing a theoretical alternative to classical methods for high-dimensional approximation.


Keywords: Banach spacesDeep learningDeep neural networksHigh-dimensional approximationUncertainty quantification


Links

PubMed: https://pubmed.ncbi.nlm.nih.gov/39454372/

DOI: 10.1016/j.neunet.2024.106761