CaltechTHESIS
  A Caltech Library Service

Operator Learning for Scientific Computing

Citation

Trautner, Margaret Katherine (2025) Operator Learning for Scientific Computing. Dissertation (Ph.D.), California Institute of Technology. doi:10.7907/s5p5-9j06. https://resolver.caltech.edu/CaltechTHESIS:05292025-172331972

Abstract

This thesis develops operator learning theory and methods for use in scientific computing. Operator learning uses data to approximate maps between infinite dimensional function spaces. As such, operator learning provides a natural framework for using machine learning in applications with partial differential equations (PDEs). While operator learning architectures have successfully modeled a variety of physical phenomena in practice, the theoretical foundations underpinning these successes remain in early stages of development.

The present work takes a step towards a complete understanding of operator learning and its potential use in scientific applications. The thesis begins by studying multiscale constitutive modeling, where operator learning models can serve as surrogates to accelerate simulation and aid in model discovery of physical laws. The work proposes, and theoretically and numerically analyzes, an operator learning architecture for modeling history dependence in homogenized constitutive equations. The thesis then addresses learning solutions to an elliptic PDE in the presence of discontinuities and corner interfaces in two-dimensional materials. By proving a key continuity result for the underlying PDE, a universal approximation result is obtained. In its second half, the thesis moves on from the setting of homogenized constitutive laws and gives insight to operator learning from a broader perspective. First, error analysis bounds a form of discretization error that arises in implementations of the Fourier Neural Operator (FNO). Next, a modified form of the FNO, the Fourier Neural Mapping, accommodates finite-dimensional data while retaining the underlying function space structure. This modification allows applications where the map of interest is governed by an infinite-dimensional operator with data, such as parameters or summary statistics, in the form of finite vectors. Finally, the thesis extends a theory-to-practice gap result in finite dimensions to the infinite-dimensional operator learning setting, asserting that even for classes of architectures whose model expressivity scales well with model size, their error convergence with respect to data size scales poorly. In summary, this thesis builds understanding of operator learning from several perspectives and contributes both theoretical advancements and practical methodologies that improve the applicability of operator learning models to scientific problems.

Item Type:Thesis (Dissertation (Ph.D.))
Subject Keywords:Learning theory, machine learning, scientific computing, operator learning, multiscale modeling, constitutive modeling, partial differential equations
Degree Grantor:California Institute of Technology
Division:Engineering and Applied Science
Major Option:Computing and Mathematical Sciences
Awards:Thomas A. Tisch Prize for Graduate Teaching in Computing and Mathematical Sciences, 2022.
Thesis Availability:Public (worldwide access)
Research Advisor(s):
  • Stuart, Andrew M.
Thesis Committee:
  • Owhadi, Houman (chair)
  • Hoffmann, Franca
  • Bhattacharya, Kaushik
  • Stuart, Andrew M.
Defense Date:23 May 2025
Funders:
Funding AgencyGrant Number
Department of Energy Computational Science Graduate FellowshipE-SC0021110
Department of Energy Computational Science Graduate FellowshipE-SC0022158
Department of Energy Computational Science Graduate FellowshipE-SC0023112
Department of Energy Computational Science Graduate FellowshipE-SC0024386
Record Number:CaltechTHESIS:05292025-172331972
Persistent URL:https://resolver.caltech.edu/CaltechTHESIS:05292025-172331972
DOI:10.7907/s5p5-9j06
Related URLs:
URLURL TypeDescription
https://doi.org/10.1137/22M149920DOIPaper "Learning Markovian homogenized models in viscoelasticity" adapted for Chapter 2.
https://doi.org/10.1016/ j.jmps.2023.105329DOIPaper "Learning macroscopic internal variables and history dependence from microscopic models" appearing as Chapter 2.
https://doi.org/10.1137/23M1585015DOIPaper "Learning homogenization for elliptic operators" adapted for Chapter 3.
https://doi.org/10.48550/arXiv.2405.02221arXivPreprint "Discretization error of Fourier neural operators" adapted for Chapter 4.
https://doi.org/10.48550/arXiv.2502.05463arXivPreprint "Learning Memory and Material Dependent Constitutive Laws" appearing as Chapter 2.
https://doi.org/10.48550/arXiv.2503.18219arXivPreprint "Theory to Practice Gap for Neural Networks and Neural Operators" adapted for Chapter 6.
https://doi.org/10.3934/ fods.2024037DOIPaper "An operator learning perspective on parameter-to-observable maps" adapted for Chapter 5.
ORCID:
AuthorORCID
Trautner, Margaret Katherine000-0001-9937-8393
Default Usage Policy:No commercial reproduction, distribution, display or performance rights in this work are provided.
ID Code:17296
Collection:CaltechTHESIS
Deposited By: Margaret Trautner
Deposited On:05 Jun 2025 17:55
Last Modified:13 Jun 2025 18:21

Thesis Files

[img] PDF - Final Version
See Usage Policy.

5MB

Repository Staff Only: item control page