The study focuses on the derivative-informed learning of nonlinear operators between infinite-dimensional separable Hilbert spaces by neural networks.
The approximation accuracy of the operator's derivatives can significantly impact the performance of the surrogate model for various outer-loop tasks in science and engineering.
The study analyzes the approximation errors of neural operators in Sobolev norms over infinite-dimensional Gaussian input measures.
The analysis is validated on numerical experiments with elliptic PDEs, demonstrating the accuracy of bases informed by the map.