Optimizing Variational Physics-Informed Neural Networks Using Least Squares

Carlos Uriarte, Manuela Bastidas, David Pardo, Jamie M. Taylor, Sergio Rojas

Research output: Contribution to journalArticleResearchpeer-review

5 Citations (Scopus)

Abstract

Variational Physics-Informed Neural Networks often suffer from poor convergence when using stochastic gradient-descent-based optimizers. By introducing a least squares solver for the weights of the last layer of the neural network, we improve the convergence of the loss during training in most practical scenarios. This work analyzes the computational cost of the resulting hybrid least-squares/gradient-descent optimizer and explains how to implement it efficiently. In particular, we show that a traditional implementation based on backward-mode automatic differentiation leads to a prohibitively expensive algorithm. To remedy this, we propose using either forward-mode automatic differentiation or an ultraweak-type scheme that avoids the differentiation of trial functions in the discrete weak formulation. The proposed alternatives are up to one hundred times faster than the traditional one, recovering a computational cost-per-iteration similar to that of a conventional gradient-descent-based optimizer alone. To support our analysis, we derive computational estimates and conduct numerical experiments in one- and two-dimensional problems.

Original languageEnglish
Pages (from-to)76-93
Number of pages18
JournalComputers and Mathematics with Applications
Volume185
DOIs
Publication statusPublished - 1 May 2025

Keywords

  • Gradient-descent optimization
  • Least squares
  • Neural networks
  • Variational problems

Cite this