1 of 1

Learning differentiable solvers for systems with hard constraints

1

Scientific Achievement

We develop a new method to enforce differential equation constraints into neural networks (NNs) with a high degree of accuracy. We show that our approach achieves significantly lower error compared to previous approaches, and converges much more quickly to the right solutions.

Significance and Impact

Partial differential equations (PDEs) are crucial for describing the complex phenomena of climate dynamics, and numerous other energy-related areas. NNs provide a way to approximate solutions to such systems much faster than numerical methods, but current approaches only enforce physical constraints approximately – we address this key problem through our method.

Technical Approach

  • We propose a method to enforce hard PDE constraints by creating a differentiable layer. We make the layer differentiable using implicit differentiation, thereby allowing us to train our model with gradient-based optimization methods. This layer allows us to find the optimal linear combination of functions in a learned basis, given the PDE constraint.
  • We provide empirical validation of our method on multiple PDEs. Compared to previous NN approaches, our approach takes fewer iterations to converge to the correct solution, and also requires less training time.

G. Negiar, M. W. Mahoney, A. S. Krishnapriyan, International Conference on Learning Representations (2023).

True solution

Different between ML prediction and true (our method)

Difference between ML prediction and true (previous best method)

Error compared to true solutions, our approach is more accurate and converges more quickly