Conclusion

In this section I review the impact of my work and list the open questions and challenges for model discovery.

true
August 6, 2021

Despite the massive progress in the last few years, model discover of (partial) differential equations is still a field in its infancy. At the start of this thesis in early 2019, both the sparse regression approach for PDEs and PINNs were novel; to our knowledge, no one had attempted to perform model discovery on noisy experimental data. Considering the experiments on synthetic data that would have been a futile attempt - they all showed model discovery required densely sampled datasets with very little noise (<10%, but more often <2% noise). With this we set out to construct methods able to handle noisy and sparse datasets originating from experiments. The line of work we have developed with DeepMoD makes large strides towards this goal. In each paper we show that our methods are able to handle >50% noise and work with an order of magnitude less data than classical methods on increasingly challenging (synthetic) datasets such as the Kuramoto-Shivashinsky equation. DeepMoD also recovered the underlying equation when applied to data generated by simple experiments, validating our approach.

A second, more implicit goal was to create accessible methods. Complex methods are less likely to be adapted than simple ones, and if the goal is to make model discovery a valuable addition to scientists’ toolboxes, complexity should be strongly limited. Ideally, scientists without a background in numerical methods should be able to understand and apply our methods. Deep learning is an ideal vehicle for this, as many scientists these days have at least a basic understanding. Automatic differentiation, the key but complex mechanism at the core of DL, is often abstracted away to frameworks such as Pytorch, making it possible to construct powerful approaches without in-depth numerical knowledge. This leads to the peculiar situation where a neural network-based approach is regarded as more accesible than splines. Indeed, our work shows that an integrated combination of simple features (a basic MLP, a simple Lasso and a constraint) can easily and strongly outperform much more complex traditional approaches - a testament to the power of differentiable programming.

Our work also carries strong implications for future work and for model discovery as a field. First and foremost, it shows that the limiting factor is the accuracy of the features, implying that accurately modelling and denoising data is just as important as the sparse regression. Perhaps paradoxically then, significant progress can be made by focussing on data-driven modelling. In all cases, neural networks (especially PINNs) should feature prominently; the benefits of automatic differentiation and excellent inter- and extrapolation only become more pronounced in high-dimensional data. That is not to say that non-DL-based approaches should be neglected. A main thread in our work has been to show how ‘classical’ methods such as sparse and Bayesian regression can be integrated in DL approaches, and most progress can be made by synthesizing these approaches. The Bayesian regression and model selection literature is especially rich, and combining DL-based modelling with these methods should prove fruitful.

Taken together, our work strongly establishes the argument for physics-constrained, neural network-based surrogates for model discovery of PDEs on experimental data.

Challenges and questions unanswered

Model discovery is a young and exciting field, and as with all young fields, many limitations, challenges and questions remain. As such I list these in no particular order below - I’ve mentioned some of them before, others might have occured to you while reading this thesis and a few of them are of a more philosophical nature.

Brunton, B. W., S. L. Brunton, J. L. Proctor, and J. N. Kutz. 2013. “Optimal Sensor Placement and Enhanced Sparsity for Classification.” arXiv:1310.4217 [Cs], October. http://arxiv.org/abs/1310.4217.
Champion, Kathleen, Steven L. Brunton, and J. Nathan Kutz. 2019. “Discovery of Nonlinear Multiscale Systems: Sampling Strategies and Embeddings.” SIAM Journal on Applied Dynamical Systems 18 (1): 312–33. https://doi.org/10.1137/18M1188227.
Chen, Aoxue, and Guang Lin. 2021. “Robust Data-Driven Discovery of Partial Differential Equations with Time-Dependent Coefficients.” arXiv:2102.01432 [Cs, Stat], February. http://arxiv.org/abs/2102.01432.
Cranmer, Miles, Sam Greydanus, Stephan Hoyer, Peter Battaglia, David Spergel, and Shirley Ho. 2020. “Lagrangian Neural Networks.” arXiv:2003.04630 [Physics, Stat], March. http://arxiv.org/abs/2003.04630.
Finzi, Marc, Max Welling, and Andrew Gordon Wilson. 2021. “A Practical Method for Constructing Equivariant Multilayer Perceptrons for Arbitrary Matrix Groups.” arXiv:2104.09459 [Cs, Math, Stat], April. http://arxiv.org/abs/2104.09459.
Greydanus, Sam, Misko Dzamba, and Jason Yosinski. 2019. “Hamiltonian Neural Networks.” arXiv:1906.01563 [Cs], September. http://arxiv.org/abs/1906.01563.
Rudy, Samuel, Alessandro Alla, Steven L. Brunton, and J. Nathan Kutz. 2018. “Data-Driven Identification of Parametric Partial Differential Equations.” arXiv:1806.00732 [Math], June. http://arxiv.org/abs/1806.00732.
Silva, Brian de, David M. Higdon, Steven L. Brunton, and J. Nathan Kutz. 2019. “Discovery of Physics from Data: Universal Laws and Discrepancy Models.” arXiv:1906.07906 [Physics, Stat], June. http://arxiv.org/abs/1906.07906.
Tod, Georges, Gert-Jan Both, and Remy Kusters. 2021. “Discovering PDEs from Multiple Experiments.” arXiv:2109.11939 [Physics, Stat], September. http://arxiv.org/abs/2109.11939.

References