Bayesian; Computational modelling; Data-driven model; Model understanding; Scientific investigation; Sources of uncertainty; Theoretical modeling; Uncertainty; Uncertainty quantifications; Physics and Astronomy (all); General Physics and Astronomy
Abstract :
[en] Being able to quantify uncertainty when comparing a theoretical or computational model to observations is critical to conducting a sound scientific investigation. With the rise of data-driven modelling, understanding various sources of uncertainty and developing methods to estimate them has gained renewed attention. Five researchers discuss uncertainty quantification in machine-learned models with an emphasis on issues relevant to physics problems.
Disciplines :
Mathematics Physics Computer science
Author, co-author :
Gal, Yarin; Oxford Applied and Theoretical Machine Learning Group, Department of Computer Science, University of Oxford, Oxford, United Kingdom
Koumoutsakos, Petros ; Computational Science and Engineering Laboratory, School of Engineering and Applied Sciences, Harvard University, Cambridge, United States
Lanusse, Francois; CNRS, CEA Saclay, Saclay, France
Louppe, Gilles ; Université de Liège - ULiège > Département d'électricité, électronique et informatique (Institut Montefiore) > Big Data
Papadimitriou, Costas; Department of Mechanical Engineering, University of Thessaly, Volos, Greece
Language :
English
Title :
Bayesian uncertainty quantification for machine-learned models in physics
Y.G. holds a Turing Articifical Intelligence Fellowship at the Alan Turing Institute, which is supported by Engineering and Physical Sciences Research Council (EPSRC) grant reference V030302/1.
Cranmer, K., Brehmer, J. & Louppe, G. The frontier of simulation-based inference. Proc. Natl Acad. Sci. USA 117, 30055–30062 (2020). DOI: 10.1073/pnas.1912789117
Abdar, M. et al. A review of uncertainty quantification in deep learning: techniques, applications and challenges. Inf. Fusion 76, 243–297 (2021). DOI: 10.1016/j.inffus.2021.05.008
Hüllermeier, E. & Waegeman, W. Aleatoric and epistemic uncertainty in machine learning: an introduction to concepts and methods. Mach. Learn. 110, 457–506 (2021). DOI: 10.1007/s10994-021-05946-3
Gal, Y. & Ghahramani, Z. Dropout as a Bayesian approximation: representing model uncertainty in deep learning. In Proc. 33rd International Conference on International Conference on Machine Learning Vol 48 1050–1059 (PMLR, 2016).
Neal, R. M. Bayesian Learning for Neural Networks (Springer, 1996).
van Amersfoort, J., Smith, L., Teh, Y. W. & Gal, Y. Uncertainty estimation using a single deep deterministic neural network. In Proc. 37th International Conference on Machine Learning Vol 119 9690–9700 (PMLR, 2020).
Liu, J. Z. et al. Simple and principled uncertainty estimation with deterministic deep learning via distance awareness. In Proc. 34th International Conference on Neural Information Processing Systems (NIPS’20) 7498–7512 (Curran Associates, 2020).
van Amersfoort, J., Smith, L., Jesson, A., Key, O. & Gal, Y. Improving deterministic uncertainty estimation in deep learning for classification and regression. Preprint at https://arxiv.org/abs/2102.11409v1 (2021).
Martin, S. M. et al. Korali: efficient and scalable software framework for Bayesian uncertainty quantification and stochastic optimization. Comput. Methods Appl. Mech. Eng. 389, 114264 (2022). DOI: 10.1016/j.cma.2021.114264
Bae, H. J. & Koumoutsakos, P. Scientific multi-agent reinforcement learning for wall-models of turbulent flows. Nat. Commun. 13, 1443 (2022). DOI: 10.1038/s41467-022-28957-7
Berger, J. O. & Smith, L. A. On the statistical formalism of uncertainty quantification. Annu. Rev. Stat. Appl. 6, 433–460 (2019). DOI: 10.1146/annurev-statistics-030718-105232