Neural Networks (NNs) have provided state-of-the-art results for many challenging machine learning tasks such as detection, regression and classification across the domains of computer vision, speech recognition and natural language processing. Despite their success, they are often implemented in a frequentist scheme, meaning they are unable to reason about uncertainty in their predictions. This article introduces Bayesian Neural Networks (BNNs) and the seminal research regarding their implementation. Different approximate inference methods are compared, and used to highlight where future research can improve on current methods.
CITATION STYLE
Goan, E., & Fookes, C. (2020). Bayesian Neural Networks: An Introduction and Survey. In Lecture Notes in Mathematics (Vol. 2259, pp. 45–87). Springer. https://doi.org/10.1007/978-3-030-42553-1_3
Mendeley helps you to discover research relevant for your work.