Bayesian deep learning on a quantum computer

43Citations
Citations of this article
217Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Bayesian methods in machine learning, such as Gaussian processes, have great advantages compared to other techniques. In particular, they provide estimates of the uncertainty associated with a prediction. Extending the Bayesian approach to deep architectures has remained a major challenge. Recent results connected deep feedforward neural networks with Gaussian processes, allowing training without backpropagation. This connection enables us to leverage a quantum algorithm designed for Gaussian processes and develop a new algorithm for Bayesian deep learning on quantum computers. The properties of the kernel matrix in the Gaussian process ensure the efficient execution of the core component of the protocol, quantum matrix inversion, providing at least a polynomial speedup over classical algorithms. Furthermore, we demonstrate the execution of the algorithm on contemporary quantum computers and analyze its robustness with respect to realistic noise models.

Cite

CITATION STYLE

APA

Zhao, Z., Pozas-Kerstjens, A., Rebentrost, P., & Wittek, P. (2019). Bayesian deep learning on a quantum computer. Quantum Machine Intelligence, 1(1–2), 41–51. https://doi.org/10.1007/s42484-019-00004-7

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free