University of Cambridge > Talks.cam > CQIF Seminar > Quantum Expectation Maximization and algorithms for learning representations

Quantum Expectation Maximization and algorithms for learning representations

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Sathyawageeswar Subramanian.

The Expectation-Maximization (EM) algorithm is a fundamentaltool in unsupervised machine learning. It is often used as an efficientway to solve Maximum Likelihood (ML) and Maximum A Posteriori (MAP)estimation problems, especially for models with latent variables. It isalso the algorithm of choice to fit mixture models.¬† In this talk wedefine and use a quantum version of EM to fit a Gaussian MixtureMode¬†(GMM). We start by introducing in great detail all the tools usedin quantum machine learning (QRAM, distance estimation procedures,quantum linear algebra, etc.). Then we present q-means: a quantumalgorithm for k-means. We generalize q-means algorithm to fit a GMM . Ouralgorithms are only polylogarithmic in the number of elements in thetraining set, but are polynomial in other parameters – as the dimensionof the feature space and the number of components in the mixture. We’lldiscuss some experiments concerning the runtime of these algorithms onreal datasets. We conclude by analyzing prospect relations betweenquantum iterative algorithms and the Information Bottleneck Method.

This talk is part of the CQIF Seminar series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2020 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity