|COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring.|
Using Context and Insight for the Analysis of LittleData?
If you have a question about this talk, please contact Zoubin Ghahramani.
We live in the era of Big Data with many interesting challenges for Machine Learning. In the opposite regime of “Little Data”, a limited population size, limited measurement time or a limited budget constrains the amount of data available. Examples include magnetic resonance imaging, microarray experiments or portfolio management. To obtain sensible predictions it is then crucial to take modeling assumptions into account. In this talk I will cover challenges, both computational and conceptual, that come up in the presence of a limited amount of data and how they can be overcome. I will focus on causality and high-dimensional covariance estimation and present a new algorithm that allows to compute an asymptotically unbiased, sparse and positive semidefinite estimator for covariance matrices using the SCAD penalty from high-dimensional statistics.
This talk is part of the Machine Learning @ CUED series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
Other listsTrinity Hall Natural Sciences Society Cambridge Comparative Syntax Conference (CamCoS) DNA, Cells and Cancer- A Symposium to Honour Professor Ron Laskey
Other talksTowards synchronized (and programmable) datacenters Optimal Bayes estimators for block models The Quest for Innovative Treatments in Psychiatry and Medicine: a personal perspective Prof Massimo Piepoli: TBC Bayesian Methods for Networks Inference in the infinite relational model