University of Cambridge > > Signal Processing and Communications Lab Seminars > Shrinkage Estimation in High Dimensions

Shrinkage Estimation in High Dimensions

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Prof. Ramji Venkataramanan.

We consider the problem of estimating a high-dimensional vector of parameters from a noisy one-time observation. The noise vector is iid Gaussian with known variance, and the performance of the estimator is measured via squared-error loss. For this problem, shrinkage estimators, which shrink the observed data towards a point or a target subspace, have evoked a lot of interest because they dominate the simple maximum-likelihood estimator (when the number of dimensions exceeds two).

In this talk, we first review the key aspects of shrinkage estimation, and then introduce shrinkage estimators that use the data to determine a “good” target subspace to shrink the data towards. We give concentration results for the squared-error loss and convergence results for the risk of the proposed estimators. We also present simulation results that validate the theory.

This is joint work with Ramji Venkataramanan.

This talk is part of the Signal Processing and Communications Lab Seminars series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.


© 2006-2023, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity