Handling Sparsity via the Horseshoe
Add to your list(s)
Download to your calendar using vCal
If you have a question about this talk, please contact rbg24.
This paper presents a general, fully Bayesian framework for
sparse supervised-learning problems based on the horseshoe prior. The
horseshoe prior is a member of the family of multivariate scale
mixtures of normals, and is therefore closely related to widely used
approaches for sparse Bayesian learning, including, among others,
Laplacian (LASSO) and Student-t priors (relevance vector machines).
The advantages of the
horseshoe are its robustness at handling unknown sparsity and large
outlying signals. These properties are justified theoretically via a
representation theorem and accompanied by comprehensive empirical
experiments that compare its performance to benchmark alternatives.
This talk is part of the Statistics series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
|