|COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring.|
Grouping strategies for denoising
If you have a question about this talk, please contact Richard Samworth.
We investigate the statistical learning approach for modeling various applications. This modeling involves several phases which need to be solved : the first one often is an approximation step, where we need to translate the observations into a dictionary. The choice of this dictionary (wavelets, needlets, variouslets,..., combinations of several bases,...) often conceals a significant part of investigation. The second phase is the treatment of very high dimensional data (ultra-high dimension means that the number of parameters may grow exponentially faster than the number of observations). This phase is requiring optimization methods of different style : $l_1$ minimizers, multi steps methods,..., as well as concentration inequalities. We concentrate on two steps thresholding methods and observe that making groups in the coefficients can seriously improve the selection and prediction rates. We provide a ‘boosting-grouping’ strategy, taking into account this observation.
This talk is part of the Statistics series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
Other listsMini Courses in Theoretical Computer Science Is there Enough for All of Us? Global Growth, Climate Change and Food Security Fluid Mechanics (CUED)
Other talksAutumn Cactus & Succulent Show MASTERCLASS: Case study research in applied linguistics: What can case studies of L2 learners tell us? Is the All-There-Is all there is? Intimations of our place in the cosmos Evolution of vertebrate photoreception and retinoid cycle: Clues from the eye transcriptome of basal vertebrates TBC Stability and breakup of liquid threads