M-estimation strategies for the ranking problem
Add to your list(s)
Download to your calendar using vCal
If you have a question about this talk, please contact Richard Nickl.
Statistical learning theory was mainly developed in the framework
of binary classification under the assumption that observations in the
training set form an i.i.d. sample. The techniques involved in order to
provide statistical guarantees for state-of-the-art learning algorithms are
borrowed from the theory of empirical processes. This is made possible not
only because of the “i.i.d.” assumption on the data but also because of the
nature of the performance measures, such as classification error or margin
error, which are statistics of order one. In the talk, I will discuss a
variety of questions which arise in the theory when more involved criteria
are considered. The problem of bipartite ranking through ROC curve
optimization provides a prolific source of optimization functionals which
are statistics of order strictly larger than one and several examples will
be presented.
http://www.cmla.ens-cachan.fr/Membres/vayatis.html
This talk is part of the Statistics series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
|