Learning and Representing: a Jointly Optimal Approach
- π€ Speaker: Xinhua Zhang, University of Alberta
- π Date & Time: Tuesday 10 April 2012, 10:00 - 11:00
- π Venue: Large lecture theatre, Microsoft Research Ltd, 7 J J Thomson Avenue (Off Madingley Road), Cambridge
Abstract
Data representations, and transformations of data representations, are fundamental to the design of effective machine learning methods. Previous research has established that expressing complex data objects such as documents or images as feature vectors reveals important structure, both in collections of data and in individual data items. For any particular application, however, one does not often know which features to use. Automatically discovering useful features as part of training has therefore been a long standing goal of machine learning research. Unfortunately the resulting training problem—simultaneously learning a feature representation and a data reconstruction model—has been deemed intractable in general.
I will demonstrate in this talk a fundamental reformulation of representation learning that enables the training problem to be solved both globally and efficiently, even when a feature representation and data reconstruction model are learned simultaneously. This is a major advance over the current state-of-the-art, where globally optimal representations cannot be guaranteed in general. I will show that this work has led to significant improvements in both generalization accuracy and training time over state-of-the-art methods.
In addition, we have taken a step towards scaling up the method to large datasets. Arguably, optimization underlies almost all branches in machine learning and a major difficulty is the nonsmooth objectives. I will show that a novel framework based on smoothing, pioneered by Nesterov, provably improves the convergence rates. In particular, I will show applications of our idea to optimizing multivariate performance measures and structure prediction. Empirical evaluation on some of the largest publicly available datasets from a variety of domains shows that our method learns the optimal model significantly faster than the state-of-the-art solvers. A broader application of the smoothing technique includes graphical model inference and compressive sensing.
Series This talk is part of the Microsoft Research Cambridge, public talks series.
Included in Lists
- All Talks (aka the CURE list)
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge talks
- Chris Davis' list
- Guy Emerson's list
- Interested Talks
- Large lecture theatre, Microsoft Research Ltd, 7 J J Thomson Avenue (Off Madingley Road), Cambridge
- Microsoft Research Cambridge, public talks
- ndk22's list
- ob366-ai4er
- Optics for the Cloud
- personal list
- PMRFPS's
- rp587
- School of Technology
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Xinhua Zhang, University of Alberta
Tuesday 10 April 2012, 10:00-11:00