COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > NLIP Seminar Series > Predict, Price and Cut: Column and Row Generation for Structured Prediction
Predict, Price and Cut: Column and Row Generation for Structured PredictionAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Ekaterina Kochmar. Many problems in NLP , and structured prediction in general, can be cast as finding high-scoring structures based on a large set of candidate parts. For example, in second order tagging, we have to select high-scoring transitions between tags in a globally consistent fashion. In second order graph-based dependency parsing we have to choose a quadratic number of first order and a cubic number of second order edges such that the graph is both high-scoring and a tree. What makes such problems challenging is the large number of possible parts to consider. This number not only affects the cost of search or optimization but also slows down the process of scoring parts before they enter the optimisation problem, and extracting features. In this talk I present an approach that can solve problems with large sets of candidate parts without considering all of these parts in either optimization or scoring. In contrast to most pruning heuristics, our algorithm can give certificates of optimality before having optimized over, or even scored, all parts. It does so without the need of auxiliary models or tuning of threshold parameters. This is achieved by a delayed column and row generation algorithm that iteratively solves an LP relaxation over a small subset of current candidate parts, and then finds new candidates with high scores that can be inserted into the current optimal solution without removing high scoring existing structure. The latter step subtracts from the cost of a part the price of resources the part requires, and is often referred as pricing. Sometimes parts may score highly after pricing, but are necessary in order to make the current solution feasible. We add such parts in a step that roughly amounts to violated cuts to the LP. We evaluate our approach on two applications: second order dependency parsing and first order tagging with large domains. In both cases we dramatically reduce the number of parts considered, and observe about an order of magnitude speed-up. This is possible without loss of optimality guarantees, and hence accuracy. This talk is part of the NLIP Seminar Series series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsType the title of a new list here Caius-Trinity MedSoc Talks - 'The Future of Medicine' Exploring modern South Asian history with visual research methods: theories and practices DevBio Queens' Linguistics Fest 2012 Making more from your research – Impact & valueOther talksThe Digital Doctor: Hope, Hype, and Harm at the Dawn of Medicine’s Computer Age CANCELLED in solidarity with strike action: Permanent Sovereignty over Natural Resources and the Unsettling of Mainstream Narratives of International Legal History From dry to wet granular media Colorectal cancer. Part 1. Presentation, Diagnosis and Intervention. Part 2. Cellular signalling networks in colon cancer and the models to study them - a basic research perspective Trees as keys, ladders, maps: a revisionist history of early systematic trees Index of Suspicion: Predicting Cancer from Prescriptions 'Cryptocurrency and BLOCKCHAIN – PAST, PRESENT AND FUTURE' Atiyah Floer conjecture Single Cell Seminars (October) Validation & testing of novel therapeutic targets to treat osteosarcoma Networks, resilience and complexity St Catharine’s Political Economy Seminar - ‘Bank Credit Rating Changes, Capital Structure Adjustments and Lending’ by Claudia Girardone |