COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > CUED Speech Group Seminars > Paraphrastic Language Models / Structured SVMs for ASR
Paraphrastic Language Models / Structured SVMs for ASRAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Catherine Breslin. PARAPHRASTIC LANGUAGE MODELS , Andrew Liu In natural languages multiple word sequences can represent the same underlying meaning. Only modelling the observed surface word sequence can result in poor context coverage, for example, when using N-gram language models (LM). To handle this issue, this paper presents a novel form of language model, the paraphrastic LM. A phrase level paraphrase model that is statistically learned from standard text data is used to generate paraphrase variants. LM probabilities are then estimated by maximizing their marginal probability. Significant error rate reductions of 0.5%-0.6% absolute were obtained over the baseline N-gram LMs on two state-of-the-art recognition tasks for English conversational telephone speech and Mandarin Chinese broadcast speech using a paraphrastic multi-level LM modelling both word and phrase sequences. When it is further combined with word and phrase level neural network LMs, significant error rate reduction of 0.9% absolute (9% relative) and 0.5% absolute (5% relative) were obtained over the baseline N-gram and neural network LMs respectively. STRUCTURED SV Ms FOR ASR , Austin Zhang Combining generative and discriminative models offers a flexible sequence classification framework. This talk describes a structured support vector machines (S-SVM) approach in this framework suitable for medium to large vocabulary speech recognition. An important aspect of S-SVMs is the form of the joint feature spaces. Here, generative models, hidden Markov models, are used to obtain the features. To apply this form of combined generative and discriminative model to speech recognition tasks, a number of issues need to be addressed. First, the features extracted are a function of the segmentation of the utterance. A Viterbi-like scheme for obtaining the “optimal” segmentation is described. Second, we will show that S-SVMs can be viewed as large margin trained log linear models. Finally to speed up the training process, a 1-slack algorithm, caching competing hypotheses and parallelization strategies will also be presented. This talk is part of the CUED Speech Group Seminars series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsTackling Obesity with Big Data: methods & models - One Day Seminar University of Cambridge, New Zealand Studies Seminar Group Logic & Semantics for Dummies ICT4D: ICT for Development Type the title of a new list here Managing Open Data with an Industrial PartnerOther talksCambridge-Lausanne Workshop 2018 - Day 1 Not 'just a GP' Dynamical large deviations in glassy systems Treatment Centre Simulation ADMM for Exploiting Structure in MPC Problems Slaying (or at least taming) a dreadful monster: Louis de Serres' treatise of 1625 for women suffering from infertility ***PLEASE NOTE THIS SEMINAR IS CANCELLED*** An approach to the four colour theorem via Donaldson- Floer theory Microtubule Modulation of Myocyte Mechanics Molecular mechanisms of cardiomyopathies in patients with severe non-ischemic heart failure 'Politics in Uncertain Times: What will the world look like in 2050 and how do you know? Seminar – The Cambridge Sustainable Food Hub |