COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > CUED Speech Group Seminars > Multi-Head State Space Model for Sequence Modeling
Multi-Head State Space Model for Sequence ModelingAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Dr Kate Knill. Recently, state space models (SSMs) have shown promising results on sequence modeling tasks. However, a potential challenge of existing works is that SSMs are usually introduced or initialized in a homogeneous way, encouraging the model to only capture similar temporal dynamics on different features. In this talk, we propose a multi-head state space model (MSSM), in which parallel heads are introduced to learn different temporal dynamics on sequence data. Furthermore, we propose a novel variant of the Transformer, referred to as the Stateformer, which combines MSS Ms with attention. Experiments on large-scale automatic speech recognition (ASR) and language modeling tasks show the MSSM outperforming a range of attention-based baselines. The Stateformer further improves performance, achieving the state-of-the-art performance on the LibriSpeech ASR task. Research performed in Research Internship at Meta (AI Speech), California. This talk is part of the CUED Speech Group Seminars series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsrc781 Economics i-TeamsOther talksA day in the life of an adsorbate: Observing dynamics in an open quantum system Joined Up Thinking Battery Metals: Formation and the race to Net Zero Capitalism in a Colonial Context: African Merchants in Lagos, 1851ā1921 Gateway The Whitham Approach to the cā0 limit of The Lieb-Liniger Model and Generalized Hydrodynamics |