University of Cambridge > Talks.cam > Data Science and AI in Medicine  > Context Aware Deep Learning for Multi Modal Depression Detection

Context Aware Deep Learning for Multi Modal Depression Detection

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Pietro Lio.

In this talk, we focus on automated approaches to detect depression from clinical interviews using multi-modal machine learning (ML). Our approach differentiates from other successful ML methods such as context-aware analysis through feature engineering and end-to-end deep neural networks for depression detection utilizing the Distress Analysis Interview Corpus. We propose a novel method that incorporates: (1) pre-trained Transformer combined with data augmentation based on topic modeling for textual data; and (2) deep 1D convolutional neural network (CNN) for acoustic feature modeling. The simulation results demonstrate the effectiveness of the proposed method for training multi-modal deep learning models. Our deep 1D CNN and Transformer models achieved state-of-the-art performance for audio and text modalities respectively. Combining them in a multi-modal framework also outperforms state-of-the-art for the combined setting.

This work was previously presented as an Oral at International Conference on Acoustics, Speech and Signal Processing 2019, United Kingdom

This talk is part of the Data Science and AI in Medicine series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity