University of Cambridge > Talks.cam > NLIP Seminar Series > Interpretability in NLP: Moving Beyond Vision

Interpretability in NLP: Moving Beyond Vision

Add to your list(s) Download to your calendar using vCal

  • UserShuoyang Ding (Johns Hopkins University)
  • ClockFriday 28 May 2021, 13:00-14:00
  • HouseVirtual (Zoom).

If you have a question about this talk, please contact Huiyuan Xie.

Note unusual time

Join Zoom Meeting https://cl-cam-ac-uk.zoom.us/j/92766937414?pwd=bHJ1TDRqbHRHN0l0eEdPUkxHNlVYQT09

Meeting ID: 927 6693 7414 Passcode: 751190

Deep neural network models have been extremely successful for natural language processing (NLP) applications in recent years, but one complaint they often suffer from is their lack of interpretability. On the other hand, the field of computer vision has navigated their own way of improving interpretability for deep learning models, most notably with post-hoc interpretation methods such as saliency. In this talk, we investigate the possibility of deploying these interpretation methods to natural language processing applications. Our study covers common NLP applications such as language modeling and neural machine translation, and we stress the necessity of quantitative evaluations of interpretations apart from qualitative evaluations. We show that this adaptation is feasible at least in some scenarios, while also pointing out some shortcomings of the current practice that may shed light on future research directions.

This talk is part of the NLIP Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity