University of Cambridge > Talks.cam > Language Technology Lab Seminars > What do sentence transformers know, and how can we find out?

What do sentence transformers know, and how can we find out?

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Panagiotis Fytas.

Transformer architectures specialised for producing full-sentence representations, notably SBERT , often achieve better performance on downstream tasks than sentence embeddings extracted from vanilla BERT . However, compared to Vanilla transformers, we still have a limited understanding of which linguistic properties of the inputs are represented well (or less well) within these models.

In my presentation, I will report on two angles from which we have analyzed SBERT : (a), a black-box testing approach where we build minimal pairs of synthetic sentences to observe and analyze differences in the model’s predictions [1]; and (b), a white-box testing approach where we extend the Integrated Gradients attribution method to the Siamese case. This permits us to decompose model predictions on arbitrary input in terms of the contributions of individual token pairs [2,3].

[1] Dmitry Nikolaev and Sebastian Padó. Representation biases in sentence transformers. Proceedings of EACL . Dubrovnik, Croatia, 2023. https://aclanthology.org/2023.eacl-main.268

[2] Lucas Möller, Dmitry Nikolaev and Sebastian Padó. An Attribution Method for Siamese Encoders. Proceedings of EMNLP . Singapore, 2023. https://aclanthology.org/2023.emnlp-main.980

[3] Lucas Möller, Dmitry Nikolaev and Sebastian Padó. Approximate Attributions for Off-the-Shelf Siamese Transformers. Proceedings of EACL . St Julian’s, Malta, 2024. https://arxiv.org/abs/2402.02883

This talk is part of the Language Technology Lab Seminars series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity