Talks.cam will close on 1 July 2026, further information is available on the UIS Help Site
 

University of Cambridge > Talks.cam > NLIP Seminar Series > Efficient Pre-Training and Inference Methods for Language Models

Efficient Pre-Training and Inference Methods for Language Models

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Suchir Salhan.

This seminar is rescheduled to a later date in November.

Abstract: With the advancement of large language models (LLMs), their high training and inference costs have become a major bottleneck. This report focuses on cutting-edge algorithms to improve LLM efficiency. For pre-training, we will discuss data optimization methods that accelerate training by enhancing data quality. For inference, we will explore model compression (knowledge distillation) and architecture optimization (efficient attention mechanisms) as pathways to next-generation efficient model design.

This talk is part of the NLIP Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity