University of Cambridge > Talks.cam > Language Technology Lab Seminars > Pretraining, Instruction Tuning, Alignment: Towards Building Large Language Models from First Principles

Pretraining, Instruction Tuning, Alignment: Towards Building Large Language Models from First Principles

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Panagiotis Fytas.

Recently, the field has been greatly impressed and inspired by Large Language Models (LLMs). LLMs’ multi-dimensional abilities are significantly beyond many AI researchers’ and practitioners’ expectations and thus reshaping the AI research paradigm. A natural question is how LLMs get there, and where these fantastic abilities come from. In this talk, we try to dissect the strong LLMs’ capabilities and trace them to their sources. We first review the generic recipe for building large language models from first principles. Then we discuss recipes for improving language models’ reasoning capabilities. Finally, we consider further improvements by complexity-based prompting, distilling chain-of-thought, and learning from AI feedback.

This talk is part of the Language Technology Lab Seminars series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity