COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Language Technology Lab Seminars > Pretraining, Instruction Tuning, Alignment: Towards Building Large Language Models from First Principles
Pretraining, Instruction Tuning, Alignment: Towards Building Large Language Models from First PrinciplesAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Panagiotis Fytas. Recently, the field has been greatly impressed and inspired by Large Language Models (LLMs). LLMs’ multi-dimensional abilities are significantly beyond many AI researchers’ and practitioners’ expectations and thus reshaping the AI research paradigm. A natural question is how LLMs get there, and where these fantastic abilities come from. In this talk, we try to dissect the strong LLMs’ capabilities and trace them to their sources. We first review the generic recipe for building large language models from first principles. Then we discuss recipes for improving language models’ reasoning capabilities. Finally, we consider further improvements by complexity-based prompting, distilling chain-of-thought, and learning from AI feedback. This talk is part of the Language Technology Lab Seminars series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsGiving What We Can: Cambridge ECNM Group, Department of Materials Science and Metallurgy West African Archaeology; Papers in Honour of Thurstan ShawOther talksInstabilities in Liquid Crystal Elastomers Unveiling the Cosmic Dawn and the Epoch of Re-ionization with Radio Observations The homology of big mapping class groups Social structure, collective decision-making and responses to the rise of inequality in animal societies Looking into the first billion years with JWST Introduction by Chris Breward & Camp Rules |