University of Cambridge > Talks.cam > CBL Research Talks  > LLM Processes for Regression and Classification

LLM Processes for Regression and Classification

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Cat Spencer.

Machine learning practitioners often face significant challenges in formally integrating their prior knowledge and beliefs into predictive models. Our goal is to build prediction models that can process numerical data and make probabilistic predictions, guided by natural language text which describes a user’s prior knowledge. Large Language Models (LLMs) provide a useful starting point for designing such a tool since they prove 1) an interface where users can incorporate expert insights in natural language and 2) an opportunity for leveraging latent problem-relevant knowledge encoded in LLMs that users may not have themselves. We show how LLMs can compute joint posterior predictive distributions over an arbitrary number of outputs that may be numeric or categorical in settings such as time series forecasting, multi-dimensional regression, black-box optimization, image modeling, and tabular data. Finally, we demonstrate the ability to usefully incorporate text into numerical predictions, showing how the text influences the predictive distribution and improves predictive performance.

References: LLM Processes: Numerical Predictive Distributions Conditioned on Natural Language https://arxiv.org/pdf/2405.12856 JoLT: Joint Probabilistic Predictions on Tabular Data Using LLMs https://arxiv.org/pdf/2502.11877

This talk is part of the CBL Research Talks series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity