University of Cambridge > Talks.cam > Astro Data Science Discussion Group > Scaling laws for large time-series models: More data, more parameters

Scaling laws for large time-series models: More data, more parameters

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact km723.

Scaling laws for large language models (LLMs) offer valuable insights into how increasing model size and training data leads to predictable performance improvements. Time series forecasting, which shares a sequential structure similar to language, is well-suited to large-scale transformer architectures. In this talk, I will demonstrate that foundational decoder-only time series transformer models exhibit scaling behaviour analogous to LLMs. I will begin with a general introduction to scaling laws and how they can inform efficient, optimised model training. I will then focus on their specific application to time series data, highlighting the emergence of power law behaviour. Finally, I will discuss the broader implications of these findings, and potential scientific applications.

Related papers: https://arxiv.org/abs/2001.08361 https://arxiv.org/abs/2405.13867

This talk is part of the Astro Data Science Discussion Group series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity