University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Algorithmic stability for heavy-tailed SGD

Algorithmic stability for heavy-tailed SGD

Add to your list(s) Download to your calendar using vCal

  • UserLingjiong Zhu (Florida State University)
  • ClockWednesday 24 April 2024, 14:30-15:15
  • HouseExternal.

If you have a question about this talk, please contact nobody.

TMLW02 - SGD: stability, momentum acceleration and heavy tails

Recent studies have shown that heavy tails can emerge in stochastic optimization and that the heaviness of the tails have links to the generalization error. In this study, we establish novel links between the tail behavior and generalization properties of stochastic gradient descent (SGD), through the lens of algorithmic stability. We develop generalization bounds for a general class of objective functions, which includes non-convex functions as well. Our approach is based on developing Wasserstein stability bounds for heavy-tailed SGD , which we then convert to generalization bounds, indicating a non-monotonic relationship between the generalization error and heavy tails. We support our theory with synthetic and real neural network experiments.   

This talk is part of the Isaac Newton Institute Seminar Series series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2024 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity