COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Algorithmic stability for heavy-tailed SGD
Algorithmic stability for heavy-tailed SGDAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact nobody. TMLW02 - SGD: stability, momentum acceleration and heavy tails Recent studies have shown that heavy tails can emerge in stochastic optimization and that the heaviness of the tails have links to the generalization error. In this study, we establish novel links between the tail behavior and generalization properties of stochastic gradient descent (SGD), through the lens of algorithmic stability. We develop generalization bounds for a general class of objective functions, which includes non-convex functions as well. Our approach is based on developing Wasserstein stability bounds for heavy-tailed SGD , which we then convert to generalization bounds, indicating a non-monotonic relationship between the generalization error and heavy tails. We support our theory with synthetic and real neural network experiments. This talk is part of the Isaac Newton Institute Seminar Series series. This talk is included in these lists:This talk is not included in any other list Note that ex-directory lists are not shown. |
Other listsTechnology List Type the title of a new list hereOther talksTechnology for Bioelectronic Medicine Applications of proofs to network security MATAR (2023) film screening and Q&A with Director Hassan Akkad Tagging Anglo-Saxon Stone Sculptures Using Multi-Label Image Classification ML Techniques Rethinking Development from the Ethics of Care Gram-scale explosions to inform numerical modelling and FREM development |