COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > Isaac Newton Institute Seminar Series > Algorithmic stability for heavy-tailed SGD
Algorithmic stability for heavy-tailed SGDAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact nobody. TMLW02 - SGD: stability, momentum acceleration and heavy tails Recent studies have shown that heavy tails can emerge in stochastic optimization and that the heaviness of the tails have links to the generalization error. In this study, we establish novel links between the tail behavior and generalization properties of stochastic gradient descent (SGD), through the lens of algorithmic stability. We develop generalization bounds for a general class of objective functions, which includes non-convex functions as well. Our approach is based on developing Wasserstein stability bounds for heavy-tailed SGD , which we then convert to generalization bounds, indicating a non-monotonic relationship between the generalization error and heavy tails. We support our theory with synthetic and real neural network experiments. This talk is part of the Isaac Newton Institute Seminar Series series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsCAMbFIRE email hosting uk Special LectureOther talksEarly phase neuroplasticity induced by transcranial ultrasound stimulation Introduction to statistical learning theory, SGD, neural networks Translating the biology of epigenetics into clinical applications for cancer Smyth’s conjecture and a probabilistic local-to-global principle |