COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > CCIMI Seminars > Coordinate Descent on the Orthogonal Group for Recurrent Neural Network Training
Coordinate Descent on the Orthogonal Group for Recurrent Neural Network TrainingAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Hamza Fawzi. To address the poor scalability of training algorithms for orthogonal recurrent neural networks, we propose to use a coordinate descent method on the orthogonal group. This algorithm has a cost per iteration that evolves linearly with the number of recurrent states, in contrast with the cubic dependency of typical algorithms such as stochastic Riemannian gradient descent. We numerically show that the Riemannian gradient in recurrent neural network training has an approximately sparse structure. Leveraging this observation, we propose a variant of the proposed algorithm that relies on Gauss-Southwell coordinate selection. Experiments on a benchmark recurrent neural network training problem show that the proposed approach is a very promising step towards the training of orthogonal recurrent neural networks with big architectures. Join Zoom Meeting https://maths-cam-ac-uk.zoom.us/j/93776043287?pwd=UDIrNDdkeUU1NmFtZXpNUzd6ZjRrdz09 Meeting ID: 937 7604 3287 Passcode: p1Co4skf This talk is part of the CCIMI Seminars series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsActive Materials Young Nanoscientist India Award Winner's Lecture sponsored by Oxford Instruments Zoology Postdoc Summer Seminar SeriesOther talksStatistics Clinic Summer 2022 IV Severe Space Weather At the limits of China: frontiers, borders, and political geography in early modern Sino-Western cartographic exchanges Local yield stress in model amorphous solids Functional Macroevolution of Transcription Factors in Land Plants HOLIDAY BREAK |