Ohad Shamir (Weizmann) - Training Neural Networks: The Bigger the Better?Return to Full Calendar
- November 1, 2019 at 10:30am - 11:30am
- JCL, Rm. 390
- Event Audience:
Speaker: Ohad Shamir Faculty Member, Department of Computer Science and Applied Mathematics, Weizmann Institute
Ohad Shamir is a faculty member at the Department of Computer Science and Applied Mathematics at the Weizmann Institute. He received his PhD in 2010 at the Hebrew University, and between 2010-2013 and 2017-2018 was a researcher at Microsoft Research in Boston. His research focuses on theoretical machine learning, in areas such as theory of deep learning, learning with information and communication constraints, and topics at the intersection of machine learning and optimization. He received several awards, and served as program co-chair of COLT as well as a member of its steering committee.
Abstract: Training Neural Networks: The Bigger the Better?
Artificial neural networks are nowadays routinely trained to solve challenging learning tasks, but our theoretical understanding of this phenomenon remains quite limited. One increasingly popular approach, which is aligned with practice, is to study how making the network sufficiently large (a.k.a. ``over-parameterized'') makes the associated training problem easier. In this talk, I'll describe some of the possibilities and challenges in understanding neural networks using this approach. Based on joint works with Itay Safran and Gilad Yehudai.
Sponsor: CS / Toyota Technological Institute of Chicago Machine Learning Seminar Series