国家天元数学中部中心高性能计算系列报告 | 张景昭 (清华大学交叉信息科学院)

发布时间: 2022-09-27 20:22

报告目:On the (Non)smoothness of Neural Network Training

报告时间:2022-10-11  15:30 - 16:30

报告人:张景昭  清华大学交叉信息科学院

ZOOMID:561 420 9883   密码:tmcc2022

Abstract: In this talk, we will discuss the following questions―why is neural network training non-smooth from an optimization perspective, and how should we analyze convergence in nonsmooth settings? We will start by showing that the non-smoothness is essential to standard neural network training procedures, and that network training converges in an unstable manner. We then provide theoretical models for understanding why optimization in neural network is unstable. We conclude by showing that new definitions of convergence in the nonsmooth settings can reconcile theory with practice.