2024年12月21日 星期六 登录 EN

学术活动
Implicit biases of optimization algorithms for neural networks and their effects on generalization
首页 - 学术活动
报告人:
Dr. Chao Ma, Stanford University
邀请人:
Pingbing Ming, Professor
题目:
Implicit biases of optimization algorithms for neural networks and their effects on generalization
时间地点:
9:00-10:00 December 23( Friday),Tencent Meeting ID: 247-520-003
摘要:
Modern neural networks are usually over-parameterized—the number of parameters exceeds the number of training data. In this case the loss functions tend to have many (or even infinite) global minima, which imposes an additional challenge of minima selection on optimization algorithms besides the convergence. Specifically, when training a neural network, the algorithm not only has to find a global minimum, but also needs to select minima with good generalization among many other bad ones. In this talk, we connect the implicit bias of optimization algorithms and the generalization performance via two steps. First, with a linear stability analysis around global minima, we show that stochastic gradient descent (SGD) favors flat and uniform global minima. Then, we build a theoretical connection of flatness and generalization performance based on a special multiplicative structure of neural networks. Together, we show that SGD tends to find global minima with good generalization. Bounds for generalization error and adversarial robustness depending on SGD hyperparameters are derived.