A Mean-Field Perspective on Neural Network Optimization: From Single-Stage to Multi-Stage Learning

发布者:梁慧丽发布时间:2026-01-13浏览次数:11

报告人


陈宗昊

University College London

时间

2026年1月13日 星期二

下午 14:00-15:00

地点


602会议室



Abstract


Optimization of neural networks is notoriously challenging due to the highly non-convex nature of their parameter landscapes. In recent years, a mean-field perspective has emerged as a powerful framework for understanding neural network training in the large-width limit, where the training of network parameters can be described by a gradient flow over probability measures. This talk will first review recent developments in the mean-field analysis of neural network optimization for single-stage learning problems, such as standard supervised learning.

I will then present my recent work on extending the mean-field framework to multi-stage learning problems, with a particular focus on bilevel optimization in nonparametric instrumental variable (NPIV) regression.

Finally, I will discuss the limitations and open challenges of the mean-field perspective for neural network optimization, and outline several directions for future research.




Biography


图片

陈宗昊目前在英国伦敦大学学院(University College London)基础人工智能中心攻读博士学位,师从 François-Xavier Briol 与 Arthur Gretton。他的研究兴趣聚焦于通过优化与泛化两大视角理解机器学习算法。目前的研究方向包括生成模型、蒙特卡洛方法以及因果推断。在攻读博士之前,他于清华大学电子工程系获得学士学位(2022),并曾获清华大学特等奖学金。



搜索
您想要找的