Theory of the Frequency Principle for General Deep Neural Networks

Tao Luo , Zheng Ma , Zhi-Qin John Xu , Yaoyu Zhang

CSIAM Trans. Appl. Math. ›› 2021, Vol. 2 ›› Issue (3) : 484 -507.

PDF (48KB)
CSIAM Trans. Appl. Math. ›› 2021, Vol. 2 ›› Issue (3) : 484 -507. DOI: 10.4208/csiam-am.SO-2020-0005
research-article

Theory of the Frequency Principle for General Deep Neural Networks

Author information +
History +
PDF (48KB)

Abstract

Along with fruitful applications of Deep Neural Networks (DNNs) to realistic problems, recently, empirical studies reported a universal phenomenon of Frequency Principle (F-Principle), that is, a DNN tends to learn a target function from low to high frequencies during the training. The F-Principle has been very useful in providing both qualitative and quantitative understandings of DNNs. In this paper, we rigorously investigate the F-Principle for the training dynamics of a general DNN at three stages: initial stage, intermediate stage, and final stage. For each stage, a theorem is provided in terms of proper quantities characterizing the F-Principle. Our results are general in the sense that they work for multilayer networks with general activation functions, population densities of data, and a large class of loss functions. Our work lays a theoretical foundation of the F-Principle for a better understanding of the training process of DNNs.

Keywords

Frequency principle / Deep Neural Networks / dynamical system / training process

Cite this article

Download citation ▾
Tao Luo, Zheng Ma, Zhi-Qin John Xu, Yaoyu Zhang. Theory of the Frequency Principle for General Deep Neural Networks. CSIAM Trans. Appl. Math., 2021, 2(3): 484-507 DOI:10.4208/csiam-am.SO-2020-0005

登录浏览全文

4963

注册一个新账户 忘记密码

References

AI Summary AI Mindmap
PDF (48KB)

208

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/