On the Banach Spaces Associated with Multi-Layer ReLU Networks: Function Representation, Approximation Theory and Gradient Descent Dynamics
Weinan E , Stephan Wojtowytsch
CSIAM Trans. Appl. Math. ›› 2020, Vol. 1 ›› Issue (3) : 387 -440.
On the Banach Spaces Associated with Multi-Layer ReLU Networks: Function Representation, Approximation Theory and Gradient Descent Dynamics
We develop Banach spaces for ReLU neural networks of finite depth L and infinite width. The spaces contain all finite fully connected L-layer networks and their L2-limiting objects under bounds on the natural path-norm. Under this norm, the unit ball in the space for L-layer networks has low Rademacher complexity and thus favorable generalization properties. Functions in these spaces can be approximated by multi-layer neural networks with dimension-independent convergence rates.
The key to this work is a new way of representing functions in some form of expec-tations, motivated by multi-layer neural networks. This representation allows us to define a new class of continuous models for machine learning. We show that the gra-dient flow defined this way is the natural continuous analog of the gradient descent dynamics for the associated multi-layer neural networks. We show that the path-norm increases at most polynomially under this continuous gradient flow dynamics.
Barron space / multi-layer space / deep neural network / representations of functions / machine learning / infinitely wide network / ReLU activation / Banach space / path-norm / continu-ous gradient descent dynamics / index representation
/
| 〈 |
|
〉 |