In this paper we characterize the set of functions that can be represented by infinite width neural networks with RePU activation function max(0,x), when the network coefficients are regularized by an ℓ (quasi)norm. Compared to the more well-known ReLU activation function (which corresponds to p=1), the RePU activation functions exhibit a greater degree of smoothness which makes them preferable in several applications. Our main result shows that such representations are possible for a given function if and only if the function is κ-order Lipschitz and its R-norm is finite. This extends earlier work on this topic that has been restricted to the case of the ReLU activation function and coefficient bounds with respect to the ℓ norm. Since for q<2, ℓ regularizations are known to promote sparsity, our results also shed light on the ability to obtain sparse neural network representations.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1016/j.neunet.2022.09.005 | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!