Better Approximations of High Dimensional Smooth Functions by Deep Neural Networks with Rectified Power Units
Year: 2020
Author: Bo Li, Shanshan Tang, Haijun Yu
Communications in Computational Physics, Vol. 27 (2020), Iss. 2 : pp. 379–411
Abstract
Deep neural networks with rectified linear units (ReLU) are getting more and more popular due to their universal representation power and successful applications. Some theoretical progress regarding the approximation power of deep ReLU network for functions in Sobolev space and Korobov space have recently been made by [D. Yarotsky, Neural Network, 94:103-114, 2017] and [H. Montanelli and Q. Du, SIAM J Math. Data Sci., 1:78-92, 2019], etc. In this paper, we show that deep networks with rectified power units (RePU) can give better approximations for smooth functions than deep ReLU networks. Our analysis bases on classical polynomial approximation theory and some efficient algorithms proposed in this paper to convert polynomials into deep RePU networks of optimal size with no approximation error. Comparing to the results on ReLU networks, the sizes of RePU networks required to approximate functions in Sobolev space and Korobov space with an error tolerance ε, by our constructive proofs, are in general $O$($log\frac{1}{ε}$) times smaller than the sizes of corresponding ReLU networks constructed in most of the existing literature. Comparing to the classical results of Mhaskar [Mhaskar, Adv. Comput. Math. 1:61-80, 1993], our constructions use less number of activation functions and numerically more stable, they can be served as good initials of deep RePU networks and further trained to break the limit of linear approximation theory. The functions represented by RePU networks are smooth functions, so they naturally fit in the places where derivatives are involved in the loss function.
You do not have full access to this article.
Already a Subscriber? Sign in as an individual or via your institution
Journal Article Details
Publisher Name: Global Science Press
Language: English
DOI: https://doi.org/10.4208/cicp.OA-2019-0168
Communications in Computational Physics, Vol. 27 (2020), Iss. 2 : pp. 379–411
Published online: 2020-01
AMS Subject Headings: Global Science Press
Copyright: COPYRIGHT: © Global Science Press
Pages: 33
Keywords: Deep neural network high dimensional approximation sparse grids rectified linear unit rectified power unit rectified quadratic unit.
Author Details
-
Theoretical guarantees for neural control variates in MCMC
Belomestny, Denis | Goldman, Artur | Naumov, Alexey | Samsonov, SergeyMathematics and Computers in Simulation, Vol. 220 (2024), Iss. P.382
https://doi.org/10.1016/j.matcom.2024.01.019 [Citations: 2] -
q-Neurons: Neuron Activations Based on Stochastic Jackson’s Derivative Operators
Nielsen, Frank | Sun, KeIEEE Transactions on Neural Networks and Learning Systems, Vol. 32 (2021), Iss. 6 P.2782
https://doi.org/10.1109/TNNLS.2020.3005167 [Citations: 1] -
Neural Control of Parametric Solutions for High-Dimensional Evolution PDEs
Gaby, Nathan | Ye, Xiaojing | Zhou, HaominSIAM Journal on Scientific Computing, Vol. 46 (2024), Iss. 2 P.C155
https://doi.org/10.1137/23M1549870 [Citations: 1] -
Simultaneous approximation of a smooth function and its derivatives by deep neural networks with piecewise-polynomial activations
Belomestny, Denis | Naumov, Alexey | Puchkin, Nikita | Samsonov, SergeyNeural Networks, Vol. 161 (2023), Iss. P.242
https://doi.org/10.1016/j.neunet.2023.01.035 [Citations: 7] -
Approximation results for Gradient Flow Trained Shallow Neural Networks in 1d
Gentile, Russell | Welper, GerritConstructive Approximation, Vol. 60 (2024), Iss. 3 P.547
https://doi.org/10.1007/s00365-024-09694-0 [Citations: 0] -
Lyapunov-Net: A Deep Neural Network Architecture for Lyapunov Function Approximation
Gaby, Nathan | Zhang, Fumin | Ye, Xiaojing2022 IEEE 61st Conference on Decision and Control (CDC), (2022), P.2091
https://doi.org/10.1109/CDC51059.2022.9993006 [Citations: 21] -
Approximation rates for neural networks with encodable weights in smoothness spaces
Gühring, Ingo | Raslan, MonesNeural Networks, Vol. 134 (2021), Iss. P.107
https://doi.org/10.1016/j.neunet.2020.11.010 [Citations: 40] -
Near-optimal learning of Banach-valued, high-dimensional functions via deep neural networks
Adcock, Ben | Brugiapaglia, Simone | Dexter, Nick | Moraga, SebastianNeural Networks, Vol. 181 (2025), Iss. P.106761
https://doi.org/10.1016/j.neunet.2024.106761 [Citations: 0] -
The robust physics-informed neural networks for a typical fourth-order phase field model
Zhang, Wen | Li, JianComputers & Mathematics with Applications, Vol. 140 (2023), Iss. P.64
https://doi.org/10.1016/j.camwa.2023.03.016 [Citations: 2] -
Efficient Spectral Methods for Quasi-Equilibrium Closure Approximations of Symmetric Problems on Unit Circle and Sphere
Jiang, Shan | Yu, HaijunJournal of Scientific Computing, Vol. 89 (2021), Iss. 2
https://doi.org/10.1007/s10915-021-01646-1 [Citations: 3] -
Approximation Algorithm Applications in Artificial Intelligence: A Review
Abdul Ameer, Hawraa R. | Jaddoa, Ahmed Sami | AL-Challabi, Hawraa A. | Abdul-Zahra, Dalael Saad2022 Fifth College of Science International Conference of Recent Trends in Information Technology (CSCTIT), (2022), P.43
https://doi.org/10.1109/CSCTIT56299.2022.10145686 [Citations: 0] -
Solving Parametric Partial Differential Equations with Deep Rectified Quadratic Unit Neural Networks
Lei, Zhen | Shi, Lei | Zeng, ChenyuJournal of Scientific Computing, Vol. 93 (2022), Iss. 3
https://doi.org/10.1007/s10915-022-02015-2 [Citations: 5] -
Uncertainty, Constraints, and Decision Making
Why Rectified Power (RePU) Activation Functions are Efficient in Deep Learning: A Theoretical Explanation
Bokati, Laxman | Kreinovich, Vladik | Baca, Joseph | Rovelli, Natasha2023
https://doi.org/10.1007/978-3-031-36394-8_2 [Citations: 0] -
Smooth Function Approximation by Deep Neural Networks with General Activation Functions
Ohn, Ilsang | Kim, YongdaiEntropy, Vol. 21 (2019), Iss. 7 P.627
https://doi.org/10.3390/e21070627 [Citations: 46] -
Discovering Phase Field Models from Image Data with the Pseudo-Spectral Physics Informed Neural Networks
Zhao, Jia
Communications on Applied Mathematics and Computation, Vol. 3 (2021), Iss. 2 P.357
https://doi.org/10.1007/s42967-020-00105-2 [Citations: 5] -
Constructing custom thermodynamics using deep learning
Chen, Xiaoli | Soh, Beatrice W. | Ooi, Zi-En | Vissol-Gaudin, Eleonore | Yu, Haijun | Novoselov, Kostya S. | Hippalgaonkar, Kedar | Li, QianxiaoNature Computational Science, Vol. 4 (2023), Iss. 1 P.66
https://doi.org/10.1038/s43588-023-00581-5 [Citations: 1] -
A convergent Deep Learning algorithm for approximation of polynomials
Després, Bruno
Comptes Rendus. Mathématique, Vol. 361 (2023), Iss. G6 P.1029
https://doi.org/10.5802/crmath.462 [Citations: 0] -
Numerical Analysis Meets Machine Learning
Learning smooth functions in high dimensions
Adcock, Ben | Brugiapaglia, Simone | Dexter, Nick | Moraga, Sebastian2024
https://doi.org/10.1016/bs.hna.2024.05.001 [Citations: 1] -
Sparse Approximation of Triangular Transports, Part I: The Finite-Dimensional Case
Zech, Jakob | Marzouk, YoussefConstructive Approximation, Vol. 55 (2022), Iss. 3 P.919
https://doi.org/10.1007/s00365-022-09569-2 [Citations: 5] -
OnsagerNet: Learning stable and interpretable dynamics using a generalized Onsager principle
Yu, Haijun | Tian, Xinyuan | E, Weinan | Li, QianxiaoPhysical Review Fluids, Vol. 6 (2021), Iss. 11
https://doi.org/10.1103/PhysRevFluids.6.114402 [Citations: 25] -
Constructive Deep ReLU Neural Network Approximation
Herrmann, Lukas | Opschoor, Joost A. A. | Schwab, ChristophJournal of Scientific Computing, Vol. 90 (2022), Iss. 2
https://doi.org/10.1007/s10915-021-01718-2 [Citations: 12] -
Optimal approximation of infinite-dimensional holomorphic functions
Adcock, Ben | Dexter, Nick | Moraga, SebastianCalcolo, Vol. 61 (2024), Iss. 1
https://doi.org/10.1007/s10092-023-00565-x [Citations: 3] -
ChebNet: Efficient and Stable Constructions of Deep Neural Networks with Rectified Power Units via Chebyshev Approximation
Tang, Shanshan | Li, Bo | Yu, HaijunCommunications in Mathematics and Statistics, Vol. (2024), Iss.
https://doi.org/10.1007/s40304-023-00392-0 [Citations: 0] -
Exponential ReLU DNN Expression of Holomorphic Maps in High Dimension
Opschoor, J. A. A. | Schwab, Ch. | Zech, J.Constructive Approximation, Vol. 55 (2022), Iss. 1 P.537
https://doi.org/10.1007/s00365-021-09542-5 [Citations: 33] -
Two-layer networks with the $$\text {ReLU}^k$$ activation function: Barron spaces and derivative approximation
Li, Yuanyuan | Lu, Shuai | Mathé, Peter | Pereverzev, Sergei V.Numerische Mathematik, Vol. 156 (2024), Iss. 1 P.319
https://doi.org/10.1007/s00211-023-01384-6 [Citations: 1]