TY - JOUR
T1 - ReLU deep neural networks from the hierarchical basis perspective[Formula presented]
AU - He, Juncai
AU - Li, Lin
AU - Xu, Jinchao
N1 - Funding Information:
This work was partially supported by the Center for Computational Mathematics and Applications (CCMA) at The Pennsylvania State University, the Verne M. William Professorship Fund from The Pennsylvania State University, and the National Science Foundation (Grant No. DMS-1819157 ).
Publisher Copyright:
© 2022 Elsevier Ltd
PY - 2022/8/15
Y1 - 2022/8/15
N2 - We study ReLU deep neural networks (DNNs) by investigating their connections with the hierarchical basis method in finite element methods. First, we show that the approximation schemes of ReLU DNNs for x2 and xy are composition versions of the hierarchical basis approximation for these two functions. Based on this fact, we obtain a geometric interpretation and systematic proof for the approximation result of ReLU DNNs for polynomials, which plays an important role in a series of recent exponential approximation results of ReLU DNNs. Through our investigation of connections between ReLU DNNs and the hierarchical basis approximation for x2 and xy, we show that ReLU DNNs with this special structure can be applied only to approximate quadratic functions. Furthermore, we obtain a concise representation to explicitly reproduce any linear finite element function on a two-dimensional uniform mesh by using ReLU DNNs with only two hidden layers.
AB - We study ReLU deep neural networks (DNNs) by investigating their connections with the hierarchical basis method in finite element methods. First, we show that the approximation schemes of ReLU DNNs for x2 and xy are composition versions of the hierarchical basis approximation for these two functions. Based on this fact, we obtain a geometric interpretation and systematic proof for the approximation result of ReLU DNNs for polynomials, which plays an important role in a series of recent exponential approximation results of ReLU DNNs. Through our investigation of connections between ReLU DNNs and the hierarchical basis approximation for x2 and xy, we show that ReLU DNNs with this special structure can be applied only to approximate quadratic functions. Furthermore, we obtain a concise representation to explicitly reproduce any linear finite element function on a two-dimensional uniform mesh by using ReLU DNNs with only two hidden layers.
UR - http://www.scopus.com/inward/record.url?scp=85133894312&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85133894312&partnerID=8YFLogxK
U2 - 10.1016/j.camwa.2022.06.006
DO - 10.1016/j.camwa.2022.06.006
M3 - Article
AN - SCOPUS:85133894312
SN - 0898-1221
VL - 120
SP - 105
EP - 114
JO - Computers and Mathematics with Applications
JF - Computers and Mathematics with Applications
ER -