当前位置: X-MOL 学术Neural Comput. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Deep Network with Approximation Error Being Reciprocal of Width to Power of Square Root of Depth
Neural Computation ( IF 2.7 ) Pub Date : 2021-01-29 , DOI: 10.1162/neco_a_01364
Zuowei Shen 1 , Haizhao Yang 1 , Shijun Zhang 1
Affiliation  

A new network with super-approximation power is introduced. This network is built with Floor (x) or ReLU (max{0,x}) activation function in each neuron; hence, we call such networks Floor-ReLU networks. For any hyperparameters NN+ and LN+, we show that Floor-ReLU networks with width max{d,5N+13} and depth 64dL+3 can uniformly approximate a Hölder function f on [0,1]d with an approximation error 3λdα/2N-αL, where α(0,1] and λ are the Hölder order and constant, respectively. More generally for an arbitrary continuous function f on [0,1]d with a modulus of continuity ωf(·), the constructive approximation rate is ωf(dN-L)+2ωf(d)N-L. As a consequence, this new class of networks overcomes the curse of dimensionality in approximation power when the variation of ωf(r) as r0 is moderate (e.g., ωf(r)rα for Hölder continuous functions), since the major term to be considered in our approximation rate is essentially d times a function of N and L independent of d within the modulus of continuity.



中文翻译:

具有近似误差为宽度与深度平方根次方的倒数的深度网络

介绍了一种具有超逼近能力的新网络。这个网络是用 Floor (X) 或 ReLU (最大限度{0,X}) 每个神经元的激活函数;因此,我们称这种网络为 Floor-ReLU 网络。对于任何超参数NN+N+,我们展示了具有宽度的 Floor-ReLU 网络 最大限度{d,5N+13} 和深度 64d+3 可以一致地逼近一个 Hölder 函数 F[0,1]d 有近似误差 3λdα/2N——α, 在哪里 α(0,1]λ分别是 Hölder 阶和常数。更一般地用于任意连续函数F[0,1]d 具有连续性模数 ωF(·),建设性逼近率为 ωF(dN——)+2ωF(d)N——. 因此,当ωF(r) 作为 r0 是中等的(例如, ωF(r)rα 对于 Hölder 连续函数),因为在我们的近似率中要考虑的主要项本质上是 d 次函数 N 独立于 d 在连续性模数内。

更新日期:2021-01-31
down
wechat
bug