跳到主要导航 跳到搜索 跳到主要内容

Deep neural networks for rotation-invariance approximation and learning

科研成果: 期刊稿件文章同行评审

32 引用 (Scopus)

摘要

Based on the tree architecture, the objective of this paper is to design deep neural networks with two or more hidden layers (called deep nets) for realization of radial functions so as to enable rotational invariance for near-optimal function approximation in an arbitrarily high-dimensional Euclidian space. It is shown that deep nets have much better performance than shallow nets (with only one hidden layer) in terms of approximation accuracy and learning capabilities. In particular, for learning radial functions, it is shown that near-optimal rate can be achieved by deep nets but not by shallow nets. Our results illustrate the necessity of depth in neural network design for realization of rotation-invariance target functions.

源语言英语
页(从-至)737-772
页数36
期刊Analysis and Applications
17
5
DOI
出版状态已出版 - 1 9月 2019
已对外发布

学术指纹

探究 'Deep neural networks for rotation-invariance approximation and learning' 的科研主题。它们共同构成独一无二的指纹。

引用此