跳到主要导航 跳到搜索 跳到主要内容

Penetrating the influence of regularizations on neural network based on information bottleneck theory

  • Xi'an Jiaotong University

科研成果: 期刊稿件文章同行评审

10 引用 (Scopus)

摘要

Regularization is a very effective algorithm to solve overfitting problem in neural network, which improves the generalization ability of the model. However, their working mechanisms and the impact on the model performance have not been fully explored. In this paper, we study and analyze them using information bottleneck theory and one theory from human brain sensory system. We propose a metric to characterise the encoding length of hidden layers, named as AEntry value. Then, we implement extensive experiments on MNIST and FashionMNIST datasets with several commonly used regularization algorithms, and calculate the corresponding AEntry values. We analyze these results and obtain three conclusions. (1) The introduction of regularization influences the encoding of relative features with prediction task in neural network. The early stopping technique avoids introducing unrelated information with the task into the model by stopping the training process as an appropriate iterations. Laplace, Gaussian and Sparse Response regularizations compress the related representation and improve the performance of neural network by introducing the prior information into the model. In contrast, Dropout, Batch Normalization, and Layer Normalization increase the encoding length of features by adopting redundant representation to improve the performance. (2) The encoding of neural network does not satisfy the data processing inequality of information theory, which is mainly caused by redundant coding of extracted features. (3) The overfitting is caused by introducing irrelative information with the target. These results can give us insight into building more efficient regularization algorithm to improve the performance of neural network model.

源语言英语
页(从-至)76-82
页数7
期刊Neurocomputing
393
DOI
出版状态已出版 - 14 6月 2020

学术指纹

探究 'Penetrating the influence of regularizations on neural network based on information bottleneck theory' 的科研主题。它们共同构成独一无二的指纹。

引用此