TY - JOUR
T1 - Efficient and robust deep learning with Correntropy-induced loss function
AU - Chen, Liangjun
AU - Qu, Hua
AU - Zhao, Jihong
AU - Chen, Badong
AU - Principe, Jose C.
N1 - Publisher Copyright:
© 2015, The Natural Computing Applications Forum.
PY - 2016/5/1
Y1 - 2016/5/1
N2 - Deep learning systems aim at using hierarchical models to learning high-level features from low-level features. The progress in deep learning is great in recent years. The robustness of the learning systems with deep architectures is however rarely studied and needs further investigation. In particular, the mean square error (MSE), a commonly used optimization cost function in deep learning, is rather sensitive to outliers (or impulsive noises). Robust methods are needed to improve the learning performance and immunize the harmful influences caused by outliers which are pervasive in real-world data. In this paper, we propose an efficient and robust deep learning model based on stacked auto-encoders and Correntropy-induced loss function (CLF), called CLF-based stacked auto-encoders (CSAE). CLF as a nonlinear measure of similarity is robust to outliers and can approximate different norms (from (Formula presented.) to (Formula presented.) ) of data. Essentially, CLF is an MSE in reproducing kernel Hilbert space. Different from conventional stacked auto-encoders, which use, in general, the MSE as the reconstruction loss and KL divergence as the sparsity penalty term, the reconstruction loss and sparsity penalty term in CSAE are both built with CLF. The fine-tuning procedure in CSAE is also based on CLF, which can further enhance the learning performance. The excellent and robust performance of the proposed model is confirmed by simulation experiments on MNIST benchmark dataset.
AB - Deep learning systems aim at using hierarchical models to learning high-level features from low-level features. The progress in deep learning is great in recent years. The robustness of the learning systems with deep architectures is however rarely studied and needs further investigation. In particular, the mean square error (MSE), a commonly used optimization cost function in deep learning, is rather sensitive to outliers (or impulsive noises). Robust methods are needed to improve the learning performance and immunize the harmful influences caused by outliers which are pervasive in real-world data. In this paper, we propose an efficient and robust deep learning model based on stacked auto-encoders and Correntropy-induced loss function (CLF), called CLF-based stacked auto-encoders (CSAE). CLF as a nonlinear measure of similarity is robust to outliers and can approximate different norms (from (Formula presented.) to (Formula presented.) ) of data. Essentially, CLF is an MSE in reproducing kernel Hilbert space. Different from conventional stacked auto-encoders, which use, in general, the MSE as the reconstruction loss and KL divergence as the sparsity penalty term, the reconstruction loss and sparsity penalty term in CSAE are both built with CLF. The fine-tuning procedure in CSAE is also based on CLF, which can further enhance the learning performance. The excellent and robust performance of the proposed model is confirmed by simulation experiments on MNIST benchmark dataset.
KW - Correntropy
KW - Deep learning
KW - Stacked auto-encoders
KW - Unsupervised feature learning
UR - https://www.scopus.com/pages/publications/84928386842
U2 - 10.1007/s00521-015-1916-x
DO - 10.1007/s00521-015-1916-x
M3 - 文章
AN - SCOPUS:84928386842
SN - 0941-0643
VL - 27
SP - 1019
EP - 1031
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 4
ER -