Skip to main navigation Skip to search Skip to main content

Online efficient learning with quantized KLMS and L 1 regularization

  • Badong Chen
  • , Songlin Zhao
  • , Sohan Seth
  • , Jose C. Principe

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

26 Scopus citations

Abstract

In a recent work, we have proposed the quantized kernel least mean square (QKLMS) algorithm, which is quite effective in online learning sequentially a nonlinear mapping with a slowly growing radial basis function (RBF) structure. In this paper, in order to further reduce the network size, we propose a sparse QKLMS algorithm, which is derived by adding a sparsity inducing l 1 norm penalty of the coefficients to the squared error cost. Simulation examples show that the new algorithm works efficiently, and results in a much sparser network while preserving a desirable performance.

Original languageEnglish
Title of host publication2012 International Joint Conference on Neural Networks, IJCNN 2012
DOIs
StatePublished - 2012
Externally publishedYes
Event2012 Annual International Joint Conference on Neural Networks, IJCNN 2012, Part of the 2012 IEEE World Congress on Computational Intelligence, WCCI 2012 - Brisbane, QLD, Australia
Duration: 10 Jun 201215 Jun 2012

Publication series

NameProceedings of the International Joint Conference on Neural Networks

Conference

Conference2012 Annual International Joint Conference on Neural Networks, IJCNN 2012, Part of the 2012 IEEE World Congress on Computational Intelligence, WCCI 2012
Country/TerritoryAustralia
CityBrisbane, QLD
Period10/06/1215/06/12

Keywords

  • QKLMS
  • kernel adaptive filtering
  • l norm penalty
  • online learning

Fingerprint

Dive into the research topics of 'Online efficient learning with quantized KLMS and L 1 regularization'. Together they form a unique fingerprint.

Cite this