Skip to main navigation Skip to search Skip to main content

SPLBoost: An Improved Robust Boosting Algorithm Based on Self-Paced Learning

  • Xi'an Jiaotong University
  • Macau University of Science and Technology

Research output: Contribution to journalArticlepeer-review

28 Scopus citations

Abstract

It is known that boosting can be interpreted as an optimization technique to minimize an underlying loss function. Specifically, the underlying loss being minimized by the traditional AdaBoost is the exponential loss, which proves to be very sensitive to random noise/outliers. Therefore, several boosting algorithms, e.g., LogitBoost and SavageBoost, have been proposed to improve the robustness of AdaBoost by replacing the exponential loss with some designed robust loss functions. In this article, we present a new way to robustify AdaBoost, that is, incorporating the robust learning idea of self-paced learning (SPL) into the boosting framework. Specifically, we design a new robust boosting algorithm based on the SPL regime, that is, SPLBoost, which can be easily implemented by slightly modifying off-the-shelf boosting packages. Extensive experiments and a theoretical characterization are also carried out to illustrate the merits of the proposed SPLBoost.

Original languageEnglish
Article number8943296
Pages (from-to)1556-1570
Number of pages15
JournalIEEE Transactions on Cybernetics
Volume51
Issue number3
DOIs
StatePublished - Mar 2021

Keywords

  • AdaBoost
  • loss function
  • majorization minimization
  • robustness
  • self-paced learning (SPL)

Fingerprint

Dive into the research topics of 'SPLBoost: An Improved Robust Boosting Algorithm Based on Self-Paced Learning'. Together they form a unique fingerprint.

Cite this