PROXSGD: TRAINING STRUCTURED NEURAL NETWORKS UNDER REGULARIZATION AND CONSTRAINTS

  • Yang Yang
  • , Yaxiong Yuan
  • , Avraam Chatzimichailidis
  • , Ruud J.G. van Sloun
  • , Lei Lei
  • , Symeon Chatzinotas

Research output: Contribution to conferencePaperpeer-review

10 Scopus citations

Abstract

In this paper, we consider the problem of training structured neural networks (NN) with nonsmooth regularization (e.g. ℓ1-norm) and constraints (e.g. interval constraints). We formulate training as a constrained nonsmooth nonconvex optimization problem, and propose a convergent proximal-type stochastic gradient descent (ProxSGD) algorithm. We show that under properly selected learning rates, with probability 1, every limit point of the sequence generated by the proposed ProxSGD algorithm is a stationary point. Finally, to support the theoretical analysis and demonstrate the flexibility of ProxSGD, we show by extensive numerical tests how ProxSGD can be used to train either sparse or binary neural networks through an adequate selection of the regularization function and constraint set.

Original languageEnglish
StatePublished - 2020
Externally publishedYes
Event8th International Conference on Learning Representations, ICLR 2020 - Addis Ababa, Ethiopia
Duration: 30 Apr 2020 → …

Conference

Conference8th International Conference on Learning Representations, ICLR 2020
Country/TerritoryEthiopia
CityAddis Ababa
Period30/04/20 → …

Fingerprint

Dive into the research topics of 'PROXSGD: TRAINING STRUCTURED NEURAL NETWORKS UNDER REGULARIZATION AND CONSTRAINTS'. Together they form a unique fingerprint.

Cite this