摘要
In this paper, we consider the problem of training structured neural networks (NN) with nonsmooth regularization (e.g. ℓ1-norm) and constraints (e.g. interval constraints). We formulate training as a constrained nonsmooth nonconvex optimization problem, and propose a convergent proximal-type stochastic gradient descent (ProxSGD) algorithm. We show that under properly selected learning rates, with probability 1, every limit point of the sequence generated by the proposed ProxSGD algorithm is a stationary point. Finally, to support the theoretical analysis and demonstrate the flexibility of ProxSGD, we show by extensive numerical tests how ProxSGD can be used to train either sparse or binary neural networks through an adequate selection of the regularization function and constraint set.
| 源语言 | 英语 |
|---|---|
| 出版状态 | 已出版 - 2020 |
| 已对外发布 | 是 |
| 活动 | 8th International Conference on Learning Representations, ICLR 2020 - Addis Ababa, 埃塞俄比亚 期限: 30 4月 2020 → … |
会议
| 会议 | 8th International Conference on Learning Representations, ICLR 2020 |
|---|---|
| 国家/地区 | 埃塞俄比亚 |
| 市 | Addis Ababa |
| 时期 | 30/04/20 → … |
学术指纹
探究 'PROXSGD: TRAINING STRUCTURED NEURAL NETWORKS UNDER REGULARIZATION AND CONSTRAINTS' 的科研主题。它们共同构成独一无二的指纹。引用此
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver