TY - JOUR
T1 - A Simple and Effective Architecture Selection Method for Differentiable Architecture Search
AU - Chen, Boxu
AU - Yang, Le
AU - Zheng, Ziwei
AU - Li, Fan
AU - Song, Shiji
AU - Huang, Gao
AU - Shen, Chao
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2025
Y1 - 2025
N2 - Although differentiable architecture search (DARTS) improves the searching efficiency of neural architecture search (NAS), the widely applied magnitude-based selection method of DARTS can frequently lead to deteriorating architectures with degenerated performance. Most existing works propose to address this issue by improving the supernet's optimization to guarantee the applicability of the magnitude-based method, while little attention has been paid to the selection criterion to obtain the final architecture. In this brief, we introduce a novel, simple, and effective architecture selection method, Manda (Magnitudes and activations), which estimates the contribution of an operation in an optimized supernet by both its architecture parameter's magnitude and corresponding generated activation. Notably, Manda can effectively address the notorious degeneration issue in DARTS without any modification of the supernet's optimization procedure, indicating the instability in DARTS can be attributed to the widely applied magnitude-based selection method. The experimental results on both NAS-Bench-201 and DARTS search spaces show the effectiveness of our method.
AB - Although differentiable architecture search (DARTS) improves the searching efficiency of neural architecture search (NAS), the widely applied magnitude-based selection method of DARTS can frequently lead to deteriorating architectures with degenerated performance. Most existing works propose to address this issue by improving the supernet's optimization to guarantee the applicability of the magnitude-based method, while little attention has been paid to the selection criterion to obtain the final architecture. In this brief, we introduce a novel, simple, and effective architecture selection method, Manda (Magnitudes and activations), which estimates the contribution of an operation in an optimized supernet by both its architecture parameter's magnitude and corresponding generated activation. Notably, Manda can effectively address the notorious degeneration issue in DARTS without any modification of the supernet's optimization procedure, indicating the instability in DARTS can be attributed to the widely applied magnitude-based selection method. The experimental results on both NAS-Bench-201 and DARTS search spaces show the effectiveness of our method.
KW - AutoML
KW - differentiable architecture search
KW - Neural architecture search
KW - neural networks
UR - https://www.scopus.com/pages/publications/105017181528
U2 - 10.1109/TAI.2025.3610384
DO - 10.1109/TAI.2025.3610384
M3 - 文章
AN - SCOPUS:105017181528
SN - 2691-4581
JO - IEEE Transactions on Artificial Intelligence
JF - IEEE Transactions on Artificial Intelligence
ER -