当前位置: X-MOL 学术IEEE Trans. Pattern Anal. Mach. Intell. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
You Only Search Once: Single Shot Neural Architecture Search via Direct Sparse Optimization
IEEE Transactions on Pattern Analysis and Machine Intelligence ( IF 20.8 ) Pub Date : 8-31-2020 , DOI: 10.1109/tpami.2020.3020300
xinbang zhang , zehao huang , Naiyan Wang , Shiming XIANG , Chunhong Pan

Recently neural architecture search (NAS) has raised great interest in both academia and industry. However, it remains challenging because of its huge and non-continuous search space. Instead of applying evolutionary algorithm or reinforcement learning as previous works, this paper proposes a direct sparse optimization NAS (DSO-NAS) method. The motivation behind DSO-NAS is to address the task in the view of model pruning. To achieve this goal, we start from a completely connected block, and then introduce scaling factors to scale the information flow between operations. Next, sparse regularizations are imposed to prune useless connections in the architecture. Lastly, an efficient and theoretically sound optimization method is derived to solve it. Our method enjoys both advantages of differentiability and efficiency, therefore it can be directly applied to large datasets like ImageNet and tasks beyond classification. Particularly, on the CIFAR-10 dataset, DSO-NAS achieves an average test error 2.74 percent, while on the ImageNet dataset DSO-NAS achieves 25.4 percent test error under 600M FLOPs with 8 GPUs in 18 hours. As for semantic segmentation task, DSO-NAS also achieve competitive result compared with manually designed architectures on the PASCAL VOC dataset. Code is available at https://github.com/XinbangZhang/DSO-NAS .

中文翻译:


您只需搜索一次:通过直接稀疏优化进行单次神经架构搜索



最近,神经架构搜索(NAS)引起了学术界和工业界的极大兴趣。然而,由于其巨大且不连续的搜索空间,它仍然具有挑战性。本文没有像以前的工作那样应用进化算法或强化学习,而是提出了一种直接稀疏优化NAS(DSO-NAS)方法。 DSO-NAS 背后的动机是从模型剪枝的角度解决该任务。为了实现这个目标,我们从一个完全连接的块开始,然后引入缩放因子来缩放操作之间的信息流。接下来,施加稀疏正则化来修剪架构中无用的连接。最后,推导了一种高效且理论上合理的优化方法来解决该问题。我们的方法兼具可微性和效率的优点,因此它可以直接应用于 ImageNet 等大型数据集和分类之外的任务。特别是,在 CIFAR-10 数据集上,DSO-NAS 的平均测试误差为 2.74%,而在 ImageNet 数据集上,DSO-NAS 在 18 小时内在 8 个 GPU 的 600M FLOP 下实现了 25.4% 的测试误差。至于语义分割任务,与 PASCAL VOC 数据集上手动设计的架构相比,DSO-NAS 也取得了有竞争力的结果。代码可在 https://github.com/XinbangZhang/DSO-NAS 获取。
更新日期:2024-08-22
down
wechat
bug