Journal of Computer Applications

    Next Articles

Survey on differentiable neural architecture search methods

WANG Han, YU Ruiqi, REN Shuang   

  1. School of Computer Science and Technology, Beijing Jiaotong University
  • Received:2025-09-28 Revised:2025-12-09 Online:2025-12-26 Published:2025-12-26
  • About author:WANG Han, born in 2001, M. S. candidate. His research interests include artificial intelligence, quantum computing. YU Ruiqi, born in 2002, Ph. D. candidate. His research interests include quantum computing, machine learning. REN Shuang, born in 1981, Ph. D., associate professor. His research interests include machine learning, quantum computing, 3D computer vision.
  • Supported by:
    National Natural Science Foundation of China (62072025)

可微神经架构搜索方法综述

王涵,于瑞祺,任爽   

  1. 北京交通大学 计算机科学与技术学院
  • 通讯作者: 任爽
  • 作者简介:王涵(2001—),男,山西大同人,硕士研究生,主要研究方向:人工智能、量子计算;于瑞祺(2002—),男,黑龙江鹤岗人,博士研究生,主要研究方向:量子计算、机器学习;任爽(1981—),男,吉林长春人,副教授,博士生导师,博士,CCF会员,主要研究方向:机器学习、量子计算、3D计算机视觉。
  • 基金资助:
    国家自然科学基金资助项目(62072025)

Abstract: Neural architecture search has been recognized as an automated paradigm for designing neural network structures, through which the limitations of conventional trial-and-error–based manual design have been transcended. By this means, the reliance of deep learning model development on expert knowledge has been substantially reduced, while the overall model performance has been significantly enhanced. Among the diverse approaches, Differentiable Neural Architecture Search (DNAS), which leverages the one-shot framework in conjunction with gradient-based optimization, has drastically shortened the search time and has emerged as the predominant research direction. Nevertheless, a systematic review of differentiable methods and their developmental trajectory remains absent. To bridge this gap and provide a reference for subsequent studies, recent advances in differentiable neural architecture search are synthesized and analyzed. The paper begins by introducing the fundamental theories and benchmark datasets of differentiable neural architecture search, followed by an analysis of its existing challenges. A hierarchical review of current algorithms is then provided, and the paper concludes with a discussion on future research directions.

Key words: Differentiable Neural Architecture Search (DNAS), one-shot architecture, super net, gradient descent, architecture collapse, hardware latency 

摘要: 神经架构搜索通过自动化设计神经网络结构,突破传统手动设计的试错局限,显著降低了深度学习模型开发对专业经验的依赖,大幅提升模型表现。在众多方法中,可微神经架构搜索(DNAS)借助one-shot框架与梯度下降优化,大幅缩短了搜索时间,已成为当前研究的主流方向。然而,现有研究仍缺乏对可微方法及其发展脉络的系统化梳理。为了填补相关空白,并为相关研究提供借鉴参考,对近年来出现的可微神经架构搜索方法进行归纳分析。首先介绍可微神经架构搜索算法的基本理论和实验数据集,梳理了可微神经架构搜索的现存挑战,之后对现有算法进行层次化分析,最后对未来发展方向进行展望。

关键词: 可微神经架构搜索, one-shot架构, 超网, 梯度下降, 架构崩溃, 硬件延迟

CLC Number: