Maintained by Difan Deng and Marius Lindauer.
The following list considers papers related to neural architecture search. It is by no means complete. If you miss a paper on the list, please let us know.
Please note that although NAS methods steadily improve, the quality of empirical evaluations in this field are still lagging behind compared to other areas in machine learning, AI and optimization. We would therefore like to share some best practices for empirical evaluations of NAS methods, which we believe will facilitate sustained and measurable progress in the field. If you are interested in a teaser, please read our blog post or directly jump to our checklist.
Transformers have gained increasing popularity in different domains. For a comprehensive list of papers focusing on Neural Architecture Search for Transformer-Based spaces, the awesome-transformer-search repo is all you need.
2025
(Ed.)
PQNAS: Mixed-precision Quantization-aware Neural Architecture Search with Pseudo Quantizer Collection
2025.
@collection{gao-icassp25a,
title = {PQNAS: Mixed-precision Quantization-aware Neural Architecture Search with Pseudo Quantizer},
author = {Tianxiao Gao and Li Guo and Shihao Wang and Shiai Zhu and Dajiang Zhou},
url = {https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=10888233},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
booktitle = {2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
keywords = {},
pubstate = {published},
tppubtype = {collection}
}
(Ed.)
Neural Architecture Search: Tradeoff Between Performance and Efficiency Collection
2025.
@collection{nokey,
title = {Neural Architecture Search: Tradeoff Between Performance and Efficiency},
author = {Tien Dung Nguyen and Nassim Mokhtari and Alexis Nédélec},
url = {https://www.scitepress.org/Papers/2025/132969/132969.pdf},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
booktitle = {Proceedings of the 17th International Conference on Agents and Artificial Intelligence (ICAART 2025)},
keywords = {},
pubstate = {published},
tppubtype = {collection}
}
Fang, Xuwei; Xie, Weisheng; Li, Hui; Zhou, Wenbin; Hang, Chen; Gao, Xiangxiang
DARTS-EAST: an edge-adaptive selection with topology first differentiable architecture selection method Journal Article
In: Applied Intelligence , 2025.
@article{fang-ai25a,
title = {DARTS-EAST: an edge-adaptive selection with topology first differentiable architecture selection method},
author = {Xuwei Fang and Weisheng Xie and Hui Li and Wenbin Zhou and Chen Hang and Xiangxiang Gao
},
url = {https://link.springer.com/article/10.1007/s10489-025-06353-0},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
journal = {Applied Intelligence },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
(Ed.)
2025.
@collection{Friderikos-dte25a,
title = {OPTIMIZED LSTM NEURAL NETWORKS VIA NEURAL ARCHITECTURE SEARCH FOR PREDICTING DAMAGE EVOLUTION IN COMPOSITE LAMINATES},
author = {O. Friderikos and A. Mendoza and Emmanuel Baranger and D. Sagris and C. David},
url = {https://congressarchive.cimne.com/dte_aicomas_2025/abstracts/b8d1d10a96b711efba01000c29ddfc0c.pdf},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
booktitle = {Digital Twins in Engineering & Artificial Intelligence and Computational Methods in Applied Science, DTE - AICOMAS 2025},
keywords = {},
pubstate = {published},
tppubtype = {collection}
}
(Ed.)
HPE-DARTS: Hybrid Pruning and Proxy Evaluation in Differentiable Architecture Search Collection
2025.
@collection{lin-,
title = {HPE-DARTS: Hybrid Pruning and Proxy Evaluation in Differentiable Architecture Search},
author = {Hung-I Lin and Lin-Jing Kuo and Sheng-De Wang},
url = {https://www.scitepress.org/Papers/2025/131487/131487.pdf},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
booktitle = {Proceedings of the 17th International Conference on Agents and Artificial Intelligence (ICAART 2025)},
journal = {Proceedings of the 17th International Conference on Agents and Artificial Intelligence (ICAART 2025)},
keywords = {},
pubstate = {published},
tppubtype = {collection}
}
Li, Yuangang; Ma, Rui; Zhang, Qian; Wang, Zeyu; Zong, Linlin; Liu, Xinyue
Neural architecture search using attention enhanced precise path evaluation and efficient forward evolution Journal Article
In: scientific reports , 2025.
@article{nokey,
title = {Neural architecture search using attention enhanced precise path evaluation and efficient forward evolution},
author = {Yuangang Li and Rui Ma and Qian Zhang and Zeyu Wang and Linlin Zong and Xinyue Liu
},
url = {https://www.nature.com/articles/s41598-025-94187-8},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
booktitle = {scientific reports },
journal = {scientific reports },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Xu, Dikai; Cao, Bin
Adaptive Multiobjective Evolutionary Generative Adversarial Network for Metaverse Network Intrusion Detection Journal Article
In: Science Partner Journals, 2025.
@article{nokey,
title = {Adaptive Multiobjective Evolutionary Generative Adversarial Network for Metaverse Network Intrusion Detection},
author = {Dikai Xu and Bin Cao},
url = {https://spj.science.org/doi/pdf/10.34133/research.0665},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
journal = {Science Partner Journals},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Naayini, Prudhvi; Kamatala, Srikanth; Myakala, Praveen Kumar
Transforming Performance Engineering with Generative AI Journal Article
In: Journal of Computer and Communications , vol. 13, no. 3, 2025.
@article{Naayini-jcc25a,
title = { Transforming Performance Engineering with Generative AI },
author = { Prudhvi Naayini and Srikanth Kamatala and Praveen Kumar Myakala},
url = {https://www.scirp.org/journal/paperinformation?paperid=141454},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
journal = {Journal of Computer and Communications },
volume = {13},
number = {3},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
HERNANDEZ, ESAU ALAIN HERVERT; CAO, YAN; KEHTARNAVAZ, NASSER
Computationally Efficient Neural Architecture Search for Image Denoising Bachelor Thesis
2025.
@bachelorthesis{nokey,
title = {Computationally Efficient Neural Architecture Search for Image Denoising},
author = {ESAU ALAIN HERVERT HERNANDEZ and YAN CAO and NASSER KEHTARNAVAZ},
url = {https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=10948435},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
journal = {IEEE Access},
keywords = {},
pubstate = {published},
tppubtype = {bachelorthesis}
}
Liu, Guoqing; Qian, Yuhua; Liang, Xinyan; Fu, Pinhan
Core structure-guided multi-modal classification via Monte Carlo Tree Search Journal Article
In: International Journal of Machine Learning and Cybernetics , 2025.
@article{liu-ijmlc25a,
title = {Core structure-guided multi-modal classification via Monte Carlo Tree Search},
author = { Guoqing Liu and Yuhua Qian and Xinyan Liang and Pinhan Fu
},
url = {https://link.springer.com/article/10.1007/s13042-025-02606-z},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
journal = {International Journal of Machine Learning and Cybernetics },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Ju, Moran; Niu, Buniu; Li, Mulin; Mao, Tengkai; Jin, Si-nian
Toward Better Accuracy-Efficiency Trade-Offs for Oriented SAR Ship Object Detection Bachelor Thesis
2025.
@bachelorthesis{ju-jstaeors,
title = {Toward Better Accuracy-Efficiency Trade-Offs for Oriented SAR Ship Object Detection},
author = {Moran Ju and Buniu Niu and Mulin Li and Tengkai Mao and Si-nian Jin},
url = {https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=10944503},
year = {2025},
date = {2025-03-01},
urldate = {2025-03-01},
journal = { IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing. },
keywords = {},
pubstate = {published},
tppubtype = {bachelorthesis}
}
Herterich, Nils; Liu, Kai; Stein, Anthony
Multi-objective neural architecture search for real-time weed detection on embedded system Miscellaneous
2025.
@misc{Herterich,
title = {Multi-objective neural architecture search for real-time weed detection on embedded system},
author = {Nils Herterich and Kai Liu and Anthony Stein},
url = {https://dl.gi.de/server/api/core/bitstreams/29a49f8d-304e-4073-8a92-4bef6483c087/content},
year = {2025},
date = {2025-02-01},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Huang, Tao
Efficient Deep Neural Architecture Design and Training PhD Thesis
2025.
@phdthesis{nokey,
title = {Efficient Deep Neural Architecture Design and Training},
author = { Huang, Tao },
url = {https://ses.library.usyd.edu.au/handle/2123/33598},
year = {2025},
date = {2025-02-01},
urldate = {2025-02-01},
keywords = {},
pubstate = {published},
tppubtype = {phdthesis}
}
Zhao, Yi-Heng; Pang, Shen-Wen; Huang, Heng-Zhi; Wu, Shao-Wen; Sun, Shao-Hua; Liu, Zhen-Bing; Pan, Zhi-Chao
Automatic clustering of single-molecule break junction data through task-oriented representation learning Journal Article
In: Rare Metals , 2025.
@article{zhao-rarem25a,
title = {Automatic clustering of single-molecule break junction data through task-oriented representation learning},
author = {
Yi-Heng Zhao and Shen-Wen Pang and Heng-Zhi Huang and Shao-Wen Wu and Shao-Hua Sun and Zhen-Bing Liu and Zhi-Chao Pan
},
url = {https://link.springer.com/article/10.1007/s12598-024-03089-7},
year = {2025},
date = {2025-02-01},
urldate = {2025-02-01},
journal = { Rare Metals },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Fu, Jintao; Cong, Peng; Xu, Shuo; Chang, Jiahao; Liu, Ximing; Sun, Yuewen
Neural architecture search with Deep Radon Prior for sparse-view CT image reconstruction Journal Article
In: Med Phys , 2025.
@article{Fu-medphs25a,
title = { Neural architecture search with Deep Radon Prior for sparse-view CT image reconstruction },
author = {Jintao Fu and Peng Cong and Shuo Xu and Jiahao Chang and Ximing Liu and Yuewen Sun
},
url = {https://pubmed.ncbi.nlm.nih.gov/39930320/},
year = {2025},
date = {2025-02-01},
urldate = {2025-02-01},
journal = { Med Phys },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Yu, Sixing
2025.
@phdthesis{yu-phd25a,
title = {Scalable and resource-efcient federated learning: Techniques for resource-constrained heterogeneous systems},
author = {Sixing Yu},
url = {https://www.proquest.com/docview/3165602177?pq-origsite=gscholar&fromopenview=true&sourcetype=Dissertations%20&%20Theses},
year = {2025},
date = {2025-02-01},
urldate = {2025-02-01},
keywords = {},
pubstate = {published},
tppubtype = {phdthesis}
}
(Ed.)
TinyDevID: TinyML-Driven IoT Devices IDentification Using Network Flow Data Collection
2025.
@collection{Rushikesh-csp25a,
title = {TinyDevID: TinyML-Driven IoT Devices IDentification Using Network Flow Data},
author = {Priyanka Rushikesh Chaudhary and Anand Agrawal and Rajib Ranjan Maiti},
url = {https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=10885715},
year = {2025},
date = {2025-02-01},
urldate = {2025-02-01},
booktitle = {COMSNETS 2025 - Cybersecurity & Privacy Workshop (CSP)},
keywords = {},
pubstate = {published},
tppubtype = {collection}
}
Tabak, Gabriel Couto; Molenaar, Dylan; Curi, Mariana
An evolutionary neural architecture search for item response theory autoencoders Journal Article
In: Behaviormetrika , 2025.
@article{nokey,
title = {An evolutionary neural architecture search for item response theory autoencoders},
author = {Gabriel Couto Tabak and Dylan Molenaar and Mariana Curi
},
url = {https://link.springer.com/article/10.1007/s41237-024-00250-5},
year = {2025},
date = {2025-01-27},
urldate = {2025-01-27},
journal = {Behaviormetrika },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Hao, Debei; Pei, Songwei
MIG-DARTS: towards effective differentiable architecture search by gradually mitigating the initial-channel gap between search and evaluation Journal Article
In: Neural Computing and Applications, 2025.
@article{nokey,
title = {MIG-DARTS: towards effective differentiable architecture search by gradually mitigating the initial-channel gap between search and evaluation},
author = {
Debei Hao and Songwei Pei
},
url = {https://link.springer.com/article/10.1007/s00521-024-10681-6},
year = {2025},
date = {2025-01-09},
urldate = {2025-01-09},
journal = {Neural Computing and Applications},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
(Ed.)
2025.
@collection{nokey,
title = {H4H: Hybrid Convolution-Transformer Architecture Search for NPU-CIM Heterogeneous Systems for AR/VR Applications},
author = {Yiwei Zhao and Jinhui Chen and Sai Qian Zhang and Syed Shakib Sarwar and Kleber Hugo Stangherlin and Jorge Tomas Gomez and Jae-Sun Seo and Barbara De Salvo and Chiao Liu and Phillip B. Gibbons and Ziyun Li},
url = {https://www.pdl.cmu.edu/PDL-FTP/associated/ASP-DAC2025-1073-12.pdf},
year = {2025},
date = {2025-01-02},
urldate = {2025-01-02},
booktitle = {ASPDAC ’25},
keywords = {},
pubstate = {published},
tppubtype = {collection}
}
Öcal, Göktuğ; Özgövde, Atay
Network-aware federated neural architecture search Journal Article
In: Future Generation Computer Systems, vol. 162, pp. 107475, 2025, ISSN: 0167-739X.
@article{OCAL2025107475,
title = {Network-aware federated neural architecture search},
author = {Göktuğ Öcal and Atay Özgövde},
url = {https://www.sciencedirect.com/science/article/pii/S0167739X24004205},
doi = {https://doi.org/10.1016/j.future.2024.07.053},
issn = {0167-739X},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Future Generation Computer Systems},
volume = {162},
pages = {107475},
abstract = {The cooperation between Deep Learning (DL) and edge devices has further advanced technological developments, allowing smart devices to serve as both data sources and endpoints for DL-powered applications. However, the success of DL relies on optimal Deep Neural Network (DNN) architectures, and manually developing such systems requires extensive expertise and time. Neural Architecture Search (NAS) has emerged to automate the search for the best-performing neural architectures. Meanwhile, Federated Learning (FL) addresses data privacy concerns by enabling collaborative model development without exchanging the private data of clients. In a FL system, network limitations can lead to biased model training, slower convergence, and increased communication overhead. On the other hand, traditional DNN architecture design, emphasizing validation accuracy, often overlooks computational efficiency and size constraints of edge devices. This research aims to develop a comprehensive framework that effectively balances trade-offs between model performance, communication efficiency, and the incorporation of FL into an iterative NAS algorithm. This framework aims to overcome challenges by addressing the specific requirements of FL, optimizing DNNs through NAS, and ensuring computational efficiency while considering the network constraints of edge devices. To address these challenges, we introduce Network-Aware Federated Neural Architecture Search (NAFNAS), an open-source federated neural network pruning framework with network emulation support. Through comprehensive testing, we demonstrate the feasibility of our approach, efficiently reducing DNN size and mitigating communication challenges. Additionally, we propose Network and Distribution Aware Client Grouping (NetDAG), a novel client grouping algorithm tailored for FL with diverse DNN architectures, considerably enhancing efficiency of communication rounds and update balance.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Rahman, Abdur; Street, Jason; Wooten, James; Marufuzzaman, Mohammad; Gude, Veera G.; Buchanan, Randy; Wang, Haifeng
MoistNet: Machine vision-based deep learning models for wood chip moisture content measurement Journal Article
In: Expert Systems with Applications, vol. 259, pp. 125363, 2025, ISSN: 0957-4174.
@article{Rahman_2025,
title = {MoistNet: Machine vision-based deep learning models for wood chip moisture content measurement},
author = {Abdur Rahman and Jason Street and James Wooten and Mohammad Marufuzzaman and Veera G. Gude and Randy Buchanan and Haifeng Wang},
url = {http://dx.doi.org/10.1016/j.eswa.2024.125363},
doi = {10.1016/j.eswa.2024.125363},
issn = {0957-4174},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Expert Systems with Applications},
volume = {259},
pages = {125363},
publisher = {Elsevier BV},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Jiang, Zhiying; Liu, Risheng; Yang, Shuzhou; Zhang, Zengxi; Fan, Xin
DRNet: Learning a dynamic recursion network for chaotic rain streak removal Journal Article
In: Pattern Recognition, vol. 158, pp. 111004, 2025, ISSN: 0031-3203.
@article{JIANG2025111004,
title = {DRNet: Learning a dynamic recursion network for chaotic rain streak removal},
author = {Zhiying Jiang and Risheng Liu and Shuzhou Yang and Zengxi Zhang and Xin Fan},
url = {https://www.sciencedirect.com/science/article/pii/S0031320324007556},
doi = {https://doi.org/10.1016/j.patcog.2024.111004},
issn = {0031-3203},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Pattern Recognition},
volume = {158},
pages = {111004},
abstract = {Image deraining refers to removing the visible rain streaks to restore the rain-free scenes. Existing methods rely on manually crafted networks to model the distribution of rain streaks. However, complex scenes disrupt the uniformity of rain streak characteristics assumed in ideal conditions, resulting in rain streaks of varying directions, intensities, and brightness intersecting within the same scene, challenging the deep learning based deraining performance. To address the chaotic rain streak removal, we handle the rain streaks with similar distribution characteristics in the same layer and employ a dynamic recursive mechanism to extract and unveil them progressively. Specifically, we employ neural architecture search to determine the models of different rain streaks. To avoid the loss of texture details associated with overly deep structures, we integrate multi-scale modeling and cross-scale recruitment within the dynamic structure. Considering the application of real-world scenes, we incorporate contrastive training to improve the generalization. Experimental results indicate superior performance in rain streak depiction compared to existing methods. Practical evaluation confirms its effectiveness in object detection and semantic segmentation tasks. Code is available at https://github.com/Jzy2017/DRNet.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Wu, Zhenpeng; Chen, Jiamin; Al-Sabri, Raeed; Oloulade, Babatounde Moctard; Gao, Jianliang
Asymmetric augmented paradigm-based graph neural architecture search Journal Article
In: Information Processing & Management, vol. 62, no. 1, pp. 103897, 2025, ISSN: 0306-4573.
@article{WU2025103897,
title = {Asymmetric augmented paradigm-based graph neural architecture search},
author = {Zhenpeng Wu and Jiamin Chen and Raeed Al-Sabri and Babatounde Moctard Oloulade and Jianliang Gao},
url = {https://www.sciencedirect.com/science/article/pii/S0306457324002565},
doi = {https://doi.org/10.1016/j.ipm.2024.103897},
issn = {0306-4573},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Information Processing & Management},
volume = {62},
number = {1},
pages = {103897},
abstract = {In most scenarios of graph-based tasks, graph neural networks (GNNs) are trained end-to-end with labeled samples. Labeling graph samples, a time-consuming and expert-dependent process, leads to huge costs. Graph data augmentations can provide a promising method to expand labeled samples cheaply. However, graph data augmentations will damage the capacity of GNNs to distinguish non-isomorphic graphs during the supervised graph representation learning process. How to utilize graph data augmentations to expand labeled samples while preserving the capacity of GNNs to distinguish non-isomorphic graphs is a challenging research problem. To address the above problem, we abstract a novel asymmetric augmented paradigm in this paper and theoretically prove that it offers a principled approach. The asymmetric augmented paradigm can preserve the capacity of GNNs to distinguish non-isomorphic graphs while utilizing augmented labeled samples to improve the generalization capacity of GNNs. To be specific, the asymmetric augmented paradigm will utilize similar yet distinct asymmetric weights to classify the real sample and augmented sample, respectively. To systemically explore the benefits of asymmetric augmented paradigm under different GNN architectures, rather than studying individual asymmetric augmented GNN (A2GNN) instance, we then develop an auto-search engine called Asymmetric Augmented Graph Neural Architecture Search (A2GNAS) to save human efforts. We empirically validate our asymmetric augmented paradigm on multiple graph classification benchmarks, and demonstrate that representative A2GNN instances automatically discovered by our A2GNAS method achieve state-of-the-art performance compared with competitive baselines. Our codes are available at: https://github.com/csubigdata-Organization/A2GNAS.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Wang, Weibo; Li, Hua
NAS FD Lung: A novel lung assist diagnostic system based on neural architecture search Journal Article
In: Biomedical Signal Processing and Control, vol. 100, pp. 107022, 2025, ISSN: 1746-8094.
@article{WANG2025107022,
title = {NAS FD Lung: A novel lung assist diagnostic system based on neural architecture search},
author = {Weibo Wang and Hua Li},
url = {https://www.sciencedirect.com/science/article/pii/S1746809424010802},
doi = {https://doi.org/10.1016/j.bspc.2024.107022},
issn = {1746-8094},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Biomedical Signal Processing and Control},
volume = {100},
pages = {107022},
abstract = {In the detection and recognition of lung nodules, pulmonary nodules vary in size and shape and contain many similar tissues and organs around them, leading to the problems of both missed detection and false detection in existing detection algorithms. Designing proprietary detection and recognition networks manually requires substantial professional expertise. This process is time-consuming and labour-intensive and leads to issues like parameter redundancy and improper feature selection. Therefore, this paper proposes a new pulmonary CAD (computer-aided diagnosis) system for pulmonary nodules, NAS FD Lung (Using the NAS approach to search deep FPN and DPN networks), that can automatically learn and generate a deep learning network tailored to pulmonary nodule detection and recognition task requirements. NAS FD Lung aims to use automatic search to generate deep learning networks in the auxiliary diagnosis of pulmonary nodules to replace the manual design of deep learning networks. NAS FD Lung comprises two automatic search networks: BM NAS-FPN (Using NAS methods to search for deep FPN structures with Binary operation and Matrix multiplication fusion methods) network for nodule detection and NAS-A-DPN (Using the NAS approach to search deep DPN networks with attention mechanism) for nodule identification. The proposed technique is tested on the LUNA16 dataset, and the experimental results show that the model is superior to many existing state-of-the-art approaches. The detection accuracy of lung nodules is 98.23%. Regarding the lung nodules classification, the accuracy, specificity, sensitivity, and AUC values achieved 96.32%,97.14%,95.82%, and 98.33%, respectively.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Solis-Martin, David; Galan-Paez, Juan; Borrego-Diaz, Joaquin
Bayesian Model Selection Pruning in Predictive Maintenance Proceedings Article
In: Quintián, Héctor; Corchado, Emilio; Lora, Alicia Troncoso; García, Hilde Pérez; Pérez, Esteban Jove; Rolle, José Luis Calvo; de Pisón, Francisco Javier Martínez; Bringas, Pablo García; Álvarez, Francisco Martínez; Herrero, Álvaro; Fosci, Paolo (Ed.): Hybrid Artificial Intelligent Systems, pp. 263–274, Springer Nature Switzerland, Cham, 2025, ISBN: 978-3-031-74183-8.
@inproceedings{10.1007/978-3-031-74183-8_22,
title = {Bayesian Model Selection Pruning in Predictive Maintenance},
author = {David Solis-Martin and Juan Galan-Paez and Joaquin Borrego-Diaz},
editor = {Héctor Quintián and Emilio Corchado and Alicia Troncoso Lora and Hilde Pérez García and Esteban Jove Pérez and José Luis Calvo Rolle and Francisco Javier Martínez de Pisón and Pablo García Bringas and Francisco Martínez Álvarez and Álvaro Herrero and Paolo Fosci},
url = {https://link.springer.com/chapter/10.1007/978-3-031-74183-8_22},
isbn = {978-3-031-74183-8},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {Hybrid Artificial Intelligent Systems},
pages = {263–274},
publisher = {Springer Nature Switzerland},
address = {Cham},
abstract = {Deep Neural Network architecture design significantly impacts the final model performance. The process of searching for optimal architectures, known as Neural Architecture Search (NAS), involves training and evaluating an important number of models. Therefore, mechanisms to reduce the resources required for NAS are highly valuable.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Solis-Martin, David; Galan-Paez, Juan; Borrego-Diaz, Joaquin
Bayesian Model Selection Pruning in Predictive Maintenance Proceedings Article
In: Quintián, Héctor; Corchado, Emilio; Lora, Alicia Troncoso; García, Hilde Pérez; Pérez, Esteban Jove; Rolle, José Luis Calvo; de Pisón, Francisco Javier Martínez; Bringas, Pablo García; Álvarez, Francisco Martínez; Herrero, Álvaro; Fosci, Paolo (Ed.): Hybrid Artificial Intelligent Systems, pp. 263–274, Springer Nature Switzerland, Cham, 2025, ISBN: 978-3-031-74183-8.
@inproceedings{10.1007/978-3-031-74183-8_22b,
title = {Bayesian Model Selection Pruning in Predictive Maintenance},
author = {David Solis-Martin and Juan Galan-Paez and Joaquin Borrego-Diaz},
editor = {Héctor Quintián and Emilio Corchado and Alicia Troncoso Lora and Hilde Pérez García and Esteban Jove Pérez and José Luis Calvo Rolle and Francisco Javier Martínez de Pisón and Pablo García Bringas and Francisco Martínez Álvarez and Álvaro Herrero and Paolo Fosci},
url = {https://link.springer.com/chapter/10.1007/978-3-031-74183-8_22},
isbn = {978-3-031-74183-8},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {Hybrid Artificial Intelligent Systems},
pages = {263–274},
publisher = {Springer Nature Switzerland},
address = {Cham},
abstract = {Deep Neural Network architecture design significantly impacts the final model performance. The process of searching for optimal architectures, known as Neural Architecture Search (NAS), involves training and evaluating an important number of models. Therefore, mechanisms to reduce the resources required for NAS are highly valuable.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Garcia-Garcia, Cosijopii; Derbel, Bilel; Morales-Reyes, Alicia; Escalante, Hugo Jair
Speeding up the Multi-objective NAS Through Incremental Learning Proceedings Article
In: Martínez-Villaseñor, Lourdes; Ochoa-Ruiz, Gilberto (Ed.): Advances in Soft Computing, pp. 3–15, Springer Nature Switzerland, Cham, 2025, ISBN: 978-3-031-75543-9.
@inproceedings{10.1007/978-3-031-75543-9_1,
title = {Speeding up the Multi-objective NAS Through Incremental Learning},
author = {Cosijopii Garcia-Garcia and Bilel Derbel and Alicia Morales-Reyes and Hugo Jair Escalante},
editor = {Lourdes Martínez-Villaseñor and Gilberto Ochoa-Ruiz},
url = {https://link.springer.com/chapter/10.1007/978-3-031-75543-9_1},
isbn = {978-3-031-75543-9},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {Advances in Soft Computing},
pages = {3–15},
publisher = {Springer Nature Switzerland},
address = {Cham},
abstract = {Deep neural networks (DNNs), particularly convolutional neural networks (CNNs), have garnered significant attention in recent years for addressing a wide range of challenges in image processing and computer vision. Neural architecture search (NAS) has emerged as a crucial field aiming to automate the design and configuration of CNN models. In this paper, we propose a novel strategy to speed up the performance estimation of neural architectures by gradually increasing the size of the training set used for evaluation as the search progresses. We evaluate this approach using the CGP-NASV2 model, a multi-objective NAS method, on the CIFAR-100 dataset. Experimental results demonstrate a notable acceleration in the search process, achieving a speedup of 4.6 times compared to the baseline. Despite using limited data in the early stages, our proposed method effectively guides the search towards competitive architectures. This study highlights the efficacy of leveraging lower-fidelity estimates in NAS and paves the way for further research into accelerating the design of efficient CNN architectures.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Liu, Wenbo; Deng, Tao; An, Rui; Yan, Fei
DARTS-CGW: Research on Differentiable Neural Architecture Search Algorithm Based on Coarse Gradient Weighting Proceedings Article
In: Lin, Zhouchen; Cheng, Ming-Ming; He, Ran; Ubul, Kurban; Silamu, Wushouer; Zha, Hongbin; Zhou, Jie; Liu, Cheng-Lin (Ed.): Pattern Recognition and Computer Vision, pp. 31–44, Springer Nature Singapore, Singapore, 2025, ISBN: 978-981-97-8502-5.
@inproceedings{10.1007/978-981-97-8502-5_3,
title = {DARTS-CGW: Research on Differentiable Neural Architecture Search Algorithm Based on Coarse Gradient Weighting},
author = {Wenbo Liu and Tao Deng and Rui An and Fei Yan},
editor = {Zhouchen Lin and Ming-Ming Cheng and Ran He and Kurban Ubul and Wushouer Silamu and Hongbin Zha and Jie Zhou and Cheng-Lin Liu},
url = {https://link.springer.com/chapter/10.1007/978-981-97-8502-5_3},
isbn = {978-981-97-8502-5},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {Pattern Recognition and Computer Vision},
pages = {31–44},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Differential architecture search (DARTS) has emerged as a prominent research area, yet it grapples with a longstanding challenge: the discretization discrepancy problem. This issue directly impedes the search for an optimal model architecture and undermines search algorithm performance. To alleviate this issue, we propose a novel coarse gradient weighting algorithm. Our proposed algorithm has the capability to simulate the discretization process, wherein the architectural parameters move toward both ends. And we integrate this discretization process into the training phase of the architectural parameters, enabling the model to adapt to the discretization process in a trial-and-error fashion. Specifically, based on the architectural parameters in training, we divide the candidate operations into two regions, i.e., the easy-to-select region and the hard-to-be-selected region. The different weighting strategies are implemented in different regions so that the architectural parameters are pushed to the ends. The processed architecture parameters are used for training, which is equivalent to introducing the discretization process into the search phase. Additionally, we use the coarse gradient algorithm to optimize the updating process of the weighting algorithm and theoretically justify the rationality of the coarse gradient weighting algorithm. Extensive experimental results demonstrate that our proposed method can improve the performance of the searched model and make DARTS more robust without adding additional search time.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Man, Wenxing; Xu, Liming; He, Chunlin
Evolutionary architecture search for generative adversarial networks using an aging mechanism-based strategy Journal Article
In: Neural Networks, vol. 181, pp. 106877, 2025, ISSN: 0893-6080.
@article{MAN2025106877,
title = {Evolutionary architecture search for generative adversarial networks using an aging mechanism-based strategy},
author = {Wenxing Man and Liming Xu and Chunlin He},
url = {https://www.sciencedirect.com/science/article/pii/S0893608024008050},
doi = {https://doi.org/10.1016/j.neunet.2024.106877},
issn = {0893-6080},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Neural Networks},
volume = {181},
pages = {106877},
abstract = {Generative Adversarial Networks (GANs) have emerged as a key technology in artificial intelligence, especially in image generation. However, traditionally hand-designed GAN architectures often face significant training stability challenges, which are effectively addressed by our Evolutionary Neural Architecture Search (ENAS) algorithm for GANs, named EAMGAN. This one-shot model automates the design of GAN architectures and employs an Operation Importance Metric (OIM) to enhance training stability. It also incorporates an aging mechanism to optimize the selection process during architecture search. Additionally, the use of a non-dominated sorting algorithm ensures the generation of Pareto-optimal solutions, promoting diversity and preventing premature convergence. We evaluated our method on benchmark datasets, and the results demonstrate that EAMGAN is highly competitive in terms of efficiency and performance. Our method identified an architecture achieving Inception Scores (IS) of 8.83±0.13 and Fréchet Inception Distance (FID) of 9.55 on CIFAR-10 with only 0.66 GPU days. Results on the STL-10, CIFAR-100, and ImageNet32 datasets further demonstrate the robust portability of our architecture.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Zhong, Rui; Xu, Yuefeng; Zhang, Chao; Yu, Jun
Efficient multiplayer battle game optimizer for numerical optimization and adversarial robust neural architecture search Journal Article
In: Alexandria Engineering Journal, vol. 113, pp. 150-168, 2025, ISSN: 1110-0168.
@article{ZHONG2025150,
title = {Efficient multiplayer battle game optimizer for numerical optimization and adversarial robust neural architecture search},
author = {Rui Zhong and Yuefeng Xu and Chao Zhang and Jun Yu},
url = {https://www.sciencedirect.com/science/article/pii/S1110016824014935},
doi = {https://doi.org/10.1016/j.aej.2024.11.035},
issn = {1110-0168},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Alexandria Engineering Journal},
volume = {113},
pages = {150-168},
abstract = {This paper introduces a novel metaheuristic algorithm, known as the efficient multiplayer battle game optimizer (EMBGO), specifically designed for addressing complex numerical optimization tasks. The motivation behind this research stems from the need to rectify identified shortcomings in the original MBGO, particularly in search operators during the movement phase, as revealed through ablation experiments. EMBGO mitigates these limitations by integrating the movement and battle phases to simplify the original optimization framework and improve search efficiency. Besides, two efficient search operators: differential mutation and Lévy flight are introduced to increase the diversity of the population. To evaluate the performance of EMBGO comprehensively and fairly, numerical experiments are conducted on benchmark functions such as CEC2017, CEC2020, and CEC2022, as well as engineering problems. Twelve well-established MA approaches serve as competitor algorithms for comparison. Furthermore, we apply the proposed EMBGO to the complex adversarial robust neural architecture search (ARNAS) tasks and explore its robustness and scalability. The experimental results and statistical analyses confirm the efficiency and effectiveness of EMBGO across various optimization tasks. As a potential optimization technique, EMBGO holds promise for diverse applications in real-world problems and deep learning scenarios. The source code of EMBGO is made available in https://github.com/RuiZhong961230/EMBGO.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Wang, Yu; Fan, Jiahao; Sun, Yanan
Classification of sewer pipe defects based on an automatically designed convolutional neural network Journal Article
In: Expert Systems with Applications, vol. 264, pp. 125806, 2025, ISSN: 0957-4174.
@article{WANG2025125806,
title = {Classification of sewer pipe defects based on an automatically designed convolutional neural network},
author = {Yu Wang and Jiahao Fan and Yanan Sun},
url = {https://www.sciencedirect.com/science/article/pii/S0957417424026733},
doi = {https://doi.org/10.1016/j.eswa.2024.125806},
issn = {0957-4174},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Expert Systems with Applications},
volume = {264},
pages = {125806},
abstract = {Accurate classification of sewer pipe defects allows for timely prevention of system failures, preventing environmental pollution caused by sewage leakage, thereby safeguarding public health and ensuring urban infrastructure operates normally. However, existing methods with deep convolutional neural networks not only require a significant amount of time to design network architectures, but also demand substantial computational resources. Therefore, this paper automatically designed high-performance and lightweight AutoSewerNet based on neural architecture search (NAS) for classification of sewer pipe defects. First, Super-net is designed to enrich the diversity of search space and include high-performance network architecture. Second, to reduce the search time of NAS, a gradient-based search strategy is designed. Third, AutoSewerNet is designed to be lightweight and perform real-time inspections of sewer pipes. Fourth, weight balance is introduced to solve the imbalanced dataset problem. Experimental results demonstrate that AutoSewerNet achieved an F1-score of 0.6251 on the benchmark dataset, which is far superior to that of ResNet-50 (F1-score: 0.4523) and InceptionV3 (F1-score: 0.4611). Moreover, AutoSewerNet requires only 11.6% of VGG-16. Thus, AutoSewerNet is better than state-of-the-art methods. Our source code and models are anonymously available at https://anonymous.4open.science/r/ASN-26AB/.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Fu, Wei; Lou, Wenqi; Qin, Yunji; Gong, Lei; Wang, Chao; Zhou, Xuehai
MFNAS: Multi-fidelity Exploration in Neural Architecture Search with Stable Zero-Shot Proxy Proceedings Article
In: Hadfi, Rafik; Anthony, Patricia; Sharma, Alok; Ito, Takayuki; Bai, Quan (Ed.): PRICAI 2024: Trends in Artificial Intelligence, pp. 348–360, Springer Nature Singapore, Singapore, 2025, ISBN: 978-981-96-0116-5.
@inproceedings{10.1007/978-981-96-0116-5_29,
title = {MFNAS: Multi-fidelity Exploration in Neural Architecture Search with Stable Zero-Shot Proxy},
author = {Wei Fu and Wenqi Lou and Yunji Qin and Lei Gong and Chao Wang and Xuehai Zhou},
editor = {Rafik Hadfi and Patricia Anthony and Alok Sharma and Takayuki Ito and Quan Bai},
isbn = {978-981-96-0116-5},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {PRICAI 2024: Trends in Artificial Intelligence},
pages = {348–360},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Neural architecture search (NAS) automates the design of neural networks for specific tasks. Recently, zero-shot NAS has attracted much attention. Unlike traditional NAS, which relies on training to rank architectures, zero-shot NAS uses gradients or activation information to evaluate architecture performance. However, existing zero-shot NAS methods are limited by their inconsistent architecture ranking and the evaluation bias of their search algorithm, making it challenging to discover networks with high accuracy efficiently. To address this dilemma, this paper proposes an efficient and stable search framework for zero-shot NAS. Firstly, we design a stable zero-shot proxy, which achieves good consistency with network accuracy by utilizing filtered gradient information. On this basis, we employ a multi-fidelity evolutionary algorithm for efficient exploration. This algorithm utilizes multi-fidelity proxies to correct the bias towards certain types of networks and enhances the ability to distinguish high-performing architectures, achieving rapid convergence through performance-directed multi-point crossover and mutation. Experimental results conducted on NATS-Bench demonstrate that our framework can discover high-performance architectures within minutes of GPU time, outperforming existing training-free and training-based NAS methods. The code is available at https://github.com/mine7777/MFNAS.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Verma, Sahil; Kumar, Prabhat; Singh, Jyoti Prakash
MLP-GNAS: Meta-learning-based predictor-assisted Genetic Neural Architecture Search system Journal Article
In: Applied Soft Computing, vol. 169, pp. 112527, 2025, ISSN: 1568-4946.
@article{VERMA2025112527,
title = {MLP-GNAS: Meta-learning-based predictor-assisted Genetic Neural Architecture Search system},
author = {Sahil Verma and Prabhat Kumar and Jyoti Prakash Singh},
url = {https://www.sciencedirect.com/science/article/pii/S1568494624013012},
doi = {https://doi.org/10.1016/j.asoc.2024.112527},
issn = {1568-4946},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Applied Soft Computing},
volume = {169},
pages = {112527},
abstract = {Convolutional neural networks achieve state-of-the-art results on many image classification tasks. However, according to the No Free Lunch Theorem, no single model performs optimally for all the datasets. Hence, the manual development of these models using the hit-and-trial approach requires high computational overhead. Transfer learning mitigates this issue to some extent by transferring knowledge learned from similar domain tasks to target tasks through model architecture and weights. Yet, the model remains designed for the source task, and using the same model on the target task may result in over-parameterization or sub-optimal results. Henceforth, the current work proposes a meta-learning-based predictor-assisted genetic Neural Architecture Search (MLP-GNAS) system to automate the model generation process for plant disease detection tasks to validate its efficacy in real-world scenarios. To this cause, the MLP-GNAS system employs a meta-learning component to recommend the top 3 suitable models for a target dataset, followed by a genetic algorithm-based NAS to fine-tune the recommended model. In addition, the proposed approach uses a CNN-based performance predictor designed to discard models unlikely to surpass the optimal performance recorded thus far. Further, extensive comparative experiments demonstrate that MLP-GNAS-generated models outperform 14 state-of-the-art models on two distinct plant disease datasets with an accuracy of 96.94% and 92.04%. Moreover, further comparison with the manually fine-tuned and NAS-generated models reveals that the proposed system outperforms both to obtain an accuracy of 98.35% and 99.95% on respective datasets.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Oloulade, Babatounde Moctard; Gao, Jianliang; Chen, Jiamin; Al-Sabri, Raeed; Wu, Zhenpeng; Abdullah, Monir
Shapley-guided pruning for efficient graph neural architecture prediction in distributed learning environments Journal Article
In: Information Sciences, vol. 695, pp. 121695, 2025, ISSN: 0020-0255.
@article{OLOULADE2025121695,
title = {Shapley-guided pruning for efficient graph neural architecture prediction in distributed learning environments},
author = {Babatounde Moctard Oloulade and Jianliang Gao and Jiamin Chen and Raeed Al-Sabri and Zhenpeng Wu and Monir Abdullah},
url = {https://www.sciencedirect.com/science/article/pii/S0020025524016098},
doi = {https://doi.org/10.1016/j.ins.2024.121695},
issn = {0020-0255},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Information Sciences},
volume = {695},
pages = {121695},
abstract = {Graph neural architecture search (GNAS) methods have made significant strides in predictive modeling; however, challenges persist in optimizing these approaches for improved accuracy and scalability. In this paper, we propose GraphNAP++, an innovative framework that combines Shapley-value-guided search space pruning with data-centric distributed learning (DCDL) to address the scalability and efficiency limitations of existing GNAS methodologies. The proposed method begins by selecting a subset of architectures from the search space, which are evaluated on a graph validation dataset using DCDL. These architectures are encoded, and a neural predictor is trained to predict their performance. Shapley values are utilized to prune the search space by retaining only the most influential architectures. The neural predictor then estimates the performance of all architectures within the reduced search space, and the highest-performing architectures are selected for final evaluation. Experimental results on benchmark datasets in distributed learning environments demonstrate the effectiveness of GraphNAP++ in both graph and node classification tasks, highlighting its potential to advance the state-of-the-art for graph neural networks.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Chen, Jiakun; Xu, Jie; Hu, Jiahui; Qiao, Liqiang; Wang, Shuo; Huang, Feiran; Li, Chaozhuo
Context-Aware Structural Adaptive Graph Neural Networks Proceedings Article
In: Hadfi, Rafik; Anthony, Patricia; Sharma, Alok; Ito, Takayuki; Bai, Quan (Ed.): PRICAI 2024: Trends in Artificial Intelligence, pp. 467–479, Springer Nature Singapore, Singapore, 2025, ISBN: 978-981-96-0116-5.
@inproceedings{10.1007/978-981-96-0116-5_39,
title = {Context-Aware Structural Adaptive Graph Neural Networks},
author = {Jiakun Chen and Jie Xu and Jiahui Hu and Liqiang Qiao and Shuo Wang and Feiran Huang and Chaozhuo Li},
editor = {Rafik Hadfi and Patricia Anthony and Alok Sharma and Takayuki Ito and Quan Bai},
url = {https://link.springer.com/chapter/10.1007/978-981-96-0116-5_39},
isbn = {978-981-96-0116-5},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {PRICAI 2024: Trends in Artificial Intelligence},
pages = {467–479},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Graph-based data structures are prevalent in various real-world applications, for example, protein molecules and social connection networks, necessitating effective representation learning techniques. Graph Neural Networks (GNNs) have demonstrated significant advancements in tasks like node classification and social network analysis through recursive information aggregation. However, current GNN approaches are predominantly static, lacking adaptability to specific graph structures. Inspired by Neural Architecture Search (NAS) in designing dataset-specific architectures, we propose Context-Aware Structure Adaptive Graph Neural Networks (CAS-GNN). This framework is capable of automatically selecting the appropriate aggregator for each node which is determined by both node attributes and local contextual information. The selection is formulated as the Markov Decision Process (MDP) optimized via Deep-Q-Network (DQN) training. Our contributions include a flexible framework incorporating various aggregators for individual nodes based on their attributes and local context, improved performance through node-specific aggregator selection, and extensive experimental validation demonstrating the effectiveness of CAS-GNN on multiple real-world datasets.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Reddy, P. N. Aravinda; Ramachandra, Raghavendra; Rao, K. Sreenivasa; Mitra, Pabitra
NeuralMultiling: A Novel Neural Architecture Search for Smartphone Based Multilingual Speaker Verification Proceedings Article
In: Antonacopoulos, Apostolos; Chaudhuri, Subhasis; Chellappa, Rama; Liu, Cheng-Lin; Bhattacharya, Saumik; Pal, Umapada (Ed.): Pattern Recognition, pp. 406–423, Springer Nature Switzerland, Cham, 2025, ISBN: 978-3-031-78341-8.
@inproceedings{10.1007/978-3-031-78341-8_26,
title = {NeuralMultiling: A Novel Neural Architecture Search for Smartphone Based Multilingual Speaker Verification},
author = {P. N. Aravinda Reddy and Raghavendra Ramachandra and K. Sreenivasa Rao and Pabitra Mitra},
editor = {Apostolos Antonacopoulos and Subhasis Chaudhuri and Rama Chellappa and Cheng-Lin Liu and Saumik Bhattacharya and Umapada Pal},
url = {https://link.springer.com/chapter/10.1007/978-3-031-78341-8_26},
isbn = {978-3-031-78341-8},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {Pattern Recognition},
pages = {406–423},
publisher = {Springer Nature Switzerland},
address = {Cham},
abstract = {Multilingual speaker verification introduces the challenge of verifying a speaker in multiple languages. Existing systems were built using i-vector/x-vector approaches along with Bi-LSTMs, which were trained to discriminate speakers, irrespective of the language. Instead of exploring the design space manually, we propose a neural architecture search for multilingual speaker verification suitable for mobile devices, called NeuralMultiling. First, our algorithm searches for an optimal operational combination of neural cells with different architectures for normal cells and reduction cells and then derives a CNN model by stacking neural cells. Using the derived architecture, we performed two different studies:1) language agnostic condition and 2) interoperability between languages and devices on the publicly available Multilingual Audio-Visual Smartphone (MAVS) dataset. The experimental results suggest that the derived architecture significantly outperforms the existing Autospeech method by a 5–6% reduction in the Equal Error Rate (EER) with fewer model parameters.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Li, Jie; Wang, Yuxia; Wang, Yifan; Yu, Ruiyun; Wang, Xingwei
HN-Darts:Hybrid Network Differentiable Architecture Search for Industrial Scenarios Proceedings Article
In: Hadfi, Rafik; Anthony, Patricia; Sharma, Alok; Ito, Takayuki; Bai, Quan (Ed.): PRICAI 2024: Trends in Artificial Intelligence, pp. 322–327, Springer Nature Singapore, Singapore, 2025, ISBN: 978-981-96-0116-5.
@inproceedings{10.1007/978-981-96-0116-5_26,
title = {HN-Darts:Hybrid Network Differentiable Architecture Search for Industrial Scenarios},
author = {Jie Li and Yuxia Wang and Yifan Wang and Ruiyun Yu and Xingwei Wang},
editor = {Rafik Hadfi and Patricia Anthony and Alok Sharma and Takayuki Ito and Quan Bai},
url = {https://link.springer.com/chapter/10.1007/978-981-96-0116-5_26},
isbn = {978-981-96-0116-5},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {PRICAI 2024: Trends in Artificial Intelligence},
pages = {322–327},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Neural architecture search is a powerful tool in image processing, automating model construction and reducing human involvement. However, its deployment on edge devices with limited computing resources is often impeded by the size of large models, a concern overlooked by most NAS methods focused solely on accuracy. We propose a hybrid network search approach that integrates the glore_unit, a novel component that replaces traditional cells to optimize model size without sacrificing accuracy. By leveraging the Differentiable Architecture Search (Darts) and a Googlenet-like hypernet, we've redefined the search space to prioritize compactness and precision, enhanced by a temperature factor for more reliable search selections. Our experiments on cifar10 and ImageNet showcase a model with a 2.35% error rate and 2.76M parameters on cifar10, and a Top-1 error rate of 23.75%, Top-5 error rate of 7.13% with 3.9M parameters on ImageNet, demonstrating SOTA accuracy with a significant reduction in model parameters, making it suitable for environments with constrained computational resources.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Jin, Xiao; Yu, Wen; Chen, Dai-Wei; Shi, Wei
DFD-NAS: General deepfake detection via efficient neural architecture search Journal Article
In: Neurocomputing, vol. 619, pp. 129129, 2025, ISSN: 0925-2312.
@article{JIN2025129129,
title = {DFD-NAS: General deepfake detection via efficient neural architecture search},
author = {Xiao Jin and Wen Yu and Dai-Wei Chen and Wei Shi},
url = {https://www.sciencedirect.com/science/article/pii/S0925231224019003},
doi = {https://doi.org/10.1016/j.neucom.2024.129129},
issn = {0925-2312},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Neurocomputing},
volume = {619},
pages = {129129},
abstract = {As the saying goes, “seeing is believing”. However, with the development of digital face editing tools, we can no longer trust what we can see. Though face forgery detection has made promising progress, most current methods are manually designed by human experts, which is labor-intensive. In this paper, we develop an end-to-end framework based on neural architecture search (NAS) for deepfake detection, which can automatically design network architectures without human intervention. First, a forgery-oriented search space is created to choose appropriate operations for this task, which facilitates the search process in finding the most valuable gradient information for face forgery detection. Second, inspired by the fact that the gap between training error and test error is a good indicator of generalization ability for a classification task, we propose a novel performance estimation metric. This metric encourages the error in the performance estimation phase to be close to the error in the search phase, which guides the search process to select more general models. The cross-dataset search is also considered to develop more general architectures. Eventually, we design a cell cascaded pyramid network (C2PN) for final detection, which aggregates multiscale features for performance improvements. Compared with state-of-the-art networks artificially designed, our method achieves competitive performance in both in-dataset and cross-dataset scenarios.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Sarker, Prodip Kumar
Transformer-based neural architecture search for effective visible-infrared person re-identification Journal Article
In: Neurocomputing, vol. 620, pp. 129257, 2025, ISSN: 0925-2312.
@article{SARKER2025129257,
title = {Transformer-based neural architecture search for effective visible-infrared person re-identification},
author = {Prodip Kumar Sarker},
url = {https://www.sciencedirect.com/science/article/pii/S0925231224020289},
doi = {https://doi.org/10.1016/j.neucom.2024.129257},
issn = {0925-2312},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Neurocomputing},
volume = {620},
pages = {129257},
abstract = {Visible-infrared person re-identification (VI-reID) is a complex task in security and video surveillance that aims to identify and match a person captured by various non-overlapping cameras. In recent years, there has been a notable advancement in reID owing to the development of transformer-based architectures. Although many existing methods emphasize on learning both modality-specific and shared features, challenges remain in fully exploiting the complementary information between infrared and visible modalities. Consequently, there is still opportunity to increase retrieval performance by effectively comprehending and integrating cross-modality semantic information. These designs often have problems with model complexity and time-consuming processes. To tackle these issues, we employ a novel transformer-based neural architecture search (TNAS) deep learning approach for effective VI-reID. To alleviate modality gaps, we first introduce a global–local transformer (GLT) module that captures features at both global and local levels across different modalities, contributing to better feature representation and matching. Then, an efficient neural architecture search (NAS) module is developed to search for the optimal transformer-based architecture, which further enhances the performance of VI-reID. Additionally, we introduce distillation loss and modality discriminative (MD) loss to examine the potential consistency between different modalities to promote intermodality separation between classes and intramodality compactness within classes. Experimental results on two challenging benchmark datasets illustrate that our developed model achieves state-of-the-art results, outperforming existing VI-reID methods.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Zhang, Zhengzhuo; Zhuang, Liansheng
Progressive Neural Architecture Generation with Weaker Predictors Proceedings Article
In: Ide, Ichiro; Kompatsiaris, Ioannis; Xu, Changsheng; Yanai, Keiji; Chu, Wei-Ta; Nitta, Naoko; Riegler, Michael; Yamasaki, Toshihiko (Ed.): MultiMedia Modeling, pp. 229–242, Springer Nature Singapore, Singapore, 2025, ISBN: 978-981-96-2064-7.
@inproceedings{10.1007/978-981-96-2064-7_17,
title = {Progressive Neural Architecture Generation with Weaker Predictors},
author = {Zhengzhuo Zhang and Liansheng Zhuang},
editor = {Ichiro Ide and Ioannis Kompatsiaris and Changsheng Xu and Keiji Yanai and Wei-Ta Chu and Naoko Nitta and Michael Riegler and Toshihiko Yamasaki},
isbn = {978-981-96-2064-7},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {MultiMedia Modeling},
pages = {229–242},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Neural Architecture Search (NAS) methods often suffer from low search efficiency since they have to explore a large and complex architecture search space. To accelerate the architecture search, generative methods learn a search space capturing intricate architecture distributions and generate promising architectures guided by a strong predictor within the latent space. However, since the architecture space is often exponentially large and highly non-convex, even a very strong predictor model is difficult in fitting the whole space, which may degrade their performance. To address this problem, this paper proposes a novel framework named Progressive Neural Architecture Generation with Weaker Predictors (WeakPNAG), which uses conditional diffusion models to generate promising architectures guided by weak predictors. Different from existing generative NAS methods which use a single strong predictor, our WeakPNAG progressively shrinks the sample space based on predictions from previous weak predictors, and updates new weak predictors towards the subspace of better architectures. In this way, our WeakPNAG iteratively learns to generate samples from increasingly promising latent subspaces. Extensive experiments on standard benchmarks demonstrate that our WeakPNAG achieves superior performance with reduced evaluation time compared with SOTA NAS methods.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Su, Yulan; Zhang, Sisi; Lin, Zechao; Wang, Xingbin; Zhao, Lutan; Meng, Dan; Hou, Rui
Poseidon: A NAS-Based Ensemble Defense Method Against Multiple Perturbations Proceedings Article
In: Ide, Ichiro; Kompatsiaris, Ioannis; Xu, Changsheng; Yanai, Keiji; Chu, Wei-Ta; Nitta, Naoko; Riegler, Michael; Yamasaki, Toshihiko (Ed.): MultiMedia Modeling, pp. 215–228, Springer Nature Singapore, Singapore, 2025, ISBN: 978-981-96-2064-7.
@inproceedings{10.1007/978-981-96-2064-7_16,
title = {Poseidon: A NAS-Based Ensemble Defense Method Against Multiple Perturbations},
author = {Yulan Su and Sisi Zhang and Zechao Lin and Xingbin Wang and Lutan Zhao and Dan Meng and Rui Hou},
editor = {Ichiro Ide and Ioannis Kompatsiaris and Changsheng Xu and Keiji Yanai and Wei-Ta Chu and Naoko Nitta and Michael Riegler and Toshihiko Yamasaki},
isbn = {978-981-96-2064-7},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {MultiMedia Modeling},
pages = {215–228},
publisher = {Springer Nature Singapore},
address = {Singapore},
abstract = {Deep learning models have been proven to be severely affected by adversarial examples, which limit the widespread deployment of deep learning models. Prior research largely focused on defending against single types of perturbations using a single network. However, these methods are susceptible to sacrificing defense range because defense models that are trained to be robust against one perturbation type may not be robust against other types. Moreover, it is unrealistic to assume that neural networks would only be affected by a single type of perturbation. To defend against multiple perturbations, recent works have attempted to improve the overall robustness against multiple perturbations. Nonetheless, when evaluating the model's robustness against each type of perturbation, multi-perturbation defenses are still significantly less effective than models that are robust against a single perturbation type. To address these issues, we propose Poseidon, an ensemble defense method based on neural architecture search to defend against multiple perturbations. We first highlight the importance of architecture evolution in enhancing model robustness. And a novel robust architecture search method is proposed to identify perturbation-tailored architectures for sub-models. Furthermore, we explore a dedicated ensemble method that can combine these diverse sub-model architectures to be robust against multiple types of perturbations. The experimental results demonstrate that Poseidon outperforms the state-of-the-art multiple perturbation defense methods by 10.9% and 7.4% in robustness on the CIFAR-10 and CIFAR-100 datasets, respectively.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Amoury, Sofia El; Smili, Youssef; Fakhri, Youssef
Design of an Optimal Convolutional Neural Network Architecture for MRI Brain Tumor Classification by Exploiting Particle Swarm Optimization Journal Article
In: Preprints, 2025.
@article{202501.0040,
title = {Design of an Optimal Convolutional Neural Network Architecture for MRI Brain Tumor Classification by Exploiting Particle Swarm Optimization},
author = {Sofia El Amoury and Youssef Smili and Youssef Fakhri},
url = {https://doi.org/10.20944/preprints202501.0040.v1},
doi = {10.20944/preprints202501.0040.v1},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Preprints},
publisher = {Preprints},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Capello, Alessio; Berta, Riccardo; Ballout, Hadi; Fresta, Matteo; Soltanmuradov, Vafali; Bellotti, Francesco
Enhancing μNAS for 1D CNNs on Microcontrollers Proceedings Article
In: Valle, Maurizio; Gastaldo, Paolo; Limiti, Ernesto (Ed.): Proceedings of SIE 2024, pp. 481–486, Springer Nature Switzerland, Cham, 2025, ISBN: 978-3-031-71518-1.
@inproceedings{10.1007/978-3-031-71518-1_59,
title = {Enhancing μNAS for 1D CNNs on Microcontrollers},
author = {Alessio Capello and Riccardo Berta and Hadi Ballout and Matteo Fresta and Vafali Soltanmuradov and Francesco Bellotti},
editor = {Maurizio Valle and Paolo Gastaldo and Ernesto Limiti},
url = {https://link.springer.com/chapter/10.1007/978-3-031-71518-1_59},
isbn = {978-3-031-71518-1},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {Proceedings of SIE 2024},
pages = {481–486},
publisher = {Springer Nature Switzerland},
address = {Cham},
abstract = {Deep Learning (DL) has proved effective in a variety of application domains. However, the computational and memory demand posed by deep neural models limits the spread of DL on resource-constrained devices such as microcontrollers. An opportunity to tailor DL models to specific hardware constraints is given by Neural Architecture Search (NAS), which automatically traverses a large search space, seeking for optimal architectures both in terms of hardware and performance, based on user specifications. State of the art open-source NAS tools for microcontrollers only support 2D Convolutional Neural Network (CNN) and Multi Layer Perceptron (MLP), but do not consider 1D convolution, which is key for time series analysis and signal processing. This study focuses on enhancing the state-of-the-art μNAS framework, by adding support for 1D CNN. Preliminary tests on a dummy dataset consisting of simple gaussian-distributed waveforms, demonstrate the system ability to find appropriate architectures to satisfy the specified constraints.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Xu, Jingjing; Wu, Caesar; Li, Yuan-Fang; Danoy, Grégoire; Bouvry, Pascal
A Unified Hyperparameter Optimization Pipeline for Transformer-Based Time Series Forecasting Models Miscellaneous
2025.
@misc{xu2025unifiedhyperparameteroptimizationpipeline,
title = {A Unified Hyperparameter Optimization Pipeline for Transformer-Based Time Series Forecasting Models},
author = {Jingjing Xu and Caesar Wu and Yuan-Fang Li and Grégoire Danoy and Pascal Bouvry},
url = {https://arxiv.org/abs/2501.01394},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Ge, Chendi; Wang, Xin; Zhang, Ziwei; Qin, Yijian; Wu, Haiyang; Zhang, Yang; Yang, Yuekui; Zhu, Wenwu
Behavior Importance-Aware Graph Neural Architecture Search for Cross-Domain Recommendation Miscellaneous
2025.
@misc{nokey,
title = {Behavior Importance-Aware Graph Neural Architecture Search for Cross-Domain Recommendation},
author = {Chendi Ge and Xin Wang and Ziwei Zhang and Yijian Qin and Haiyang Wu and Yang Zhang and Yuekui Yang and Wenwu Zhu},
url = {https://mn.cs.tsinghua.edu.cn/xinwang/PDF/papers/2025_Behavior%20Importance-Aware%20Graph%20Neural%20Architecture%20Search%20for%20Cross-Domain%20Recommendation.pdf},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Wang, Mingzi; Meng, Yuan; Tang, Chen; Zhang, Weixiang; Qin, Yijian; Yao, Yang; Li, Yingxin; Feng, Tongtong; Wang, Xin; Guan, Xun; Wang, Zhi; Zhu, Wenwu
JAQ: Joint Efficient Architecture Design and Low-Bit Quantization with Hardware-Software Co-Exploration Miscellaneous
2025.
@misc{nokey,
title = {JAQ: Joint Efficient Architecture Design and Low-Bit Quantization with Hardware-Software Co-Exploration},
author = {Mingzi Wang and Yuan Meng and Chen Tang and Weixiang Zhang and Yijian Qin and Yang Yao and Yingxin Li and Tongtong Feng and Xin Wang and Xun Guan and Zhi Wang and Wenwu Zhu},
url = {https://mn.cs.tsinghua.edu.cn/xinwang/PDF/papers/2025_JAQ%20Joint%20Efficient%20Architecture%20Design%20and%20Low-Bit%20Quantization%20with%20Hardware-Software%20Co-Exploration.pdf},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Avval, Sasan Salmani Pour; Eskue, Nathan D.; Groves, Roger M.; Yaghoubi, Vahid
Systematic review on neural architecture search Journal Article
In: Artificial Intelligence Review , 2025.
@article{Avval-air25a,
title = {Systematic review on neural architecture search},
author = {Sasan Salmani Pour Avval and Nathan D. Eskue and Roger M. Groves and Vahid Yaghoubi
},
url = {https://link.springer.com/article/10.1007/s10462-024-11058-w},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Artificial Intelligence Review },
keywords = {},
pubstate = {published},
tppubtype = {article}
}
He, Yannis Y.
TART: Token-based Architecture Transformer for Neural Network Performance Prediction Technical Report
2025.
@techreport{he2025tarttokenbasedarchitecturetransformer,
title = {TART: Token-based Architecture Transformer for Neural Network Performance Prediction},
author = {Yannis Y. He},
url = {https://arxiv.org/abs/2501.02007},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
keywords = {},
pubstate = {published},
tppubtype = {techreport}
}
Archet, Agathe; Ventroux, Nicolas; Gac, Nicolas; Orieux, François
A practical HW-aware NAS flow for AI vision applications on embedded heterogeneous SoCs Proceedings Article
In: International Workshop on Design and Architecture for Signal and Image Processing 2025, Barcelone, Spain, 2025.
@inproceedings{archet:hal-04869471,
title = {A practical HW-aware NAS flow for AI vision applications on embedded heterogeneous SoCs},
author = {Agathe Archet and Nicolas Ventroux and Nicolas Gac and François Orieux},
url = {https://hal.science/hal-04869471},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {International Workshop on Design and Architecture for Signal and Image Processing 2025},
address = {Barcelone, Spain},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
