Two-stage architectural fine-tuning for neural architecture search in efficient transfer learningopen access
- Authors
- Park, Soohyun; Son, Seok Bin; Lee, Youn Kyu; Jung, Soyi; Kim, Joongheon
- Issue Date
- Dec-2023
- Publisher
- WILEY
- Keywords
- image processing; neural nets; neural net architecture
- Citation
- ELECTRONICS LETTERS, v.59, no.24
- Journal Title
- ELECTRONICS LETTERS
- Volume
- 59
- Number
- 24
- URI
- https://scholarworks.bwise.kr/hongik/handle/2020.sw.hongik/32603
- DOI
- 10.1049/ell2.13066
- ISSN
- 0013-5194
1350-911X
- Abstract
- In many deep neural network (DNN) applications, the difficulty of gathering high-quality data in industry fields hinders the practical use of DNN. Thus, the concept of transfer learning (TL) has emerged, which leverages the pretrained knowledge of the DNN which was built based on large-scale datasets. For this TL objective, this paper suggests two-stage architectural fine-tuning for reducing the costs and time while exploring the most efficient DNN model, inspired by neural architecture search (NAS). The first stage is mutation, which reduces the search costs using a priori architectural information. Moreover, the next stage is early-stopping, which reduces NAS costs by terminating the search process in the middle of computation. The data-intensive experimental results verify that the proposed method outperforms benchmarks. This paper suggests two-stage architectural fine-tuning for reducing the costs and time while exploring the most efficient neural network model, inspired by neural architecture search (NAS). The first stage is mutation, which reduces the search costs using a priori architectural information. Moreover, the next stage is early-stopping, which reduces NAS costs by terminating the search process in the middle of computation.image
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > Computer Engineering > Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.