Adaptive Dynamic Search for Multi-Task Learningopen access
- Authors
- Kim, Eunwoo
- Issue Date
- Nov-2022
- Publisher
- MDPI
- Keywords
- cost-adaptive solution; destructive interference; dynamic model search; memory efficiency; multi-task learning
- Citation
- Applied Sciences (Switzerland), v.12, no.22
- Journal Title
- Applied Sciences (Switzerland)
- Volume
- 12
- Number
- 22
- URI
- https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/61155
- DOI
- 10.3390/app122211836
- ISSN
- 2076-3417
2076-3417
- Abstract
- Multi-task learning (MTL) is a learning strategy for solving multiple tasks simultaneously while exploiting commonalities and differences between tasks for improved learning efficiency and prediction performance. Despite its potential, there remain several major challenges to be addressed. First of all, the task performance degrades when the number of tasks to solve increases or the tasks are less related. In addition, finding the prediction model for each task is typically laborious and can be suboptimal. This nature of manually designing the architecture further aggravates the problem when it comes to solving multiple tasks under different computational budgets. In this work, we propose a novel MTL approach to address these issues. The proposed method learns to search in a finely modularized base network dynamically and to discover an optimal prediction model for each instance of a task on the fly while taking the computational costs of the discovered models into account. We evaluate our learning framework on a diverse set of MTL scenarios comprising standard benchmark datasets. We achieve significant improvements in performance for all tested cases compared with existing MTL alternatives. © 2022 by the author.
- Files in This Item
-
- Appears in
Collections - College of Software > School of Computer Science and Engineering > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.