Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Not All Layers Are Equal: A Layer-Wise Adaptive Approach Toward Large-Scale DNN Training

Full metadata record
DC Field Value Language
dc.contributor.authorKo, Yunyong-
dc.contributor.authorLee, Dongwon-
dc.contributor.authorKim, Ssng Wook-
dc.date.accessioned2022-07-06T04:12:02Z-
dc.date.available2022-07-06T04:12:02Z-
dc.date.created2022-06-03-
dc.date.issued2022-04-
dc.identifier.issn0000-0000-
dc.identifier.urihttps://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/138794-
dc.description.abstractA large-batch training with data parallelism is a widely adopted approach to efficiently train a large deep neural network (DNN) model. Large-batch training, however, often suffers from the problem of the model quality degradation because of its fewer iterations. To alleviate this problem, in general, learning rate (lr) scaling methods have been applied, which increases the learning rate to make an update larger at each iteration. Unfortunately, however, we observe that large-batch training with state-of-the-art lr scaling methods still often degrade the model quality when a batch size crosses a specific limit, rendering such lr methods less useful. To this phenomenon, we hypothesize that existing lr scaling methods overlook the subtle but important differences across layersin training, which results in the degradation of the overall model quality. From this hypothesis, we propose a novel approach (LENA) toward the learning rate scaling for large-scale DNN training, employing: (1) a layer-wise adaptive lr scaling to adjust lr for each layer individually, and (2) a layer-wise state-aware warm-up to track the state of the training for each layer and finish its warm-up automatically. The comprehensive evaluation with variations of batch sizes demonstrates that LENA achieves the target accuracy (i.e., the accuracy of single-worker training): (1) within the fewest iterations across different batch sizes (up to 45.2% fewer iterations and 44.7% shorter time than the existing state-of-the-art method), and (2) for training very large-batch sizes, surpassing the limits of all baselines.-
dc.language영어-
dc.language.isoen-
dc.publisherAssociation for Computing Machinery, Inc-
dc.titleNot All Layers Are Equal: A Layer-Wise Adaptive Approach Toward Large-Scale DNN Training-
dc.typeArticle-
dc.contributor.affiliatedAuthorKim, Ssng Wook-
dc.identifier.doi10.1145/3485447.3511989-
dc.identifier.scopusid2-s2.0-85129795467-
dc.identifier.wosid000852713001087-
dc.identifier.bibliographicCitationWWW 2022 - Proceedings of the ACM Web Conference 2022, pp.1851 - 1859-
dc.relation.isPartOfWWW 2022 - Proceedings of the ACM Web Conference 2022-
dc.citation.titleWWW 2022 - Proceedings of the ACM Web Conference 2022-
dc.citation.startPage1851-
dc.citation.endPage1859-
dc.type.rimsART-
dc.type.docTypeProceedings Paper-
dc.description.journalClass1-
dc.description.isOpenAccessN-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalWebOfScienceCategoryComputer Science, Cybernetics-
dc.relation.journalWebOfScienceCategoryComputer Science, Software Engineering-
dc.relation.journalWebOfScienceCategoryComputer Science, Theory & Methods-
dc.subject.keywordAuthorlarge batch training-
dc.subject.keywordAuthorlayer-wise approach-
dc.subject.keywordAuthorlearning rate scaling-
dc.identifier.urlhttps://dl.acm.org/doi/10.1145/3485447.3511989-
Files in This Item
Go to Link
Appears in
Collections
서울 공과대학 > 서울 컴퓨터소프트웨어학부 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Kim, Sang-Wook photo

Kim, Sang-Wook
COLLEGE OF ENGINEERING (SCHOOL OF COMPUTER SCIENCE)
Read more

Altmetrics

Total Views & Downloads

BROWSE