Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Compression of Deep-Learning Models Through Global Weight Pruning Using Alternating Direction Method of Multipliers

Full metadata record
DC Field Value Language
dc.contributor.authorLee, Kichun-
dc.contributor.authorHwangbo, Sunghun-
dc.contributor.authorYang, Dongwook-
dc.contributor.authorLee, Geonseok-
dc.date.accessioned2023-11-14T02:00:08Z-
dc.date.available2023-11-14T02:00:08Z-
dc.date.created2023-03-08-
dc.date.issued2023-02-
dc.identifier.issn1875-6891-
dc.identifier.urihttps://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/192070-
dc.description.abstractDeep learning has shown excellent performance in numerous machine-learning tasks, but one practical obstacle in deep learning is that the amount of computation and required memory is huge. Model compression, especially in deep learning, is very useful because it saves memory and reduces storage size while maintaining model performance. Model compression in a layered network structure aims to reduce the number of edges by pruning weights that are deemed unnecessary during the calculation. However, existing weight pruning methods perform a layer-by-layer reduction, which requires a predefined removal-ratio constraint for each layer. Layer-by-layer removal ratios must be structurally specified depending on the task, causing a sharp increase in the training time due to a large number of tuning parameters. Thus, such a layer-by-layer strategy is hardly feasible for deep layered models. Our proposed method aims to perform weight pruning in a deep layered network, while producing similar performance, by setting a global removal ratio for the entire model without prior knowledge of the structural characteristics. Our experiments with the proposed method show reliable and high-quality performance, obviating layer-by-layer removal ratios. Furthermore, experiments with increasing layers yield a pattern in the pruned weights that could provide an insight into the layers’ structural importance. The experiment with the LeNet-5 model using MNIST data results in a higher compression ratio of 98.8% for the proposed method, outperforming existing pruning algorithms. In the Resnet-56 experiment, the performance change according to removal ratios of 10–90% is investigated, and a higher removal ratio is achieved compared to other tested models. We also demonstrate the effectiveness of the proposed method with YOLOv4, a real-life object-detection model requiring substantial computation.-
dc.language영어-
dc.language.isoen-
dc.publisherSpringer Science and Business Media B.V.-
dc.titleCompression of Deep-Learning Models Through Global Weight Pruning Using Alternating Direction Method of Multipliers-
dc.typeArticle-
dc.contributor.affiliatedAuthorLee, Kichun-
dc.identifier.doi10.1007/s44196-023-00202-z-
dc.identifier.scopusid2-s2.0-85148941525-
dc.identifier.wosid000939369800001-
dc.identifier.bibliographicCitationInternational Journal of Computational Intelligence Systems, v.16, no.1, pp.1 - 13-
dc.relation.isPartOfInternational Journal of Computational Intelligence Systems-
dc.citation.titleInternational Journal of Computational Intelligence Systems-
dc.citation.volume16-
dc.citation.number1-
dc.citation.startPage1-
dc.citation.endPage13-
dc.type.rimsART-
dc.type.docTypeArticle-
dc.description.journalClass1-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.relation.journalWebOfScienceCategoryComputer Science, Interdisciplinary Applications-
dc.subject.keywordPlusLayer by layer-
dc.subject.keywordPlusLayer removal-
dc.subject.keywordPlusLayered network-
dc.subject.keywordPlusModel compression-
dc.subject.keywordPlusNetwork compression-
dc.subject.keywordPlusNonconvex optimization-
dc.subject.keywordPlusParallel com- puting-
dc.subject.keywordPlusPerformance-
dc.subject.keywordPlusRemoval ratios-
dc.subject.keywordPlusWeight pruning-
dc.subject.keywordAuthorNetwork compression-
dc.subject.keywordAuthorNon-convex optimization-
dc.subject.keywordAuthorParallel computing-
dc.subject.keywordAuthorWeight pruning-
dc.identifier.urlhttps://link.springer.com/article/10.1007/s44196-023-00202-z-
Files in This Item
Appears in
Collections
서울 공과대학 > 서울 산업공학과 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Lee, Ki chun photo

Lee, Ki chun
COLLEGE OF ENGINEERING (DEPARTMENT OF INDUSTRIAL ENGINEERING)
Read more

Altmetrics

Total Views & Downloads

BROWSE