Regularizing activations in neural networks via distribution matching with the Wasserstein metric
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Joo, Taejong | - |
dc.contributor.author | Kang, Donggu | - |
dc.contributor.author | Kim, Byunghoon | - |
dc.date.accessioned | 2021-06-22T09:04:56Z | - |
dc.date.available | 2021-06-22T09:04:56Z | - |
dc.date.issued | 2020-04 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/1142 | - |
dc.description.abstract | Regularization and normalization have become indispensable components in training deep neural networks, resulting in faster training and improved generalization performance. We propose the projected error function regularization loss (PER) that encourages activations to follow the standard normal distribution. PER randomly projects activations onto one-dimensional space and computes the regularization loss in the projected space. PER is similar to the Pseudo-Huber loss in the projected space, thus taking advantage of both $L^1$ and $L^2$ regularization losses. Besides, PER can capture the interaction between hidden units by projection vector drawn from a unit sphere. By doing so, PER minimizes the upper bound of the Wasserstein distance of order one between an empirical distribution of activations and the standard normal distribution. To the best of the authors' knowledge, this is the first work to regularize activations via distribution matching in the probability distribution space. We evaluate the proposed method on the image classification task and the word-level language modeling task. | - |
dc.format.extent | 13 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | International Conference on Learning Representations | - |
dc.title | Regularizing activations in neural networks via distribution matching with the Wasserstein metric | - |
dc.type | Article | - |
dc.publisher.location | 미국 | - |
dc.identifier.bibliographicCitation | International Conference on Learning Representations 2020, pp 1 - 13 | - |
dc.citation.title | International Conference on Learning Representations 2020 | - |
dc.citation.startPage | 1 | - |
dc.citation.endPage | 13 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | foreign | - |
dc.subject.keywordPlus | Computer Science - Machine Learning | - |
dc.subject.keywordPlus | Statistics - Machine Learning | - |
dc.identifier.url | https://arxiv.org/abs/2002.05366 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
55 Hanyangdeahak-ro, Sangnok-gu, Ansan, Gyeonggi-do, 15588, Korea+82-31-400-4269 sweetbrain@hanyang.ac.kr
COPYRIGHT © 2021 HANYANG UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.