Exploiting Deep Neural Networks for Two-to-Five Channel Surround Decoder
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Choi, Jeonghwan | - |
dc.contributor.author | Chang, Joon-Hyuk | - |
dc.date.accessioned | 2021-08-02T08:28:53Z | - |
dc.date.available | 2021-08-02T08:28:53Z | - |
dc.date.created | 2021-05-11 | - |
dc.date.issued | 2020-12 | - |
dc.identifier.issn | 1549-4950 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/8195 | - |
dc.description.abstract | We exploited deep neural networks (DNN) for two-to-five channel surround decoding. Specifically DNNs are used to replace the primary-ambient separation and ambient-signal rendering modules. For the training, the mean-squared error of the magnitude spectra between the decoded and five-channel target signals and the interchannel level differences between the target signals were used as the loss functions. Through this procedure the DNNs can derive the spectral weights that can be used to produce the decoded signals, similar to that for the target signals. The log spectral distance, signal-to-distortion ratio, and multiple stimuli with hidden reference and anchor tests were used for objective and subjective evaluations. The experimental results show that exploiting the DNNs can generate decoded signals that are more similar to the target signals than those obtained via previous methods. | - |
dc.language | 영어 | - |
dc.language.iso | en | - |
dc.publisher | AUDIO ENGINEERING SOC | - |
dc.title | Exploiting Deep Neural Networks for Two-to-Five Channel Surround Decoder | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Chang, Joon-Hyuk | - |
dc.identifier.doi | 10.17743/jaes.2020.0020 | - |
dc.identifier.scopusid | 2-s2.0-85100908515 | - |
dc.identifier.wosid | 000615261200006 | - |
dc.identifier.bibliographicCitation | JOURNAL OF THE AUDIO ENGINEERING SOCIETY, v.68, no.12, pp.938 - 949 | - |
dc.relation.isPartOf | JOURNAL OF THE AUDIO ENGINEERING SOCIETY | - |
dc.citation.title | JOURNAL OF THE AUDIO ENGINEERING SOCIETY | - |
dc.citation.volume | 68 | - |
dc.citation.number | 12 | - |
dc.citation.startPage | 938 | - |
dc.citation.endPage | 949 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Acoustics | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalWebOfScienceCategory | Acoustics | - |
dc.relation.journalWebOfScienceCategory | Engineering, Multidisciplinary | - |
dc.subject.keywordPlus | PRIMARY-AMBIENT EXTRACTION | - |
dc.subject.keywordPlus | SPEECH ENHANCEMENT | - |
dc.identifier.url | https://www.aes.org/e-lib/browse.cfm?elib=21008 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
222, Wangsimni-ro, Seongdong-gu, Seoul, 04763, Korea+82-2-2220-1365
COPYRIGHT © 2021 HANYANG UNIVERSITY.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.