Attention-based neural network for end-to-end music separation
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Wang, Jing | - |
dc.contributor.author | Liu, Hanyue | - |
dc.contributor.author | Ying, Haorong | - |
dc.contributor.author | Qiu, Chuhan | - |
dc.contributor.author | Li, Jingxin | - |
dc.contributor.author | Anwar, Muhammad Shahid | - |
dc.date.accessioned | 2023-03-22T05:40:08Z | - |
dc.date.available | 2023-03-22T05:40:08Z | - |
dc.date.created | 2023-02-14 | - |
dc.date.issued | 2023-06 | - |
dc.identifier.issn | 2468-6557 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/87197 | - |
dc.description.abstract | The end-to-end separation algorithm with superior performance in the field of speech separation has not been effectively used in music separation. Moreover, since music signals are often dual channel data with a high sampling rate, how to model long-sequence data and make rational use of the relevant information between channels is also an urgent problem to be solved. In order to solve the above problems, the performance of the end-to-end music separation algorithm is enhanced by improving the network structure. Our main contributions include the following: (1) A more reasonable densely connected U-Net is designed to capture the long-term characteristics of music, such as main melody, tone and so on. (2) On this basis, the multi-head attention and dual-path transformer are introduced in the separation module. Channel attention units are applied recursively on the feature map of each layer of the network, enabling the network to perform long-sequence separation. Experimental results show that after the introduction of the channel attention, the performance of the proposed algorithm has a stable improvement compared with the baseline system. On the MUSDB18 dataset, the average score of the separated audio exceeds that of the current best-performing music separation algorithm based on the time-frequency domain (T-F domain). | - |
dc.language | 영어 | - |
dc.language.iso | en | - |
dc.publisher | WILEY | - |
dc.relation.isPartOf | CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY | - |
dc.title | Attention-based neural network for end-to-end music separation | - |
dc.type | Article | - |
dc.type.rims | ART | - |
dc.description.journalClass | 1 | - |
dc.identifier.wosid | 000914034200001 | - |
dc.identifier.doi | 10.1049/cit2.12163 | - |
dc.identifier.bibliographicCitation | CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, v.8, no.2, pp.355 - 363 | - |
dc.description.isOpenAccess | Y | - |
dc.identifier.scopusid | 2-s2.0-85147000869 | - |
dc.citation.endPage | 363 | - |
dc.citation.startPage | 355 | - |
dc.citation.title | CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY | - |
dc.citation.volume | 8 | - |
dc.citation.number | 2 | - |
dc.contributor.affiliatedAuthor | Anwar, Muhammad Shahid | - |
dc.type.docType | Article | - |
dc.subject.keywordAuthor | channel attention | - |
dc.subject.keywordAuthor | densely connected network | - |
dc.subject.keywordAuthor | end-to-end music separation | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Artificial Intelligence | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
1342, Seongnam-daero, Sujeong-gu, Seongnam-si, Gyeonggi-do, Republic of Korea(13120)031-750-5114
COPYRIGHT 2020 Gachon University All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.