Attention-Aided Generative Learning for Multi-Scale Multi-Modal Fundus Image Translation
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Pham, V[Pham, Van-Nguyen] | - |
dc.contributor.author | Le, DT[Le, Duc-Tai] | - |
dc.contributor.author | Bum, J[Bum, Junghyun] | - |
dc.contributor.author | Lee, EJ[Lee, Eun Jung] | - |
dc.contributor.author | Han, JC[Han, Jong Chul] | - |
dc.contributor.author | Choo, H[Choo, Hyunseung] | - |
dc.date.accessioned | 2023-07-03T02:51:35Z | - |
dc.date.available | 2023-07-03T02:51:35Z | - |
dc.date.created | 2023-07-03 | - |
dc.date.issued | 2023 | - |
dc.identifier.issn | 2169-3536 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/skku/handle/2021.sw.skku/106502 | - |
dc.description.abstract | Conventional fundus images (CFIs) and ultra-widefield fundus images (UFIs) are two fundamental image modalities in ophthalmology. While CFIs provide a detailed view of the optic nerve head and the posterior pole of an eye, their clinical use is associated with high costs and patient inconvenience due to the requirement of good pupil dilation. On the other hand, UFIs capture peripheral lesions, but their image quality is sensitive to factors such as pupil size, eye position, and eyelashes, leading to greater variability between examinations compared to CFIs. The widefield retina view of UFIs offers the theoretical possibility of generating CFIs from available UFIs to reduce patient examination costs. A recent study has shown the feasibility of this approach by leveraging deep learning techniques for the UFI-to-CFI translation task. However, the technique suffers from the heterogeneous scales of the image modalities and variations in the brightness of the training data. In this paper, we address these issues with a novel framework consisting of three stages: cropping, enhancement, and translation. The first stage is an optic disc-centered cropping strategy that helps to alleviate the scale difference between the two image domains. The second stage mitigates the variation in training data brightness and unifies the mask between the two modalities. In the last stage, we introduce an attention-aided generative learning model to translate a given UFI into the CFI domain. Our experimental results demonstrate the success of the proposed method on 1,011 UFIs, with 99.8% of the generated CFIs evaluated as good quality and usable. Expert evaluations confirm significant visual quality improvements in the generated CFIs compared to the UFIs, ranging from 10% to 80% for features such as optic nerve structure, vascular distribution, and drusen. Furthermore, using generated CFIs in an AI-based diagnosis system for age-related macular degeneration results in superior accuracy compared to UFIs and competitive performance relative to real CFIs. These results showcase the potential of our approach for automatic disease diagnosis and monitoring. | - |
dc.language | 영어 | - |
dc.language.iso | en | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Attention-Aided Generative Learning for Multi-Scale Multi-Modal Fundus Image Translation | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Pham, V[Pham, Van-Nguyen] | - |
dc.contributor.affiliatedAuthor | Le, DT[Le, Duc-Tai] | - |
dc.contributor.affiliatedAuthor | Bum, J[Bum, Junghyun] | - |
dc.contributor.affiliatedAuthor | Lee, EJ[Lee, Eun Jung] | - |
dc.contributor.affiliatedAuthor | Han, JC[Han, Jong Chul] | - |
dc.contributor.affiliatedAuthor | Choo, H[Choo, Hyunseung] | - |
dc.identifier.doi | 10.1109/ACCESS.2023.3278596 | - |
dc.identifier.scopusid | 2-s2.0-85161082780 | - |
dc.identifier.wosid | 001005656200001 | - |
dc.identifier.bibliographicCitation | IEEE ACCESS, v.11, pp.51701 - 51711 | - |
dc.relation.isPartOf | IEEE ACCESS | - |
dc.citation.title | IEEE ACCESS | - |
dc.citation.volume | 11 | - |
dc.citation.startPage | 51701 | - |
dc.citation.endPage | 51711 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.isOpenAccess | Y | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Telecommunications | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Telecommunications | - |
dc.subject.keywordAuthor | Optical imaging | - |
dc.subject.keywordAuthor | Task analysis | - |
dc.subject.keywordAuthor | Diseases | - |
dc.subject.keywordAuthor | Ophthalmology | - |
dc.subject.keywordAuthor | Mathematical models | - |
dc.subject.keywordAuthor | Deep learning | - |
dc.subject.keywordAuthor | Biomedical optical imaging | - |
dc.subject.keywordAuthor | Conventional fundus images | - |
dc.subject.keywordAuthor | deep learning | - |
dc.subject.keywordAuthor | generative learning | - |
dc.subject.keywordAuthor | ophthalmology | - |
dc.subject.keywordAuthor | unpaired image-to-image translation | - |
dc.subject.keywordAuthor | ultra wide-field fundus images | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(03063) 25-2, SUNGKYUNKWAN-RO, JONGNO-GU, SEOUL, KOREAsamsunglib@skku.edu
COPYRIGHT © 2021 SUNGKYUNKWAN UNIVERSITY ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.