Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Convolutional Network With Twofold Feature Augmentation for Diabetic Retinopathy Recognition From Multi-Modal Images

Authors
Hua, Cam-HaoKim, KiyoungThien Huynh-TheYou, Jong InYu, Seung-YoungLe-Tien, ThuongBae, Sung-HoLee, Sungyoung
Issue Date
Jul-2021
Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
Keywords
Streaming media; Feature extraction; Diabetes; Retinopathy; Retina; Image recognition; Bioinformatics; Convolutional network; diabetic retinopathy recognition; fundus photograph; multi-modal images; SS-OCT angiography; twofold feature augmentation
Citation
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, v.25, no.7, pp 2686 - 2697
Pages
12
Journal Title
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS
Volume
25
Number
7
Start Page
2686
End Page
2697
URI
https://scholarworks.bwise.kr/kumoh/handle/2020.sw.kumoh/28269
DOI
10.1109/JBHI.2020.3041848
ISSN
2168-2194
2168-2208
Abstract
Objective: With the scenario of limited labeled dataset, this paper introduces a deep learning-based approach that leverages Diabetic Retinopathy (DR) severity recognition performance using fundus images combined with wide-field swept-source optical coherence tomography angiography (SS-OCTA). Methods: The proposed architecture comprises a backbone convolutional network associated with a Twofold Feature Augmentation mechanism, namely TFA-Net. The former includes multiple convolution blocks extracting representational features at various scales. The latter is constructed in a two-stage manner, i.e., the utilization of weight-sharing convolution kernels and the deployment of a Reverse Cross-Attention (RCA) stream. Results: The proposed model achieves a Quadratic Weighted Kappa rate of 90.2% on the small-sized internal KHUMC dataset. The robustness of the RCA stream is also evaluated by the single-modal Messidor dataset, of which the obtained mean Accuracy (94.8%) and Area Under Receiver Operating Characteristic (99.4%) outperform those of the state-of-the-arts significantly. Conclusion: Utilizing a network strongly regularized at feature space to learn the amalgamation of different modalities is of proven effectiveness. Thanks to the widespread availability of multi-modal retinal imaging for each diabetes patient nowadays, such approach can reduce the heavy reliance on large quantity of labeled visual data. Significance: Our TFA-Net is able to coordinate hybrid information of fundus photos and wide-field SS-OCTA for exhaustively exploiting DR-oriented biomarkers. Moreover, the embedded feature-wise augmentation scheme can enrich generalization ability efficiently despite learning from small-scale labeled data.
Files in This Item
There are no files associated with this item.
Appears in
Collections
ETC > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE