Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Self-supervised multi-modal training from uncurated images and reports enables monitoring AI in radiology

Full metadata record
DC Field Value Language
dc.contributor.authorPark, Sangjoon-
dc.contributor.authorLee, Eun Sun-
dc.contributor.authorShin, Kyung Sook-
dc.contributor.authorLee, Jeong Eun-
dc.contributor.authorYe, Jong Chul-
dc.date.accessioned2024-01-09T14:36:06Z-
dc.date.available2024-01-09T14:36:06Z-
dc.date.issued2024-01-
dc.identifier.issn1361-8415-
dc.identifier.issn1361-8423-
dc.identifier.urihttps://scholarworks.bwise.kr/cau/handle/2019.sw.cau/70503-
dc.description.abstractThe escalating demand for artificial intelligence (AI) systems that can monitor and supervise human errors and abnormalities in healthcare presents unique challenges. Recent advances in vision-language models reveal the challenges of monitoring AI by understanding both visual and textual concepts and their semantic correspondences. However, there has been limited success in the application of vision-language models in the medical domain. Current vision-language models and learning strategies for photographic images and captions call for a web-scale data corpus of image and text pairs which is not often feasible in the medical domain. To address this, we present a model named medical cross-attention vision-language model (Medical X-VL), which leverages key components to be tailored for the medical domain. The model is based on the following components: self-supervised unimodal models in medical domain and a fusion encoder to bridge them, momentum distillation, sentencewise contrastive learning for medical reports, and sentence similarity-adjusted hard negative mining. We experimentally demonstrated that our model enables various zero-shot tasks for monitoring AI, ranging from the zero-shot classification to zero-shot error correction. Our model outperformed current state-of-the-art models in two medical image datasets, suggesting a novel clinical application of our monitoring AI model to alleviate human errors. Our method demonstrates a more specialized capacity for fine-grained understanding, which presents a distinct advantage particularly applicable to the medical domain. © 2023-
dc.language영어-
dc.language.isoENG-
dc.publisherElsevier B.V.-
dc.titleSelf-supervised multi-modal training from uncurated images and reports enables monitoring AI in radiology-
dc.typeArticle-
dc.identifier.doi10.1016/j.media.2023.103021-
dc.identifier.bibliographicCitationMedical Image Analysis, v.91-
dc.description.isOpenAccessN-
dc.identifier.wosid001112184700001-
dc.identifier.scopusid2-s2.0-85176503153-
dc.citation.titleMedical Image Analysis-
dc.citation.volume91-
dc.type.docTypeArticle-
dc.publisher.location네델란드-
dc.subject.keywordAuthorError detection-
dc.subject.keywordAuthorMonitoring AI-
dc.subject.keywordAuthorRadiograph-
dc.subject.keywordAuthorVision-language model-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaRadiology, Nuclear Medicine & Medical Imaging-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.relation.journalWebOfScienceCategoryComputer Science, Interdisciplinary Applications-
dc.relation.journalWebOfScienceCategoryEngineering, Biomedical-
dc.relation.journalWebOfScienceCategoryRadiology, Nuclear Medicine & Medical Imaging-
dc.description.journalRegisteredClassscopus-
Files in This Item
There are no files associated with this item.
Appears in
Collections
ETC > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Lee, Eun Sun photo

Lee, Eun Sun
의과대학 (의학부(임상-서울))
Read more

Altmetrics

Total Views & Downloads

BROWSE