RefCap: image captioning with referent objects attributesopen access
- Authors
- Park, Seokmok; Paik, Joonki
- Issue Date
- Dec-2023
- Publisher
- Nature Research
- Citation
- Scientific Reports, v.13, no.1
- Journal Title
- Scientific Reports
- Volume
- 13
- Number
- 1
- URI
- https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/71240
- DOI
- 10.1038/s41598-023-48916-6
- ISSN
- 2045-2322
- Abstract
- In recent years, significant progress has been made in visual-linguistic multi-modality research, leading to advancements in visual comprehension and its applications in computer vision tasks. One fundamental task in visual-linguistic understanding is image captioning, which involves generating human-understandable textual descriptions given an input image. This paper introduces a referring expression image captioning model that incorporates the supervision of interesting objects. Our model utilizes user-specified object keywords as a prefix to generate specific captions that are relevant to the target object. The model consists of three modules including: (i) visual grounding, (ii) referring object selection, and (iii) image captioning modules. To evaluate its performance, we conducted experiments on the RefCOCO and COCO captioning datasets. The experimental results demonstrate that our proposed method effectively generates meaningful captions aligned with users’ specific interests. © 2023, The Author(s).
- Files in This Item
-
- Appears in
Collections - Graduate School of Advanced Imaging Sciences, Multimedia and Film > Department of Imaging Science and Arts > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/71240)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.