Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

ISFRNet: A Deep Three-stage Identity and Structure Feature Refinement Network for Facial Image Inpaintingopen access

Authors
Wang, Y.[Wang, Y.]Shin, J.[Shin, J.]
Issue Date
1-Mar-2023
Publisher
Korean Society for Internet Information
Keywords
Deep Learning; GAN.; Image Inpainting
Citation
KSII Transactions on Internet and Information Systems, v.17, no.3, pp.881 - 895
Indexed
SCIE
SCOPUS
KCI
Journal Title
KSII Transactions on Internet and Information Systems
Volume
17
Number
3
Start Page
881
End Page
895
URI
https://scholarworks.bwise.kr/skku/handle/2021.sw.skku/104025
DOI
10.3837/tiis.2023.03.011
ISSN
1976-7277
Abstract
Modern image inpainting techniques based on deep learning have achieved remarkable performance, and more and more people are working on repairing more complex and larger missing areas, although this is still challenging, especially for facial image inpainting. For a face image with a huge missing area, there are very few valid pixels available; however, people have an ability to imagine the complete picture in their mind according to their subjective will. It is important to simulate this capability while maintaining the identity features of the face as much as possible. To achieve this goal, we propose a three-stage network model, which we refer to as the identity and structure feature refinement network (ISFRNet). ISFRNet is based on 1) a pre-Trained pSp-styleGAN model that generates an extremely realistic face image with rich structural features; 2) a shallow structured network with a small receptive field; and 3) a modified U-net with two encoders and a decoder, which has a large receptive field. We choose structural similarity index (SSIM), peak signal-To-noise ratio (PSNR), L1 Loss and learned perceptual image patch similarity (LPIPS) to evaluate our model. When the missing region is 20%-40%, the above four metric scores of our model are 28.12, 0.942, 0.015 and 0.090, respectively. When the lost area is between 40% and 60%, the metric scores are 23.31, 0.840, 0.053 and 0.177, respectively. Our inpainting network not only guarantees excellent face identity feature recovery but also exhibits state-of-The-Art performance compared to other multi-stage refinement models. © 2023 Korean Society for Internet Information. All rights reserved.
Files in This Item
There are no files associated with this item.
Appears in
Collections
Information and Communication Engineering > School of Electronic and Electrical Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher SHIN, JI TAE photo

SHIN, JI TAE
Information and Communication Engineering (Electronic and Electrical Engineering)
Read more

Altmetrics

Total Views & Downloads

BROWSE