Adversarial2Adversarial: Defending against Adversarial Fingerprint Attacks without Clean Images
- Authors
- Hong, Pyo Min; Hyun Kang, So; Kim, Jinhyeon; Kim, Ji Hoo; Kyu Lee, Youn
- Issue Date
- 2023
- Publisher
- IEEE Computer Society
- Keywords
- adversarial attack; deep learning; denoising; fingerprint liveness detection
- Citation
- International Conference on ICT Convergence, pp 1278 - 1282
- Pages
- 5
- Journal Title
- International Conference on ICT Convergence
- Start Page
- 1278
- End Page
- 1282
- URI
- https://scholarworks.bwise.kr/hongik/handle/2020.sw.hongik/32740
- DOI
- 10.1109/ICTC58733.2023.10392544
- ISSN
- 2162-1233
- Abstract
- A number of denoising-based methods have been proposed to defend against adversarial fingerprint attacks. However, these methods inherently rely on having a clean image that corresponds to each adversarial fingerprint image. In this paper, we propose a novel denoising-based defense method without the need for clean fingerprint images. Our approach leverages a Noise2Noise mechanism, which performs denoising based on the noisy dataset. This enables us to effectively eliminate any adversarial noise that may be embedded in fingerprint images without training on clean fingerprint images. The experimental results on real-world datasets confirm that our method is robust against untrained adversarial fingerprint attacks while outperforming existing methods. © 2023 IEEE.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > Computer Engineering > Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/hongik/handle/2020.sw.hongik/32740)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.