Generalized Video Deblurring for Dynamic Scenesopen access
- Authors
- Kim, Tae Hyun; Lee, Kyoung Mu
- Issue Date
- Jun-2015
- Publisher
- IEEE
- Citation
- Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp.5426 - 5434
- Indexed
- SCOPUS
- Journal Title
- Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
- Start Page
- 5426
- End Page
- 5434
- URI
- https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/156958
- ISSN
- 1063-6919
- Abstract
- Several state-of-the-art video deblurring methods are based on a strong assumption that the captured scenes are static. These methods fail to deblur blurry videos in dynamic scenes. We propose a video deblurring method to deal with general blurs inherent in dynamic scenes, contrary to other methods. To handle locally varying and general blurs caused by various sources, such as camera shake, moving objects, and depth variation in a scene, we approximate pixel-wise kernel with bidirectional optical flows. Therefore, we propose a single energy model that simultaneously estimates optical flows and latent frames to solve our deblurring problem. We also provide a framework and efficient solvers to optimize the energy model. By minimizing the proposed energy function, we achieve significant improvements in removing blurs and estimating accurate optical flows in blurry frames. Extensive experimental results demonstrate the superiority of the proposed method in real and challenging videos that state-of-the-art methods fail in either deblurring or optical flow estimation.
- Files in This Item
-
- Appears in
Collections - 서울 공과대학 > 서울 컴퓨터소프트웨어학부 > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/156958)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.