Advanced cover glass defect detection and classification based on multi-DNN model
- Authors
- Park, Jisu; Riaz, Hamza; Kim, Hyunchul; Kim, Jungsuk
- Issue Date
- Jan-2020
- Publisher
- ELSEVIER
- Keywords
- Deep learning; Defect detection; Smart factory display manufacturing
- Citation
- Manufacturing Letters, v.23, pp.53 - 61
- Journal Title
- Manufacturing Letters
- Volume
- 23
- Start Page
- 53
- End Page
- 61
- URI
- https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/17722
- DOI
- 10.1016/j.mfglet.2019.12.006
- ISSN
- 2213-8463
- Abstract
- Demands for display panels and relevant technologies are rapidly increasing with the recent advances in smart mobile devices. Many manufacturers have begun slowly investing in fully automated inspection systems that enable consistent and objective inspections. Carrying out such undertaking aims to satisfy high user requirements concerning quality while coping with high volumes. Cover glass is one of the important items for inspection because users directly interact with it. Despite the extensive use of typical machine vision-based solutions in this field, many manufacturers continue relying on human-based judgment because of a deficient understanding of defects or poor confidence in algorithms. To overcome these problems, this study proposes a deep-learning neural network (DLNN)-based defect inspection system. The DLNN has advantages over traditional computer vision- or human-based inspection in terms of flexibility and performance. We introduce a weighted multi-DLNN inspection system capable of efficiently utilizing multi-channel measurement data, with a detection rate of up to 99% and a false pass rate below 1%. © 2019 Society of Manufacturing Engineers (SME)
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - 보건과학대학 > 의용생체공학과 > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/17722)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.