Rotation-Invariant Wafer Map Pattern Classification With Convolutional Neural Networks
- Authors
- Kang, S[Kang, Seokho]
- Issue Date
- 2020
- Publisher
- IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
- Keywords
- Predictive models; Semiconductor device modeling; Feature extraction; Task analysis; Data models; Training; Training data; Wafer map pattern classification; convolutional neural network; data augmentation; rotational invariance
- Citation
- IEEE ACCESS, v.8, pp.170650 - 170658
- Indexed
- SCIE
SCOPUS
- Journal Title
- IEEE ACCESS
- Volume
- 8
- Start Page
- 170650
- End Page
- 170658
- URI
- https://scholarworks.bwise.kr/skku/handle/2021.sw.skku/7154
- DOI
- 10.1109/ACCESS.2020.3024603
- ISSN
- 2169-3536
- Abstract
- The enhancement of production yield is a continuous challenge in semiconductor manufacturing. Analyzing the spatial defect patterns of previously processed wafers is a key step in identifying the root causes of yield degradation. Predictive modeling approaches have been successful in automated wafer map pattern classification. The classification performance depends significantly on the quantity and diversity of data that can be acquired, which are often limited in practice. In this study, we demonstrate that rotation-based data augmentation can effectively improve wafer map pattern classification when training data are scarce. As rotation is a label-preserving transformation for wafer maps, we construct a convolutional neural network with rotation-augmented training data to render the classification invariant with respect to rotation. This enables us to provide consistent predictions for rotational variations of new wafer maps, thereby achieving higher classification performance. The effectiveness of our method is verified based on real-word data from a semiconductor manufacturer.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Engineering > Department of Systems Management Engineering > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.