Quantifying the Vulnerability of Attributes for Effective Privacy Preservation Using Machine Learningopen access
- Authors
- Majeed, Abdul; Hwang, Seong Oun
- Issue Date
- Jan-2023
- Publisher
- IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
- Keywords
- Data privacy; Information integrity; Information filtering; Codes; Machine learning; Data models; Data science; Personal data; privacy; utility; anonymization; vulnerability; machine learning; responsible data science; data owners; privacy-utility trade-off; imbalanced data; privacy models
- Citation
- IEEE ACCESS, v.11, pp.4400 - 4411
- Journal Title
- IEEE ACCESS
- Volume
- 11
- Start Page
- 4400
- End Page
- 4411
- URI
- https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/87078
- DOI
- 10.1109/ACCESS.2023.3235016
- ISSN
- 2169-3536
- Abstract
- Personal data have been increasingly used in data-driven applications to improve quality of life. However, privacy preservation of personal data while sharing it with analysts/ researchers has become an essential requirement to be met by data owners (hospitals, banks, insurance companies, etc.). The existing literature on privacy preservation does not precisely quantify the vulnerability of each item among user attributes, thereby leading to explicit privacy disclosures and poor data utility during published data analytics. In this work, we propose and implement an automated way of quantifying the vulnerability of each item among the attributes by using a machine learning (ML) technique to significantly preserve the privacy of users without degrading data utility. Our work can solve four technical problems in the privacy preservation field: optimization of the privacy-utility trade-off, privacy guarantees (i.e., safeguard against identity and sensitive information disclosures) in imbalanced data (or clusters), over-anonymization issues, and rectifying or enabling the applicability of prior privacy models when data have skewed distributions. The experiments were performed on two real-world benchmark datasets to prove the feasibility of the concept in practical scenarios. Compared with state-of-the-art (SOTA) methods, the proposed method effectively preserves the equilibrium between utility and privacy in the anonymized data. Furthermore, our method can significantly contribute towards responsible data science (extracting enclosed knowledge from data without violating subjects' privacy) by controlling higher changes in data during its anonymization.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - IT융합대학 > 컴퓨터공학과 > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/87078)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.