Calibrating Practical Privacy Risks for Differentially Private Machine Learning

dc.contributor.authorGu, Yuechun
dc.contributor.authorChen, Keke
dc.date.accessioned2024-12-11T17:02:05Z
dc.date.available2024-12-11T17:02:05Z
dc.date.issued2024-10-30
dc.description.abstractDifferential privacy quantifies privacy through the privacy budget ϵ, yet its practical interpretation is complicated by variations across models and datasets. Recent research on differentially private machine learning and membership inference has highlighted that with the same theoretical ϵ setting, the likelihood-ratio-based membership inference (LiRA) attacking success rate (ASR) may vary according to specific datasets and models, which might be a better indicator for evaluating real-world privacy risks. Inspired by this practical privacy measure, we study the approaches that can lower the attacking success rate to allow for more flexible privacy budget settings in model training. We find that by selectively suppressing privacy-sensitive features, we can achieve lower ASR values without compromising application-specific data utility. We use the SHAP and LIME model explainer to evaluate feature sensitivities and develop feature-masking strategies. Our findings demonstrate that the LiRA ASRᴹ on model M can properly indicate the inherent privacy risk of a dataset for modeling, and it's possible to modify datasets to enable the use of larger theoretical ϵ settings to achieve equivalent practical privacy protection. We have conducted extensive experiments to show the inherent link between ASR and the dataset's privacy risk. By carefully selecting features to mask, we can preserve more data utility with equivalent practical privacy protection and relaxed ϵ settings. The implementation details are shared online at the provided GitHub URL \url{https://anonymous.4open.science/r/On-sensitive-features-and-empirical-epsilon-lower-bounds-BF67/}.
dc.description.sponsorshipThis work is partially supported by the National Science Foundation (Aware# 2232824).
dc.description.urihttp://arxiv.org/abs/2410.22673
dc.format.extent10 pages
dc.genrejournal articles
dc.genrepreprints
dc.identifierdoi:10.13016/m2flfq-pbos
dc.identifier.urihttps://doi.org/10.48550/arXiv.2410.22673
dc.identifier.urihttp://hdl.handle.net/11603/37025
dc.language.isoen_US
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department
dc.relation.ispartofUMBC Student Collection
dc.relation.ispartofUMBC Faculty Collection
dc.rightsAttribution-NonCommercial-ShareAlike 4.0 International
dc.rights.urihttps://creativecommons.org/licenses/by-nc-sa/4.0/
dc.subjectComputer Science - Cryptography and Security
dc.subjectComputer Science - Machine Learning
dc.titleCalibrating Practical Privacy Risks for Differentially Private Machine Learning
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-9996-156X
dcterms.creatorhttps://orcid.org/0009-0006-4945-7310

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
241022673v1.pdf
Size:
2.3 MB
Format:
Adobe Portable Document Format