Why Does Differential Privacy with Large Epsilon Defend Against Practical Membership Inference Attacks?

التفاصيل البيبلوغرافية
العنوان: Why Does Differential Privacy with Large Epsilon Defend Against Practical Membership Inference Attacks?
المؤلفون: Lowy, Andrew, Li, Zhuohang, Liu, Jing, Koike-Akino, Toshiaki, Parsons, Kieran, Wang, Ye
سنة النشر: 2024
المجموعة: Computer Science
مصطلحات موضوعية: Computer Science - Cryptography and Security, Computer Science - Artificial Intelligence, Computer Science - Machine Learning, 68P27
الوصف: For small privacy parameter $\epsilon$, $\epsilon$-differential privacy (DP) provides a strong worst-case guarantee that no membership inference attack (MIA) can succeed at determining whether a person's data was used to train a machine learning model. The guarantee of DP is worst-case because: a) it holds even if the attacker already knows the records of all but one person in the data set; and b) it holds uniformly over all data sets. In practical applications, such a worst-case guarantee may be overkill: practical attackers may lack exact knowledge of (nearly all of) the private data, and our data set might be easier to defend, in some sense, than the worst-case data set. Such considerations have motivated the industrial deployment of DP models with large privacy parameter (e.g. $\epsilon \geq 7$), and it has been observed empirically that DP with large $\epsilon$ can successfully defend against state-of-the-art MIAs. Existing DP theory cannot explain these empirical findings: e.g., the theoretical privacy guarantees of $\epsilon \geq 7$ are essentially vacuous. In this paper, we aim to close this gap between theory and practice and understand why a large DP parameter can prevent practical MIAs. To tackle this problem, we propose a new privacy notion called practical membership privacy (PMP). PMP models a practical attacker's uncertainty about the contents of the private data. The PMP parameter has a natural interpretation in terms of the success rate of a practical MIA on a given data set. We quantitatively analyze the PMP parameter of two fundamental DP mechanisms: the exponential mechanism and Gaussian mechanism. Our analysis reveals that a large DP parameter often translates into a much smaller PMP parameter, which guarantees strong privacy against practical MIAs. Using our findings, we offer principled guidance for practitioners in choosing the DP parameter.
Comment: Accepted at PPAI-24: AAAI Workshop on Privacy-Preserving Artificial Intelligence
نوع الوثيقة: Working Paper
URL الوصول: http://arxiv.org/abs/2402.09540
رقم الأكسشن: edsarx.2402.09540
قاعدة البيانات: arXiv