Unifying Privacy Measures via Maximal (α,β)-Leakage (MαbeL)
arxiv(2023)
摘要
We introduce a family of information leakage measures called maximal
(α,β)-leakage (MαbeL), parameterized by real numbers α
and β greater than or equal to 1. The measure is formalized via an
operational definition involving an adversary guessing an unknown (randomized)
function of the data given the released data. We obtain a simplified computable
expression for the measure and show that it satisfies several basic properties
such as monotonicity in β for a fixed α, non-negativity, data
processing inequalities, and additivity over independent releases. We highlight
the relevance of this family by showing that it bridges several known leakage
measures, including maximal α-leakage (β=1), maximal leakage
(α=∞,β=1), local differential privacy (LDP)
(α=∞,β=∞), and local Renyi differential privacy (LRDP)
(α=β), thereby giving an operational interpretation to local Renyi
differential privacy. We also study a conditional version of MαbeL on
leveraging which we recover differential privacy and Renyi differential
privacy. A new variant of LRDP, which we call maximal Renyi leakage, appears as
a special case of MαbeL for α=∞ that smoothly tunes between
maximal leakage (β=1) and LDP (β=∞). Finally, we show that a
vector form of the maximal Renyi leakage relaxes differential privacy under
Gaussian and Laplacian mechanisms.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要