TY - GEN
T1 - Privacy-MaxEnt
T2 - 2008 ACM SIGMOD International Conference on Management of Data 2008, SIGMOD'08
AU - Du, Wenliang
AU - Teng, Zhouxuan
AU - Zhu, Zutao
PY - 2008
Y1 - 2008
N2 - Privacy-Preserving Data Publishing (PPDP) deals with the publication of microdata while preserving people' private information in the data. To measure how much private information can be preserved, privacy metrics is needed. An essential element for privacy metrics is the measure of how much adversaries can know about an individual's sensitive attributes (SA) if they know the individual's quasi-identifters (QI), i.e., we need to measure P(SA/QI). Such a measure is hard to derive when adversaries' background knowledge has to be considered. We propose a systematic approach, Privacy-MaxEnt, to integrate background knowledge in privacy quantification. Our approach is based on the maximum entropy principle. We treat all the conditional probabilities P(SA | QI) as unknown variables; we treat the background knowledge as the constraints of these variables; in addition, we also formulate constraints from the published data. Our goal becomes finding a solution to those variables (the probabilities) that satisfy all these constraints. Although many solutions may exist, the most unbiased estimate of P(SA | QI) is the one that achieves the maximum entropy.
AB - Privacy-Preserving Data Publishing (PPDP) deals with the publication of microdata while preserving people' private information in the data. To measure how much private information can be preserved, privacy metrics is needed. An essential element for privacy metrics is the measure of how much adversaries can know about an individual's sensitive attributes (SA) if they know the individual's quasi-identifters (QI), i.e., we need to measure P(SA/QI). Such a measure is hard to derive when adversaries' background knowledge has to be considered. We propose a systematic approach, Privacy-MaxEnt, to integrate background knowledge in privacy quantification. Our approach is based on the maximum entropy principle. We treat all the conditional probabilities P(SA | QI) as unknown variables; we treat the background knowledge as the constraints of these variables; in addition, we also formulate constraints from the published data. Our goal becomes finding a solution to those variables (the probabilities) that satisfy all these constraints. Although many solutions may exist, the most unbiased estimate of P(SA | QI) is the one that achieves the maximum entropy.
KW - Data publishing
KW - Privacy quantification
UR - http://www.scopus.com/inward/record.url?scp=57149146157&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=57149146157&partnerID=8YFLogxK
U2 - 10.1145/1376616.1376665
DO - 10.1145/1376616.1376665
M3 - Conference contribution
AN - SCOPUS:57149146157
SN - 9781605581026
T3 - Proceedings of the ACM SIGMOD International Conference on Management of Data
SP - 459
EP - 472
BT - SIGMOD 2008
Y2 - 9 June 2008 through 12 June 2008
ER -