TY - GEN
T1 - Differential inference testing
T2 - 2019 IEEE Symposium on Security and Privacy Workshops, SPW 2019
AU - Kassem, Ali
AU - Ács, Gergely
AU - Castelluccia, Claude
AU - Palamidessi, Catuscia
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/5/1
Y1 - 2019/5/1
N2 - In order to protect individuals' privacy, data have to be 'well-sanitized' before sharing them, i.e. one has to remove any personal information before sharing data. However, it is not always clear when data shall be deemed well-sanitized. In this paper, we argue that the evaluation of sanitized data should be based on whether the data allows the inference of sensitive information that is specific to an individual, instead of being centered around the concept of re-identification. We propose a framework to evaluate the effectiveness of different sanitization techniques on a given dataset by measuring how much an individual's record from the sanitized dataset influences the inference of his/her own sensitive attribute. Our intent is not to accurately predict any sensitive attribute but rather to measure the impact of a single record on the inference of sensitive information. We demonstrate our approach by sanitizing two real datasets in different privacy models and evaluate/compare each sanitized dataset in our framework.
AB - In order to protect individuals' privacy, data have to be 'well-sanitized' before sharing them, i.e. one has to remove any personal information before sharing data. However, it is not always clear when data shall be deemed well-sanitized. In this paper, we argue that the evaluation of sanitized data should be based on whether the data allows the inference of sensitive information that is specific to an individual, instead of being centered around the concept of re-identification. We propose a framework to evaluate the effectiveness of different sanitization techniques on a given dataset by measuring how much an individual's record from the sanitized dataset influences the inference of his/her own sensitive attribute. Our intent is not to accurately predict any sensitive attribute but rather to measure the impact of a single record on the inference of sensitive information. We demonstrate our approach by sanitizing two real datasets in different privacy models and evaluate/compare each sanitized dataset in our framework.
KW - Differential Privacy
KW - Inferences
KW - K-Anonymity
KW - Machine Learning
KW - Sanitization
KW - ℓ-Diversity
U2 - 10.1109/SPW.2019.00024
DO - 10.1109/SPW.2019.00024
M3 - Conference contribution
AN - SCOPUS:85073148456
T3 - Proceedings - 2019 IEEE Symposium on Security and Privacy Workshops, SPW 2019
SP - 72
EP - 79
BT - Proceedings - 2019 IEEE Symposium on Security and Privacy Workshops, SPW 2019
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 23 May 2019
ER -