Publications from our Researchers

Several of our current PhD candidates and fellow researchers at the Data Science Institute have published, or in the proccess of publishing, papers to present their research.  

Citation

BibTex format

@inproceedings{Gadotti:2019,
author = {Gadotti, A and Houssiau, F and Rocher, L and Livshits, B and de, Montjoye Y-A},
pages = {1081--1098},
publisher = {USENIX},
title = {When the signal is in the noise: exploiting Diffix's sticky noise},
url = {https://www.usenix.org/conference/usenixsecurity19/technical-sessions},
year = {2019}
}

RIS format (EndNote, RefMan)

TY  - CPAPER
AB - Anonymized data is highly valuable to both businesses and researchers. A large body of research has however shown the strong limits of the de-identification release-and-forget model, where data is anonymized and shared. This has led to the development of privacy-preserving query-based systems. Based on the idea of “sticky noise”, Diffix has been recently pro-posed as a novel query-based mechanism satisfying alone the EU Article 29 Working Party’s definition of anonymization. According to its authors, Diffix adds less noise to answers than solutions based on differential privacy while allowing for an unlimited number of queries.This paper presents a new class of noise-exploitation attacks, exploiting the noise added by the system to infer privateinformation about individuals in the dataset. Our first differential attack uses samples extracted from Diffix in a likelihood ratio test to discriminate between two probability distributions.We show that using this attack against a synthetic best-case dataset allows us to infer private information with 89.4% accuracy using only 5 attributes. Our second cloning attack uses dummy conditions that conditionally strongly affect the output of the query depending on the value of the private attribute. Using this attack on four real-world datasets, we show that we can infer private attributes of at least 93% of the users in the dataset with accuracy between 93.3% and 97.1%, issuing a median of 304 queries per user. We show how to optimize this attack, targeting 55.4% of the users and achieving 91.7% accuracy, using a maximum of only 32 queries per user. Our attacks demonstrate that adding data-dependent noise, as done by Diffix, is not sufficient to prevent inference of private attributes. We furthermore argue that Diffix alone fails to satisfy Art. 29 WP’s definition of anonymization. We conclude by discussing how non-provable privacy-preserving systems can be combined with fundamental security principles su
AU - Gadotti,A
AU - Houssiau,F
AU - Rocher,L
AU - Livshits,B
AU - de,Montjoye Y-A
EP - 1098
PB - USENIX
PY - 2019///
SP - 1081
TI - When the signal is in the noise: exploiting Diffix's sticky noise
UR - https://www.usenix.org/conference/usenixsecurity19/technical-sessions
UR - http://hdl.handle.net/10044/1/69958
ER -

Contact us

Data Science Institute

William Penney Laboratory
Imperial College London
South Kensington Campus
London SW7 2AZ
United Kingdom

Email us.

Sign up to our mailing list.

Follow us on Twitter, LinkedIn and Instagram.