Private Analysis of Graph Structure With Vishesh Karwa, Sofya Raskhodnikova and Adam Smith Pennsylvania State University Grigory Yaroslavtsev

Slides:



Advertisements
Similar presentations
PRIVACY IN NETWORK TRACES Ilya Mironov Microsoft Research (Silicon Valley Campus)
Advertisements

Grigory Yaroslavtsev Joint work with Piotr Berman and Sofya Raskhodnikova.
DIFFERENTIAL PRIVACY REU Project Mentors: Darakhshan Mir James Abello Marco A. Perez.
Wavelet and Matrix Mechanism CompSci Instructor: Ashwin Machanavajjhala 1Lecture 11 : Fall 12.
Publishing Set-Valued Data via Differential Privacy Rui Chen, Concordia University Noman Mohammed, Concordia University Benjamin C. M. Fung, Concordia.
Differentially Private Recommendation Systems Jeremiah Blocki Fall A: Foundations of Security and Privacy.
SplitX: High-Performance Private Analytics Ruichuan Chen (Bell Labs / Alcatel-Lucent) Istemi Ekin Akkus (MPI-SWS) Paul Francis (MPI-SWS)
Raef Bassily Adam Smith Abhradeep Thakurta Penn State Yahoo! Labs Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds Penn.
1 Differentially Private Analysis of Graphs and Social Networks Sofya Raskhodnikova Pennsylvania State University.
Seminar in Foundations of Privacy 1.Adding Consistency to Differential Privacy 2.Attacks on Anonymized Social Networks Inbal Talgam March 2008.
An brief tour of Differential Privacy Avrim Blum Computer Science Dept Your guide:
1 Preserving Privacy in Collaborative Filtering through Distributed Aggregation of Offline Profiles The 3rd ACM Conference on Recommender Systems, New.
Malicious parties may employ (a) structure-based or (b) label-based attacks to re-identify users and thus learn sensitive information about their rating.
April 13, 2010 Towards Publishing Recommendation Data With Predictive Anonymization Chih-Cheng Chang †, Brian Thompson †, Hui Wang ‡, Danfeng Yao † †‡
2. Attacks on Anonymized Social Networks. Setting A social network Edges may be private –E.g., “communication graph” The study of social structure by.
Privacy in Social Networks:
Foundations of Privacy Lecture 11 Lecturer: Moni Naor.
The Union-Split Algorithm and Cluster-Based Anonymization of Social Networks Brian Thompson Danfeng Yao Rutgers University Dept. of Computer Science Piscataway,
Right Protection via Watermarking with Provable Preservation of Distance-based Mining Spyros Zoumpoulis Joint work with Michalis Vlachos, Nick Freris and.
Differentially Private Data Release for Data Mining Benjamin C.M. Fung Concordia University Montreal, QC, Canada Noman Mohammed Concordia University Montreal,
Differentially Private Transit Data Publication: A Case Study on the Montreal Transportation System Rui Chen, Concordia University Benjamin C. M. Fung,
+ Differential Privacy in the Streaming World Aleksandar (Sasho) Nikolov Rutgers University.
Multiplicative Weights Algorithms CompSci Instructor: Ashwin Machanavajjhala 1Lecture 13 : Fall 12.
Privacy and trust in social network
Private Analysis of Graphs
Defining and Achieving Differential Privacy Cynthia Dwork, Microsoft TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Overview of Privacy Preserving Techniques.  This is a high-level summary of the state-of-the-art privacy preserving techniques and research areas  Focus.
Preserving Link Privacy in Social Network Based Systems Prateek Mittal University of California, Berkeley Charalampos Papamanthou.
Statistical Databases – Query Auditing Li Xiong CS573 Data Privacy and Anonymity Partial slides credit: Vitaly Shmatikov, Univ Texas at Austin.
CS573 Data Privacy and Security Statistical Databases
Differentially Private Data Release for Data Mining Noman Mohammed*, Rui Chen*, Benjamin C. M. Fung*, Philip S. Yu + *Concordia University, Montreal, Canada.
Protecting Sensitive Labels in Social Network Data Anonymization.
Differentially Private Marginals Release with Mutual Consistency and Error Independent of Sample Size Cynthia Dwork, Microsoft TexPoint fonts used in EMF.
Approximation for Directed Spanner Grigory Yaroslavtsev Penn State + AT&T Labs (intern) Based on a paper at ICALP’11, joint with Berman (PSU), Bhattacharyya.
Resisting Structural Re-identification in Anonymized Social Networks Michael Hay, Gerome Miklau, David Jensen, Don Towsley, Philipp Weis University of.
Personalized Social Recommendations – Accurate or Private? A. Machanavajjhala (Yahoo!), with A. Korolova (Stanford), A. Das Sarma (Google) 1.
PRISM: Private Retrieval of the Internet’s Sensitive Metadata Ang ChenAndreas Haeberlen University of Pennsylvania.
Xiaowei Ying, Xintao Wu Dept. Software and Information Systems Univ. of N.C. – Charlotte 2008 SIAM Conference on Data Mining, April 25 th Atlanta, Georgia.
Anonymized Social Networks, Hidden Patterns, and Structural Stenography Lars Backstrom, Cynthia Dwork, Jon Kleinberg WWW 2007 – Best Paper.
Part 3: Query Processing -- Data-Independent Methods 1 Marianne Winslett 1,3, Xiaokui Xiao 2, Yin Yang 3, Zhenjie Zhang 3, Gerome Miklau 4 1 University.
Differential Privacy Some contents are borrowed from Adam Smith’s slides.
Privacy Protection in Social Networks Instructor: Assoc. Prof. Dr. DANG Tran Khanh Present : Bui Tien Duc Lam Van Dai Nguyen Viet Dang.
1 Approximate XML Query Answers Presenter: Hongyu Guo Authors: N. polyzotis, M. Garofalakis, Y. Ioannidis.
Graph Data Management Lab, School of Computer Science Personalized Privacy Protection in Social Networks (VLDB2011)
Differential Privacy Xintao Wu Oct 31, Sanitization approaches Input perturbation –Add noise to data –Generalize data Summary statistics –Means,
Private Release of Graph Statistics using Ladder Functions J.ZHANG, G.CORMODE, M.PROCOPIUC, D.SRIVASTAVA, X.XIAO.
1 Differential Privacy Cynthia Dwork Mamadou H. Diallo.
Privacy Preserving in Social Network Based System PRENTER: YI LIANG.
Yang, et al. Differentially Private Data Publication and Analysis. Tutorial at SIGMOD’12 Part 4: Data Dependent Query Processing Methods Yin “David” Yang.
Space for things we might want to put at the bottom of each slide. Part 6: Open Problems 1 Marianne Winslett 1,3, Xiaokui Xiao 2, Yin Yang 3, Zhenjie Zhang.
Differential Privacy with Bounded Priors: Reconciling Utility and Privacy in Genome-Wide Association Studies Florian Tramèr, Zhicong Huang, Erman Ayday,
Privacy Issues in Graph Data Publishing Summer intern: Qing Zhang (from NC State University) Mentors: Graham Cormode and Divesh Srivastava.
Private Data Management with Verification
Techniques for Achieving Vertex Level Differential Privacy
A paper on Join Synopses for Approximate Query Answering
Understanding Generalization in Adaptive Data Analysis
Privacy-preserving Release of Statistics: Differential Privacy
Graph Analysis with Node Differential Privacy
Dieudo Mulamba November 2017
Differential Privacy in Practice
Vitaly (the West Coast) Feldman
Current Developments in Differential Privacy
Differential Privacy and Statistical Inference: A TCS Perspective
Michael Ernst CSE 140 University of Washington
Binghui Wang, Le Zhang, Neil Zhenqiang Gong
Differentially Private Analysis of Graphs and Social Networks
Published in: IEEE Transactions on Industrial Informatics
Some contents are borrowed from Adam Smith’s slides
Differential Privacy.
Presentation transcript:

Private Analysis of Graph Structure With Vishesh Karwa, Sofya Raskhodnikova and Adam Smith Pennsylvania State University Grigory Yaroslavtsev 1

Publish information about a graph Preserve privacy of relationships Publishing network data Many data sets can be represented as a graph: Friendship in online social network Financial transactions Romantic relationships American J. Sociology, Bearman, Moody, Stovel Naïve approach: anonymization 2

Goal: Publish structural information about a graph Publishing network data Database Relationships Users A queries answers ) ( Government, researchers, businesses (or) Malicious adversary 3 Anonymization not sufficient [Backström, Dwork, Kleinberg ’07, Narayanan, Shmatikov ’09, Narayanan, Shi, Rubinstein ’11] Ideal: Algorithms with rigorous privacy guarantee, no assumptions about attacker’s prior information/algorithm

Limits incremental information by hiding presence/absence of an individual relationship Database Relationships Users A queries answers ) ( Government, researchers, businesses (or) Malicious adversary Differential privacy [Dwork, McSherry, Nissim, Smith ’06] 4 Neighbors: Graphs G and G’ that differ in one edge Answers on neighboring graphs should be similar

Differential privacy for relationships 5 A(G) A(G’) Probability is over the randomness of A Definition requires that the distributions are close:

For graphs G and H: # of occurrences of H in G Subgraph counts Example: Total: 40 Total: 2 Total: 1 Triangle: 2-star: 2-triangle: k-star … k k-triangle k … 6

Subgraph counts are used in: – Exponential random graph models – Descriptive graph statistics, e.g.: Clustering coefficient = Our focus: efficient differentially private algorithms for releasing subgraph counts Subgraph counts # # 7

Previous work Smooth Sensitivity [Nissim, Raskhodnikova, Smith ‘07] – Differentially private algorithm for triangles – Open: private algorithms for other subgraphs? Private queries with joins [Rastogi, Hay, Miklau, Suciu ‘09] – Works for a wide range of subgraphs – Weaker privacy guarantee, applies only for specific class of adversaries Private degree sequence [Hay, Li, Miklau, Jensen ’09] – Guarantees differential privacy – Works for k-stars, but not for other subgraphs 8

Laplace Mechanism and Sensitivity [Dwork, McSherry, Nissim, Smith ‘06] 9

Instance-Specific Noise 10 Smooth Sensitivity [Nissim, Raskhodnikova, Smith ’07]:

Our contributions Differentially private algorithms for k-stars and k-triangles – Efficiently compute smooth sensitivity for k-stars – NP-hardness for k-triangles and k-cycles – Different approach for k-triangles Average-case analysis in G(n,p) Theoretical comparison with previous work Experimental evaluation 11

Smooth Sensitivity for k-stars ( ) 12 …

Private Approximation to Local Sensitivity: k-triangles ( ) 13 …

Evaluating our algorithms 14

Experimental results for G(n,p)

Experimental results (SNAP)

Summary Private algorithms for subgraph counts – Rigorous privacy guarantee (differential privacy) – Running time close to best algorithms for computing the subgraph counts Improvement in accuracy and (for some graph counts) in privacy over previous work Techniques: – Fast computation of smooth sensitivity – Differentially private upper bound on local sensitivity 20