Walking in the Crowd: Anonymizing Trajectory Data for Pattern Analysis

Slides:



Advertisements
Similar presentations
Wang, Lakshmanan Probabilistic Privacy Analysis of Published Views, WPES'06 Probabilistic Privacy Analysis of Published Views Hui (Wendy) Wang Laks V.S.
Advertisements

Anonymity for Continuous Data Publishing
Privacy-preserving of Trajectory Data : A Survey
Association Analysis (2). Example TIDList of item ID’s T1I1, I2, I5 T2I2, I4 T3I2, I3 T4I1, I2, I4 T5I1, I3 T6I2, I3 T7I1, I3 T8I1, I2, I3, I5 T9I1, I2,
Publishing Set-Valued Data via Differential Privacy Rui Chen, Concordia University Noman Mohammed, Concordia University Benjamin C. M. Fung, Concordia.
Differentially Private Transit Data Publication: A Case Study on the Montreal Transportation System ` Introduction With the deployment of smart card automated.
PREFIXSPAN ALGORITHM Mining Sequential Patterns Efficiently by Prefix- Projected Pattern Growth
Center for Secure Information Systems Concordia Institute for Information Systems Engineering k-Jump Strategy for Preserving Privacy in Micro-Data Disclosure.
Computer Security Lab Concordia Institute for Information Systems Engineering Concordia University Montreal, Canada A Novel Approach of Mining Write-Prints.
Mining Multiple-level Association Rules in Large Databases
Wang, Lakshmanan Probabilistic Privacy Analysis of Published Views, IDAR'07 Probabilistic Privacy Analysis of Published Views Hui (Wendy) Wang Laks V.S.
Anonymizing Location-based data Jarmanjit Singh Jar_sing(at)encs.concordia.ca Harpreet Sandhu h_san(at)encs.concordia.ca Qing Shi q_shi(at)encs.concordia.ca.
Hani AbuSharkh Benjamin C. M. Fung fung (at) ciise.concordia.ca
Secure Distributed Framework for Achieving -Differential Privacy Dima Alhadidi, Noman Mohammed, Benjamin C. M. Fung, and Mourad Debbabi Concordia Institute.
Template-Based Privacy Preservation in Classification Problems IEEE ICDM 2005 Benjamin C. M. Fung Simon Fraser University BC, Canada Ke.
Anonymizing Healthcare Data: A Case Study on the Blood Transfusion Service Benjamin C.M. Fung Concordia University Montreal, QC, Canada
Privacy-Preserving Data Mashup Benjamin C.M. Fung Concordia University Montreal, QC, Canada Noman Mohammed Concordia University.
PRIVACY AND SECURITY ISSUES IN DATA MINING P.h.D. Candidate: Anna Monreale Supervisors Prof. Dino Pedreschi Dott.ssa Fosca Giannotti University of Pisa.
1 Privacy Preserving Data Publishing Prof. Ravi Sandhu Executive Director and Endowed Chair March 29, © Ravi.
Data Mining and Intrusion Detection
Privacy Preserving Publication of Moving Object Data Joey Lei CS295 Francesco Bonchi Yahoo! Research Avinguda Diagonal 177, Barcelona, Spain 6/10/20151CS295.
K Beyond k-Anonimity: A Decision Theoretic Framework for Assessing Privacy Risk M.Scannapieco, G.Lebanon, M.R.Fouad and E.Bertino.
Temporal Pattern Matching of Moving Objects for Location-Based Service GDM Ronald Treur14 October 2003.
Anatomy: Simple and Effective Privacy Preservation Israel Chernyak DB Seminar (winter 2009)
2/8/00CSE 711 data mining: Apriori Algorithm by S. Cha 1 CSE 711 Seminar on Data Mining: Apriori Algorithm By Sung-Hyuk Cha.
Privacy Preserving Data Mining: An Overview and Examination of Euclidean Distance Preserving Data Transformation Chris Giannella cgiannel AT acm DOT org.
ACM SIGKDD Aug – Washington, DC  M. El-Hajj and O. R. Zaïane, 2003 Database Lab. University of Alberta Canada Inverted Matrix: Efficient Discovery.
Privacy Protection for RFID Data Benjamin C.M. Fung Concordia Institute for Information systems Engineering Concordia university Montreal, QC, Canada
Differentially Private Data Release for Data Mining Benjamin C.M. Fung Concordia University Montreal, QC, Canada Noman Mohammed Concordia University Montreal,
Task 1: Privacy Preserving Genomic Data Sharing Presented by Noman Mohammed School of Computer Science McGill University 24 March 2014.
Differentially Private Transit Data Publication: A Case Study on the Montreal Transportation System Rui Chen, Concordia University Benjamin C. M. Fung,
USpan: An Efficient Algorithm for Mining High Utility Sequential Patterns Authors: Junfu Yin, Zhigang Zheng, Longbing Cao In: Proceedings of the 18th ACM.
Overview of Privacy Preserving Techniques.  This is a high-level summary of the state-of-the-art privacy preserving techniques and research areas  Focus.
Trajectory Pattern Mining
Differentially Private Data Release for Data Mining Noman Mohammed*, Rui Chen*, Benjamin C. M. Fung*, Philip S. Yu + *Concordia University, Montreal, Canada.
Thwarting Passive Privacy Attacks in Collaborative Filtering Rui Chen Min Xie Laks V.S. Lakshmanan HKBU, Hong Kong UBC, Canada UBC, Canada Introduction.
CS525 DATA MINING COURSE INTRODUCTION YÜCEL SAYGIN SABANCI UNIVERSITY.
SFU Pushing Sensitive Transactions for Itemset Utility (IEEE ICDM 2008) Presenter: Yabo, Xu Authors: Yabo Xu, Benjam C.M. Fung, Ke Wang, Ada. W.C. Fu,
SECURED OUTSOURCING OF FREQUENT ITEMSET MINING Hana Chih-Hua Tai Dept. of CSIE, National Taipei University.
Security Control Methods for Statistical Database Li Xiong CS573 Data Privacy and Security.
Hybrid l-Diversity* Mehmet Ercan NergizMuhammed Zahit GökUfuk Özkanlı
TEMPLATE DESIGN © Predicate-Tree based Pretty Good Protection of Data William Perrizo, Arjun G. Roy Department of Computer.
Privacy vs. Utility Xintao Wu University of North Carolina at Charlotte Nov 10, 2008.
Intelligent DataBase System Lab, NCKU, Taiwan Josh Jia-Ching Ying 1, Wang-Chien Lee 2, Tz-Chiao Weng 1 and Vincent S. Tseng 1 1 Department of Computer.
Privacy-preserving data publishing
1/3/ A Framework for Privacy- Preserving Cluster Analysis IEEE ISI 2008 Benjamin C. M. Fung Concordia University Canada Lingyu.
Anonymizing Data with Quasi-Sensitive Attribute Values Pu Shi 1, Li Xiong 1, Benjamin C. M. Fung 2 1 Departmen of Mathematics and Computer Science, Emory.
Probabilistic km-anonymity (Efficient Anonymization of Large Set-valued Datasets) Gergely Acs (INRIA) Jagdish Achara (INRIA)
Mining Sequential Patterns © Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 Slides are adapted from Introduction to Data Mining by Tan, Steinbach,
1 Top Down FP-Growth for Association Rule Mining By Ke Wang.
Deriving Private Information from Association Rule Mining Results Zutao Zhu, Guan Wang, and Wenliang Du ICDE /3/181.
ACHIEVING k-ANONYMITY PRIVACY PROTECTION USING GENERALIZATION AND SUPPRESSION International Journal on Uncertainty, Fuzziness and Knowledge-based Systems,
Xiaokui Xiao and Yufei Tao Chinese University of Hong Kong
Sequential Pattern Mining Using A Bitmap Representation
Frequent Pattern Mining
Personalized Privacy Protection in Social Networks
Privacy Preserving Data Publishing
CARPENTER Find Closed Patterns in Long Biological Datasets
By (Group 17) Mahesha Yelluru Rao Surabhee Sinha Deep Vakharia
Mining Sequential Patterns
Towards the Automated Recovery of Complex Temporal API-Usage Patterns
Amer Zaheer PC Mohammad Ali Jinnah University, Islamabad
Personalized Privacy Protection in Social Networks
Mining Sequential Patterns
Frequent-Pattern Tree
Stratified Sampling for Data Mining on the Deep Web
Privacy preserving cloud computing
Presented by : SaiVenkatanikhil Nimmagadda
Refined privacy models
Presentation transcript:

Walking in the Crowd: Anonymizing Trajectory Data for Pattern Analysis Noman Mohammed, Benjamin C. M. Fung, Mourad Debbabi Concordia University, Montreal, Canada Introduction Anonymization Algorithm Empirical Study Privacy Model: Let L be the maximum length of the background knowledge. Let S be a set of sensitive values. A trajectory database T satisfies LKC- privacy if and only if for any sequence q with |q|<=L, |T(q)|>=K, where K > 0 is an integer anonymity threshold, and P(s|T(q))<=C, where 0 < C <=1 is a real number confidence threshold. Utility Measure: We aim at preserving the maximal frequent sequences (MFS) because MFS often serves as the information basis for different primitive data mining tasks on sequential data, such as trajectory pattern mining [4]. Identifying Violating Sequences: Any non-empty sequence q with |q|<=L in T is a violating sequence if its group T(q) does not satisfy both the conditions of LKC-privacy. We enumerate all the minimal violating sequences(MVS) by our MVS-generator algorithm. Eliminating Violating Sequences: Figure 1 depicts both an MVS-tree and an MFS-tree generated from Table 1, where minimal violating sequences, V (T) = {b2d3;b2c4; b2f6; c4c7; c4e8} and maximal frequent sequences, U(T) = {b2c5c7; b2f6c7; b2c7e8; d3c4f6; f6c7e8; c5f6; c5e8; d3c7; d3e8} with L = 2, K = 2,C = 50%, and K’ = 2. Recently, trajectory data mining has received a lot of attention in both the industry and the academic research. Publication of these trajectories for data analysis purposes threatens individuals' privacy since these raw trajectory data provide location information that identifies individuals and, potentially, their sensitive information. Example: Privacy threats of publishing trajectory data Identity linkage: ID#1 is the only record that contains b2 and d3. Sensitive value (AIDS in this case) is easily identifiable. Attribute linkage: Two out of the three records (ID#1,7,8) containing b2 and f6 have sensitive value AIDS, the adversary can infer that the victim has AIDS with 2/3 = 67% confidence. Conclusions We proposed a new LKC-privacy model based on the assumption that an adversary has limited background knowledge about the victim. We also presented an efficient algorithm for achieving LKC-privacy with the goal of preserving maximal frequent sequences, which serves as the basis of many data mining tasks on trajectory data. References Figure 1: MVS-tree and MFS-tree for efficient Score updates 1.O. Abul, F. Bonchi, and M. Nanni. Never walk alone: Uncertainty for anonymity in moving objects databases. In ICDE, 2008. 2.M. Terrovitis and N. Mamoulis. Privacy preservation in the publication of trajectories. In MDM, 2008. 3.R. Yarovoy, F. Bonchi, L. V. S. Lakshmanan, and W. H. Wang. Anonymizing moving objects: How to hide a MOB in a crowd? In EDBT, 2009. 4.F. Giannotti, M. Nanni, D. Pedreschi, and F. Pinelli. Trajectory pattern mining. In ACM SIGKDD, 2007. Table 2: Initial Score