MaskIt: Privately Releasing User Context Streams for Personalized Mobile Applications SIGMOD '12 Proceedings of the 2012 ACM SIGMOD International Conference.

Slides:



Advertisements
Similar presentations
David Rosen Goals  Overview of some of the big ideas in autonomous systems  Theme: Dynamical and stochastic systems lie at the intersection of mathematics.
Advertisements

. Markov Chains. 2 Dependencies along the genome In previous classes we assumed every letter in a sequence is sampled randomly from some distribution.
Hidden Markov Models (HMM) Rabiner’s Paper
Simulatability “The enemy knows the system”, Claude Shannon CompSci Instructor: Ashwin Machanavajjhala 1Lecture 6 : Fall 12.
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
LEARNING INFLUENCE PROBABILITIES IN SOCIAL NETWORKS Amit Goyal Francesco Bonchi Laks V. S. Lakshmanan University of British Columbia Yahoo! Research University.
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Dynamic Bayesian Networks (DBNs)
Modeling Uncertainty over time Time series of snapshot of the world “state” we are interested represented as a set of random variables (RVs) – Observable.
1 Hidden Markov Model Xiaole Shirley Liu STAT115, STAT215, BIO298, BIST520.
Automatic Speech Recognition II  Hidden Markov Models  Neural Network.
Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data John Lafferty Andrew McCallum Fernando Pereira.
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
Introduction to Hidden Markov Models
Hidden Markov Models Eine Einführung.
Hidden Markov Models Ellen Walker Bioinformatics Hiram College, 2008.
Ch 9. Markov Models 고려대학교 자연어처리연구실 한 경 수
Hidden Markov Models Theory By Johan Walters (SR 2003)
Hidden Markov Model Cryptanalysis Chris Karlof and David Wagner.
Hidden Markov Models Fundamentals and applications to bioinformatics.
Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –One exception: games with multiple moves In particular, the Bayesian.
… Hidden Markov Models Markov assumption: Transition model:
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Lecture 5: Learning models using EM
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Hidden Markov Models David Meir Blei November 1, 1999.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
Information Theory and Security
Combined Lecture CS621: Artificial Intelligence (lecture 25) CS626/449: Speech-NLP-Web/Topics-in- AI (lecture 26) Pushpak Bhattacharyya Computer Science.
CS344 : Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 21- Forward Probabilities and Robotic Action Sequences.
THE HIDDEN MARKOV MODEL (HMM)
Segmental Hidden Markov Models with Random Effects for Waveform Modeling Author: Seyoung Kim & Padhraic Smyth Presentor: Lu Ren.
Fundamentals of Hidden Markov Model Mehmet Yunus Dönmez.
Data Publishing against Realistic Adversaries Johannes Gerhrke Cornell University Ithaca, NY Michaela Götz Cornell University Ithaca, NY Ashwin Machanavajjhala.
Thwarting Passive Privacy Attacks in Collaborative Filtering Rui Chen Min Xie Laks V.S. Lakshmanan HKBU, Hong Kong UBC, Canada UBC, Canada Introduction.
Hidden Markov Models Yves Moreau Katholieke Universiteit Leuven.
Hidden Markov Models in Keystroke Dynamics Md Liakat Ali, John V. Monaco, and Charles C. Tappert Seidenberg School of CSIS, Pace University, White Plains,
CS774. Markov Random Field : Theory and Application Lecture 19 Kyomin Jung KAIST Nov
1 MANETS – An overview. 2 MANETs Model and Definitions Simulatability – mobility models Routing algorithms Security issues with routing algorithms Tracing.
Additive Data Perturbation: the Basic Problem and Techniques.
Markov Models and Simulations Yu Meng Department of Computer Science and Engineering Southern Methodist University.
Meng-Han Yang September 9, 2009 A sequence-based hybrid predictor for identifying conformationally ambivalent regions in proteins.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
CSC321: Neural Networks Lecture 16: Hidden Markov Models
Hidden Markovian Model. Some Definitions Finite automation is defined by a set of states, and a set of transitions between states that are taken based.
John Lafferty Andrew McCallum Fernando Pereira
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Discovering Evolutionary Theme Patterns from Text - An Exploration of Temporal Text Mining Qiaozhu Mei and ChengXiang Zhai Department of Computer Science.
Hidden Markov Models (HMMs) –probabilistic models for learning patterns in sequences (e.g. DNA, speech, weather, cards...) (2 nd order model)
1 Hidden Markov Models Hsin-min Wang References: 1.L. R. Rabiner and B. H. Juang, (1993) Fundamentals of Speech Recognition, Chapter.
Differential Privacy Xintao Wu Oct 31, Sanitization approaches Input perturbation –Add noise to data –Generalize data Summary statistics –Means,
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Hidden Markov Model Parameter Estimation BMI/CS 576 Colin Dewey Fall 2015.
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Graphical Models for Segmenting and Labeling Sequence Data Manoj Kumar Chinnakotla NLP-AI Seminar.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
1 Hidden Markov Model Xiaole Shirley Liu STAT115, STAT215.
PROBABILITY AND COMPUTING RANDOMIZED ALGORITHMS AND PROBABILISTIC ANALYSIS CHAPTER 1 IWAMA and ITO Lab. M1 Sakaidani Hikaru 1.
Hidden Markov Models BMI/CS 576
Markov ó Kalman Filter Localization
Hidden Markov Autoregressive Models
Uncertainty in AI.
Hidden Markov Model LR Rabiner
Hidden Markov Models By Manish Shrivastava.
Presentation transcript:

MaskIt: Privately Releasing User Context Streams for Personalized Mobile Applications SIGMOD '12 Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data

Background  Not just location.  More sensors means more privacy can be detected.

Background  User’s contexts have correlation with former context.  Some contexts are not sensitive.

Solution  System Model  x 1, x 2,…, x t o 1, o 2,…, o t  To compute o t, MASKIT employs a check deciding whether to release or suppress the current context.

Solution  Propose MASKIT: a system that decide whether to release or to suppress the current state of the use.  Probabilistic check: flips for each context a coin is chosen suitably to guarantee privacy.  Simulatable check: makes the decision only based on the released contexts so far and completely ignores the current context.  Explain how to select the better check.

Problem Statement  Utility Goal  Release as many states as possible, while satisfying the privacy goal.  The MASKIT System

Problem  What is privacy?  To preserve privacy:  When context should be suppressed?  What context should be suppressed?  How to ensure utility?

Privacy  Privacy  DEFINITION 1: We say that a system A preserves -privacy against an adversary if for all possible inputs sampled from the Markov chain M with non-zero probability, for all possible outputs, for all times t and all sensitive contexts

Utility  Measure utility as the expected number of released context: for state c i at time t’ a suppression probability

Probabilistic Privacy Check

 Prior belief:  suppression probability p i t at time t for state c i, the prior belief is 1- p i t  Posterior belief:  HMM Forward procedure Backward procedure

Probabilistic Privacy Check  Utility :  For vectors passing the check we can compute their utility  Return the one with the maximum utility  Efficiency  Use algorithms to speeding up IsPrivate & SearchAlgorithm

Simulatable Privacy Check  Only based on information available to the adversary:  The Markov chain M  Output sequence  Posterior belief  t1: last time before or at t at which a context was released  t2: earliest time after t at which a context was released  t2: end state if t2 does not exist

Simulatable Privacy Check

 Privacy :

Simulatable Privacy Check  Utility:  The simulatable check is locally optimal in the sense that if the next state is published despite the indication of the privacy check to suppress it (improving the utility) then there is a chance that future states will inevitably breach privacy.  Efficiency  Speeding up okayToRelease

Comparative Analysis Weakness of the simulatable check: It makes the suppression decision without looking at the current state. Weakness of the probabilistic check: Its decision ignores the previously released states.

Comparative Analysis  Hybrid Privacy Check:  Probabilistic check  Simulatable check Using supp i (t) we can compute recursively the expected number of suppressions following the release of X t = c i utility Simulatable (M) = T - expected number

Limited Background Knowledge  Weak adversary:  Knowing the Frequency of sensitive contexts  Knowing a Set-Labeled chain

Experiment  Continuous data on daily activities of 100 students and staff at MIT.  For each user, we train a Markov chain on the first half of his trace; the remaining half is used to for evaluation.  This paper only use location data.

Experiment  Efficiency

Experiments  Compare MASKIT using:  The simulatable check  The probabilistic check (with a granularity of d = 10)  The hybrid check with the naive approach, called MaskSensitive

Experiments

Thank you!

Problem Statement  User Model  User’s behaves like a sample from a Markov chain M  The states in M are labeled with contexts {c 1,…,c n }  Each day, the user starts at the “start” state in M and ends T steps later in the “end” state  X 1,…,X T : random variables generated from M, each taking on the value of some context c i  The independence property of Markov chains states that

Problem Statement  Adversary Model  Strong Adversary: know the Markov chain M of a user.  Week Adversary: have less knowledge about M,but they can learn more about M over time.  can access the full output sequence generated by a general suppression system A, and we assume the adversaries also know A.  adversaries have a prior belief about the user being in context ci at time t.

Problem Statement  Preliminaries: Markov chains  Markovian process with transition matrices A (1),…, A (T+1) :  PROPOSITION 1: The prior belief of an adversary about the user being in a sensitive context s at time t is equal to  The joint probability of a sequence of states is:  The probability of transitioning from state c i at time t 1 to state c j at time t 2 e i is the unit vector that is 1 at position i and 0 otherwise

Problem Statement  Preliminaries: Hidden Markov Models  Hidden Markov models help us understand how adversaries make inference about suppressed states.  Each state has a distribution over possible outputs from a set K = {k 1,…,k m }.  Define emission matrices B(t) as:  For a given output sequence, we compute the conditional probability that at tine t the hidden state was c i :

Problem Statement  Preliminaries: Hidden Markov Models  Use the forward procedure and the backward procedure to compute this ratio efficiently :  Initialize  Initialize, put everything together: