Active Sampling of Networks Joseph J. Pfeiffer III 1 Jennifer Neville 1 Paul N. Bennett 2 Purdue University 1 Microsoft Research 2 July 1, 2012 MLG, Edinburgh.

Slides:



Advertisements
Similar presentations
Sparsification and Sampling of Networks for Collective Classification
Advertisements

Classification. Introduction A discriminant is a function that separates the examples of different classes. For example – IF (income > Q1 and saving >Q2)
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Multi-label Relational Neighbor Classification using Social Context Features Xi Wang and Gita Sukthankar Department of EECS University of Central Florida.
Missing values problem in Data Mining
Combining Information from Related Regressions Duke University Machine Learning Group Presented by Kai Ni Apr. 27, 2007 F. Dominici, G. Parmigiani, K.
Patch to the Future: Unsupervised Visual Prediction
Monte Carlo Localization for Mobile Robots Karan M. Gupta 03/10/2004
Autocorrelation and Linkage Cause Bias in Evaluation of Relational Learners David Jensen and Jennifer Neville.
Relational Learning with Gaussian Processes By Wei Chu, Vikas Sindhwani, Zoubin Ghahramani, S.Sathiya Keerthi (Columbia, Chicago, Cambridge, Yahoo!) Presented.
Learning using Graph Mincuts Shuchi Chawla Carnegie Mellon University 1/11/2003.
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Probabilistic Paths and Centrality in Time Joseph J. Pfeiffer, III Jennifer Neville.
Using Structure Indices for Efficient Approximation of Network Properties Matthew J. Rattigan, Marc Maier, and David Jensen University of Massachusetts.
Lesson learnt from the UCSD datamining contest Richard Sia 2008/10/10.
Server-based Inference of Internet Performance V. N. Padmanabhan, L. Qiu, and H. Wang.
Decision Tree Algorithm
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Semi-Supervised Learning Using Randomized Mincuts Avrim Blum, John Lafferty, Raja Reddy, Mugizi Rwebangira.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Thanks to Nir Friedman, HU
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
1 Efficiently Learning the Accuracy of Labeling Sources for Selective Sampling by Pinar Donmez, Jaime Carbonell, Jeff Schneider School of Computer Science,
Determining the Significance of Item Order In Randomized Problem Sets Zachary A. Pardos, Neil T. Heffernan Worcester Polytechnic Institute Department of.
Active Learning for Networked Data Based on Non-progressive Diffusion Model Zhilin Yang, Jie Tang, Bin Xu, Chunxiao Xing Dept. of Computer Science and.
Fall 2004 TDIDT Learning CS478 - Machine Learning.
Modeling Relationship Strength in Online Social Networks Rongjing Xiang: Purdue University Jennifer Neville: Purdue University Monica Rogati: LinkedIn.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Populations, Samples, Standard errors, confidence intervals Dr. Omar Al Jadaan.
Using Transactional Information to Predict Link Strength in Online Social Networks Indika Kahanda and Jennifer Neville Purdue University.
Page 1 Ming Ji Department of Computer Science University of Illinois at Urbana-Champaign.
Active Learning on Spatial Data Christine Körner Fraunhofer AIS, Uni Bonn.
1 Passive Network Tomography Using Bayesian Inference Lili Qiu Joint work with Venkata N. Padmanabhan and Helen J. Wang Microsoft Research Internet Measurement.
Style & Topic Language Model Adaptation Using HMM-LDA Bo-June (Paul) Hsu, James Glass.
On Node Classification in Dynamic Content-based Networks.
Xu Huaping, Wang Wei, Liu Xianghua Beihang University, China.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
I. Science is not A collection of never-changing facts or beliefs about the world.
Problem Limited number of experimental replications. Postgenomic data intrinsically noisy. Poor network reconstruction.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 12 Testing for Relationships Tests of linear relationships –Correlation 2 continuous.
Major transport accidents in Norway: assessing long-term frequency and priorities for prevention TRB paper Rune Elvik.
Chapter 7. Learning through Imitation and Exploration: Towards Humanoid Robots that Learn from Humans in Creating Brain-like Intelligence. Course: Robots.
Lecture 1 Stat Applications, Types of Data And Statistical Inference.
2002 Spring Data Mining Term Project Proposal Data Mining Experimental Study with Oracle & MS SQL 012ITI12 Song Mi-Kyoung.
MCMC in structure space MCMC in order space.
1 Network Tomography Using Passive End-to-End Measurements Venkata N. Padmanabhan Lili Qiu Helen J. Wang Microsoft Research DIMACS’2002.
Statistics for Engineer. Statistics  Deals with  Collection  Presentation  Analysis and use of data to make decision  Solve problems and design.
Short Introduction to Particle Filtering by Arthur Pece [ follows my Introduction to Kalman filtering ]
MPS/MSc in StatisticsAdaptive & Bayesian - Lect 71 Lecture 7 Bayesian methods: a refresher 7.1 Principles of the Bayesian approach 7.2 The beta distribution.
1 Optimizing Decisions over the Long-term in the Presence of Uncertain Response Edward Kambour.
Bayesian Speech Synthesis Framework Integrating Training and Synthesis Processes Kei Hashimoto, Yoshihiko Nankaku, and Keiichi Tokuda Nagoya Institute.
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Bayesian Approach Jake Blanchard Fall Introduction This is a methodology for combining observed data with expert judgment Treats all parameters.
Multi-label Prediction via Sparse Infinite CCA Piyush Rai and Hal Daume III NIPS 2009 Presented by Lingbo Li ECE, Duke University July 16th, 2010 Note:
1 Guess the Covered Word Goal 1 EOC Review 2 Scientific Method A process that guides the search for answers to a question.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Bo Zong, Yinghui Wu, Ambuj K. Singh, Xifeng Yan 1 Inferring the Underlying Structure of Information Cascades
1 Network Tomography Using Passive End-to-End Measurements Lili Qiu Joint work with Venkata N. Padmanabhan and Helen J. Wang.
Bayesian Optimization. Problem Formulation Goal  Discover the X that maximizes Y  Global optimization Active experimentation  We can choose which values.
Sofus A. Macskassy Fetch Technologies
Generalization ..
The Nature of Probability and Statistics
Markov Random Fields for Edge Classification
Multitask Learning Using Dirichlet Process
Hierarchical Topic Models and the Nested Chinese Restaurant Process
Graph-based Security and Privacy Analytics via Collective Classification with Joint Weight Learning and Propagation Binghui Wang, Jinyuan Jia, and Neil.
Maintenance Sheet Due Wednesday
Sofia Pediaditaki and Mahesh Marina University of Edinburgh
Maintenance Sheet Due Wednesday
Maintenance Sheet Due Wednesday
Presentation transcript:

Active Sampling of Networks Joseph J. Pfeiffer III 1 Jennifer Neville 1 Paul N. Bennett 2 Purdue University 1 Microsoft Research 2 July 1, 2012 MLG, Edinburgh

Population

Population - Labels

Underlying Social Network

Population – No Labels, No Edges

Active Sampling

Node Subsets – Labeled Nodes – Border Nodes – Separate Nodes Acquire Positive instances into Labeled set – Minimize acquisitions Labeled set used to estimate Border set – Network structure should improve estimates Choose node(s) to investigate from Border and Separate sets Active Sampling

Estimating Border Likelihoods weighted vote Relational Neighbor 1 (wvRN) –Utilize only known edges Utilize collective inference usefully? 1 Macskassy & Provost, 2007

Estimating Border Likelihoods – Collective Inference Utilize the known 2- hop paths Weight based on the number of 2-hop paths Collective Inference becomes useful – Gibbs Sampling

Handling Uncertainty Border nodes with 1 or 2 observed edges Early Separate draws may not represent overall population Utilize the Labeled set to create priors for both Border and Separate

Handling Uncertainty - Separate Define a Beta prior based on the Labeled set – (Gamma) is used to weight the prior Use the expected value of the posterior Apply to each instance in Separate set

Handling Uncertainty - Border Use Beta prior from Labeled Create posterior using previous Border draws Use posterior as prior for individual Border instances

Evaluation Datasets AddHealth School 1: 635 Students, 24% Heavy Smokers AddHealth School 2: 576 Students, 15% Heavy Smokers Rovira Dataset: 1,133 Participants Methods Oracle – Always choose positive instance from Border nodes, if one is available Random – Randomly choose from the unlabeled instances Gibbs or NoGibbs – Proposed method using collective Inference or not Prior or NoPrior – Proposed method using a prior from previously acquired nodes, or not

Evaluation - Synthetic AddHealth School1 Rovira

Evaluation – AddHealth Schools School1School2

Conclusion and discussion Experimental results indicate that the network structure can be acquired actively, in order to improve identification of positive nodes and prediction of class labels collectively Using 2-hop network for Gibbs Sampling facilitates more accurate node predictions Priors, based on previously acquired instances, account for uncertainty associated with Border Future work: balance short term gain and long term gain; incorporate attributes to predict node labels

Questions?