Sharing Features among Dynamical Systems with Beta Processes

Slides:



Advertisements
Similar presentations
Sinead Williamson, Chong Wang, Katherine A. Heller, David M. Blei
Advertisements

Bayesian dynamic modeling of latent trait distributions Duke University Machine Learning Group Presented by Kai Ni Jan. 25, 2007 Paper by David B. Dunson,
Nonparametric hidden Markov models Jurgen Van Gael and Zoubin Ghahramani.
Bayesian Estimation in MARK
Adaption Adjusting Model’s parameters for a new speaker. Adjusting all parameters need a huge amount of data (impractical). The solution is to cluster.
Bayesian Methods with Monte Carlo Markov Chains III
Markov Chains 1.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Suggested readings Historical notes Markov chains MCMC details
Beam Sampling for the Infinite Hidden Markov Model Van Gael, et al. ICML 2008 Presented by Daniel Johnson.
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Nonparametric Bayesian Learning
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Bayesian Analysis for Extreme Events Pao-Shin Chu and Xin Zhao Department of Meteorology School of Ocean & Earth Science & Technology University of Hawaii-
Robin McDougall, Ed Waller and Scott Nokleby Faculties of Engineering & Applied Science and Energy Systems & Nuclear Science 1.
Bayes Factor Based on Han and Carlin (2001, JASA).
Adaption Def: To adjust model parameters for new speakers. Adjusting all parameters requires too much data and is computationally complex. Solution: Create.
Modeling Menstrual Cycle Length in Pre- and Peri-Menopausal Women Michael Elliott Xiaobi Huang Sioban Harlow University of Michigan School of Public Health.
MUMS a Measure of hUman Motion Similarity Francisco J Torres Reyes.
Hierarchical Dirichelet Processes Y. W. Tech, M. I. Jordan, M. J. Beal & D. M. Blei NIPS 2004 Presented by Yuting Qi ECE Dept., Duke Univ. 08/26/05 Sharing.
A General Framework for Tracking Multiple People from a Moving Camera
Hierarchical Dirichlet Process (HDP) A Dirichlet process (DP) is a discrete distribution that is composed of a weighted sum of impulse functions. Weights.
Simulation of the matrix Bingham-von Mises- Fisher distribution, with applications to multivariate and relational data Discussion led by Chunping Wang.
Spatial Dynamic Factor Analysis Hedibert Freitas Lopes, Esther Salazar, Dani Gamerman Presented by Zhengming Xing Jan 29,2010 * tables and figures are.
Lab3: Bayesian phylogenetic Inference and MCMC Department of Bioinformatics & Biostatistics, SJTU.
The Dirichlet Labeling Process for Functional Data Analysis XuanLong Nguyen & Alan E. Gelfand Duke University Machine Learning Group Presented by Lu Ren.
Discovering Deformable Motifs in Time Series Data Jin Chen CSE Fall 1.
Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation Radford M. Neal 발표자 : 장 정 호.
Summary We propose a framework for jointly modeling networks and text associated with them, such as networks or user review websites. The proposed.
Hierarchical Dirichlet Process and Infinite Hidden Markov Model Duke University Machine Learning Group Presented by Kai Ni February 17, 2006 Paper by Y.
Bayesian Generalized Kernel Mixed Models Zhihua Zhang, Guang Dai and Michael I. Jordan JMLR 2011.
Tracking Multiple Cells By Correspondence Resolution In A Sequential Bayesian Framework Nilanjan Ray Gang Dong Scott T. Acton C.L. Brown Department of.
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
Topic Models Presented by Iulian Pruteanu Friday, July 28 th, 2006.
Statistical Models for Partial Membership Katherine Heller Gatsby Computational Neuroscience Unit, UCL Sinead Williamson and Zoubin Ghahramani University.
Beam Sampling for the Infinite Hidden Markov Model by Jurgen Van Gael, Yunus Saatic, Yee Whye Teh and Zoubin Ghahramani (ICML 2008) Presented by Lihan.
Lecture #9: Introduction to Markov Chain Monte Carlo, part 3
by Ryan P. Adams, Iain Murray, and David J.C. MacKay (ICML 2009)
Adaption Def: To adjust model parameters for new speakers. Adjusting all parameters requires an impractical amount of data. Solution: Create clusters and.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
Stick-breaking Construction for the Indian Buffet Process Duke University Machine Learning Group Presented by Kai Ni July 27, 2007 Yee Whye The, Dilan.
Bayesian Speech Synthesis Framework Integrating Training and Synthesis Processes Kei Hashimoto, Yoshihiko Nankaku, and Keiichi Tokuda Nagoya Institute.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
The Phylogenetic Indian Buffet Process : A Non- Exchangeable Nonparametric Prior for Latent Features By: Kurt T. Miller, Thomas L. Griffiths and Michael.
Multi-label Prediction via Sparse Infinite CCA Piyush Rai and Hal Daume III NIPS 2009 Presented by Lingbo Li ECE, Duke University July 16th, 2010 Note:
NIPS 2013 Michael C. Hughes and Erik B. Sudderth
Introduction: Metropolis-Hasting Sampler Purpose--To draw samples from a probability distribution There are three steps 1Propose a move from x to y 2Accept.
Hierarchical Beta Process and the Indian Buffet Process by R. Thibaux and M. I. Jordan Discussion led by Qi An.
APPLICATIONS OF DIRICHLET PROCESS MIXTURES TO SPEAKER ADAPTATION Amir Harati and Joseph PiconeMarc Sobel Institute for Signal and Information Processing,
Markov-Chain-Monte-Carlo (MCMC) & The Metropolis-Hastings Algorithm P548: Intro Bayesian Stats with Psych Applications Instructor: John Miyamoto 01/19/2016:
Latent Feature Models for Network Data over Time Jimmy Foulds Advisor: Padhraic Smyth (Thanks also to Arthur Asuncion and Chris Dubois)
Edge Preserving Spatially Varying Mixtures for Image Segmentation Giorgos Sfikas, Christophoros Nikou, Nikolaos Galatsanos (CVPR 2008) Presented by Lihan.
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
Activity 1 30 Two Feet Jumps.
Bayesian Semi-Parametric Multiple Shrinkage
Nonparametric Bayesian Learning of Switching Dynamical Processes
Consistent and Efficient Reconstruction of Latent Tree Models
Accelerated Sampling for the Indian Buffet Process
Dynamical Statistical Shape Priors for Level Set Based Tracking
A Non-Parametric Bayesian Method for Inferring Hidden Causes
Robust Full Bayesian Learning for Neural Networks
Nonparametric Bayesian Texture Learning and Synthesis
NON-NEGATIVE COMPONENT PARTS OF SOUND FOR CLASSIFICATION Yong-Choon Cho, Seungjin Choi, Sung-Yang Bang Wen-Yi Chu Department of Computer Science &
Presentation transcript:

Sharing Features among Dynamical Systems with Beta Processes Emily B. Fox, Erik B. Sudderth, Michael I. Jordan, Alan S. Willsky NIPS 2009 Some pictures and slides are adopted from Emily Fox’s paper and slides in NIPS 2009

Outline Introduction Sticky HDP-HMM Beta process autoregressive HMM (BP-AR-HMM) Experiment Results Conclusion

Introduction Problem: To discover and model common behaviors in related time series. Jointly model multiple time series Transfer knowledge between related time series Uncover interesting relationships Standard way to represent the set of behaviors time series exhibit via associated features: Binary matrix Time Series Dynamic Behavior If implies that time series i exhibit feature k

picture is adopted from Emily Fox’s paper in ICML 2008 Sticky HDP-HMM HDP-HMM assumes all time series exhibit the same behaviors in same way picture is adopted from Emily Fox’s paper in ICML 2008

BP-AR-HMM Model Beta process prior masses The beta process distributed measure is represented by its masses and locations as Beta process prior Shared set of features or behaviors

BP-AR-HMM Model Features indicating behaviors of object i Transition patterns for object i

BP-AR-HMM Model Latent dynamical mode for object i Switching Vector Autoregressive (VAR) Process with order r

BP-AR-HMM Inference Metropolis-Hastings binary feature sampling “Birth and dead” reversible jump MCMC (RJMCMC) Proposal Distribution: is defined similarly, but using where

BP-AR-HMM Inference MH acceptance probability is where MH sampling is also used for updating and

Synthetic Experiments Five time series are generated from with

Motion Capture (MoCap) Experiments 6 CMU MoCap exercise routines Each routine includes some of the following 12 motions: running in place, jumping jacks, arm circles, side twists, knee raises, squats, punching, up and down, two variants of toe touches, arch over, and a reach out stretch Observation: 62-dimension 25 chains of the sampler for 15,000 burn-in and 5000 collections Chosen MCMC sample minimizes an expected Hamming distance criterion

MoCap Results Share motions: Jumping jacks, Side twists, Arm circles, Squats Unique motion: Bend Over, Punching, Toe Touch 1, Toe Touch 2 Split motions: knee raises, up and down, running in place

Comparison with Other Models

Conclusion Bayesian nonparametric framework for discovering dynamical features common to multiple time series Novel exact sampling algorithm for non-conjugate beta process models