Applications Statistical Graphical Models in Music Informatics Yushen Han Feb 10 2011 I548 Presentation.

Slides:



Advertisements
Similar presentations
Change-Point Detection Techniques for Piecewise Locally Stationary Time Series Michael Last National Institute of Statistical Sciences Talk for Midyear.
Advertisements

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
1 COMM 301: Empirical Research in Communication Lecture 15 – Hypothesis Testing Kwan M Lee.
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Statistics Review – Part II Topics: – Hypothesis Testing – Paired Tests – Tests of variability 1.
Hypothesis Testing Steps in Hypothesis Testing:
CHAPTER 2 Building Empirical Model. Basic Statistical Concepts Consider this situation: The tension bond strength of portland cement mortar is an important.
Experimental Design, Response Surface Analysis, and Optimization
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
STAT 497 APPLIED TIME SERIES ANALYSIS
Soundprism An Online System for Score-informed Source Separation of Music Audio Zhiyao Duan and Bryan Pardo EECS Dept., Northwestern Univ. Interactive.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
Evaluating Hypotheses
Chapter 2 Simple Comparative Experiments
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Inferences About Process Quality
Lecture II-2: Probability Review
Slide 1 Testing Multivariate Assumptions The multivariate statistical techniques which we will cover in this class require one or more the following assumptions.
Handwritten Character Recognition using Hidden Markov Models Quantifying the marginal benefit of exploiting correlations between adjacent characters and.
CHARM Advisory Board Meeting Institute for Historical Research, School of Advanced Study, UL Senate House, London 12 Dec 2006 Craig Stuart Sapp Centre.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
1 AUTOMATIC TRANSCRIPTION OF PIANO MUSIC - SARA CORFINI LANGUAGE AND INTELLIGENCE U N I V E R S I T Y O F P I S A DEPARTMENT OF COMPUTER SCIENCE Automatic.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 2 – Slide 1 of 25 Chapter 11 Section 2 Inference about Two Means: Independent.
Paper by Craig Stuart Sapp 2007 & 2008 Presented by Salehe Erfanian Ebadi QMUL ELE021/ELED021/ELEM021 5 March 2012.
More About Significance Tests
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Wireless and Mobile Computing Transmission Fundamentals Lecture 2.
Inferential Statistics 2 Maarten Buis January 11, 2006.
Learning Objectives In this chapter you will learn about the t-test and its distribution t-test for related samples t-test for independent samples hypothesis.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Overview Particle filtering is a sequential Monte Carlo methodology in which the relevant probability distributions are iteratively estimated using the.
Rhythmic Transcription of MIDI Signals Carmine Casciato MUMT 611 Thursday, February 10, 2005.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
CHARM Symposium Kings College London 6 January 2006 The Mazurka Project.
+ Chapter 12: More About Regression Section 12.1 Inference for Linear Regression.
Statistics - methodology for collecting, analyzing, interpreting and drawing conclusions from collected data Anastasia Kadina GM presentation 6/15/2015.
Section 10.1 Confidence Intervals
NON-NEGATIVE MATRIX FACTORIZATION FOR REAL TIME MUSICAL ANALYSIS AND SIGHT-READING EVALUATION Chih-Chieh Cheng, Diane J. Hu, and Lawrence K. Saul, UC San.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.1 One-Way ANOVA: Comparing.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
The Mazurka Project Science and Music Seminar University of Cambridge 28 Nov 2006 Craig Stuart Sapp Centre for the History and Analysis of Recorded Music.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Tracking with dynamics
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
ECE-7000: Nonlinear Dynamical Systems 2. Linear tools and general considerations 2.1 Stationarity and sampling - In principle, the more a scientific measurement.
Rao-Blackwellised Particle Filtering for Dynamic Bayesian Network Arnaud Doucet Nando de Freitas Kevin Murphy Stuart Russell.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
1 Nonparametric Statistical Techniques Chapter 18.
Copyright © 2009 Pearson Education, Inc t LEARNING GOAL Understand when it is appropriate to use the Student t distribution rather than the normal.
Rhythmic Transcription of MIDI Signals
Confidence Intervals.
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
8-1 of 23.
The general linear model and Statistical Parametric Mapping
Analyzing Redistribution Matrix with Wavelet
Hypothesis Testing: Hypotheses
Chapter 9 Hypothesis Testing.
Elementary Statistics
Discrete Event Simulation - 4
Graduate School of Information Sciences, Tohoku University
The general linear model and Statistical Parametric Mapping
Kalman Filter: Bayes Interpretation
Presentation transcript:

Applications Statistical Graphical Models in Music Informatics Yushen Han Feb I548 Presentation

Statistical Graphical Models graph-based representation for a probabilistic distribution in high-dimensional space while specifying conditional independence structure – directed acyclic graph(DAG) - Bayesian Network – undirected graph(UG) - Markov Random Field – Mixed graph

Saturated (Undirected) Graph

Markov Condition on a Bayesian Network Markov Condition: X A and X B are conditionally independent given X S whenever S separates A and B From probabilistic distribution High- dimensional space Conditional independence

Musical Application: Bayesian Identification of Chord from Audio tuning octave root Chroma (pitch- class) mode key Peak frequencies and corresponding amplitude Attributes Observation: factorization of the joint by Bayesian Inference

Application I: Score Following/Alignment Music score (given) Performance audio of this piece (given) Establishing a correspondence between the two above Score Following: Online (real time) Score Alignment: Offline (no need to be in real time)

Application I: Score Following Best “guess” of the current location given what the computer had heard SINCE THE BEGINNING UP TO THE MOMENT

Observation - “Frames” of Audio Audio (waveform) Spectrogram (via Short-time Fourier transform)

Score Following Difficulties – Tempo rubato (expressive and rhythmic freedom) – Pitch / amplitude vibrato ( ) – Polyphony music – Noise – (occasional) wrong notes etc. – Realtime computational requirement Solutions – Assuming tempo change is smooth (mostly desirable) – Robust probabilistic data model on normalized semigram – Training to learn a prior (e.g. note length distribution) – Optimized particle filtering for 2-D State-space model

Proposed Solution with 2-D State- space Model (Bayesian Network) Assuming smooth tempo change for tempo rubato Two-dimensional state-space model Proposing a unit of tempi: S(t) - Musical time elapse per audio frame at frame t Interpretation: during one audio frame of fixed length (roughly 64ms, 512 samples at 8000Hz sampling rate), how much musical time (in terms of 1/384 notes) is elapsing In another word, how much of the score the performer covers every 64ms (not precisely the conventional tempo)

2-D State-space Model State variables in one audio frame (notice conditional independence) Current “speed” at this frame 0.64ms Accumulative “speed” up to this frame – location

Physical Analogy – Integration of the Speed speed location

Physical Analogy – Integration of the Speed with stochastic components speed location Observation error also involved Speed fluctuation

Physical Analogy - to make a tractable problem speed location Observation error also involved Speed fluctuation Assuming discrete state transition with {-0.1, -0.05, 0, 0.05, 0.1} with prob distribution {0.1, 0.2, 0.4,0.2, 0.1} for smoothness. Assuming Gaussian noise in observation

Relationship to Kalman Filter Particle filtering (Using White board)

Score Following – Data Model Data model using semigram Regarding discretized observation as a histogram Chord template (pre-learned)

Score Following – Demo in R Data model in R Visualization of results in XCode

Application II: Graph Model to Estimate Expert Pianists’ Perceptual Present - with the help of audio-score alignment technique

Background Curtain eras of classical music – no improvisation, no wrong notes etc. For a certain piece of music, performance varies in tempo, dynamic, articulation, vibrato etc., depending on the interpretation of the performer (This research) focuses on the tempo change of piano music

Chopin Mazurka Op. 30, No Rubinstein Horowitz Michelangeli

Motivation Musical Perceptual Present – Recent studies in diverse fields of inquiry, including music philosophy and psychology, lend converging evidence that musical attention of both performers and listeners is primarily focused successively small “chunks” of material (hypothetically 2–10 seconds in the past) rather than larger formal relationships.

Motivation Instead of individual style, we are in search of a “common interpretation” shared among a collection of expert pianists Focus purely on tempo change per beat (since the attack of piano note is easy to capture).

Data Human corrected accumulative time per beat which is equivalent to IBI Inter Beat Interval N = 32 performances ( include different performances of the same pianist ) For the existence of the MLE, we proceed a small chunk of data at a time I = { 7, 8, 9, 10 }

Data cont. - preprocessing # !!!performance-id: pid # !!!title: Mazurka in B minor, Op. 30, No. 2 # !!!trials: 1 # !!!date: 2007/02/15/ # !!!reverse-conductor: Craig Stuart Sapp # !!!performer: Idil Biret # !!!performance-date: 1990 # !!!label: Naxos # !!!label-title: Chopin: Mazurkas (Complete) # !!!offset: : : : : : : : : : : : : : : : : : : : : : : … Original Data: Accumulative time per beat InterBeatInterval (IBI) IBI difference between beats …

Data cont. - preprocessing “Normalization” – Since the overall duration of each performance varies significantly E.g. Mazurka Op. 30. No. 2 We “stretch” the overall duration of each performance to line up with the median of all performances - can be problematic Sec. Performance index

Model For each “trunk” of timing data X of I dimensions (beats) across N performances: N performances are considered i.i.d. repetitions we assume: X ~ Ν( 0, Σ ) where the difference in IBI equals to 0 suggests that the tempo is nearly constant “on average” (of course, but could be problematic) We study the structure of I by I covariance matrix Σ Can obtain an estimate of

Model cont. – check the normality See the movie in R

Graph Models A toy example for I = 4 case -> 3 hypotheses – H: Fully saturated model (I-1=3 order Markov chain) – H 0 : A smaller model (I-2=2 order Markov chain) – H 00 : The smallest model (I-3=1 order Markov chain) diff. IBI

Graph Models cont. What does this graph mean? Conditional Independence!

Graph Models cont. Conditional independence in the graph suggests different structures in the covariance matrix ? ?? To apply reconstruction algorithm

Graph Models cont. - Testing Testing each pair of hypotheses -2Log(Q) ~ Accepting the result only when every single pair of hypotheses of the smallest difference between the alternative and the null hypotheses are not rejected (as small step as possible) Apply an appropriate degree of freedom ( = difference in number of edges between 2 graphs )

Results - Testing See R plot

Results – Interpretation “smoothed” results by using a sliding window of different lengths A “voting” mechanism Room to interpret …

Results – Interpretation “smoothed” results by using a sliding window of different lengths

Results – Interpretation “smoothed” results by using a sliding window of different lengths

Results – Interpretation Using “anchor points” to summarize the results