Learning to Align Polyphonic Music. Slide 1 Learning to Align Polyphonic Music Shai Shalev-Shwartz Hebrew University, Jerusalem Joint work with Yoram.

Slides:



Advertisements
Similar presentations
A Support Vector Method for Optimizing Average Precision
Advertisements

(SubLoc) Support vector machine approach for protein subcelluar localization prediction (SubLoc) Kim Hye Jin Intelligent Multimedia Lab
CHAPTER 2: Supervised Learning. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Learning a Class from Examples.
Online Max-Margin Weight Learning for Markov Logic Networks Tuyen N. Huynh and Raymond J. Mooney Machine Learning Group Department of Computer Science.
Structured SVM Chen-Tse Tsai and Siddharth Gupta.
SVM—Support Vector Machines
Power of Selective Memory. Slide 1 The Power of Selective Memory Shai Shalev-Shwartz Joint work with Ofer Dekel, Yoram Singer Hebrew University, Jerusalem.
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Soundprism An Online System for Score-informed Source Separation of Music Audio Zhiyao Duan and Bryan Pardo EECS Dept., Northwestern Univ. Interactive.
The loss function, the normal equation,
Present by: Fang-Hui Chu A Survey of Large Margin Hidden Markov Model Xinwei Li, Hui Jiang York University.
Feature Vector Selection and Use With Hidden Markov Models to Identify Frequency-Modulated Bioacoustic Signals Amidst Noise T. Scott Brandes IEEE Transactions.
Forgetron Slide 1 Online Learning with a Memory Harness using the Forgetron Shai Shalev-Shwartz joint work with Ofer Dekel and Yoram Singer Large Scale.
x – independent variable (input)
1 PEGASOS Primal Efficient sub-GrAdient SOlver for SVM Shai Shalev-Shwartz Yoram Singer Nati Srebro The Hebrew University Jerusalem, Israel YASSO = Yet.
1 Transfer Learning Algorithms for Image Classification Ariadna Quattoni MIT, CSAIL Advisors: Michael Collins Trevor Darrell.
Phoneme Alignment. Slide 1 Phoneme Alignment based on Discriminative Learning Shai Shalev-Shwartz The Hebrew University, Jerusalem Joint work with Joseph.
1 Robust Temporal and Spectral Modeling for Query By Melody Shai Shalev, Hebrew University Yoram Singer, Hebrew University Nir Friedman, Hebrew University.
Learning of Pseudo-Metrics. Slide 1 Online and Batch Learning of Pseudo-Metrics Shai Shalev-Shwartz Hebrew University, Jerusalem Joint work with Yoram.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Efficient and Numerically Stable Sparse Learning Sihong Xie 1, Wei Fan 2, Olivier Verscheure 2, and Jiangtao Ren 3 1 University of Illinois at Chicago,
Hidden Markov Models K 1 … 2. Outline Hidden Markov Models – Formalism The Three Basic Problems of HMMs Solutions Applications of HMMs for Automatic Speech.
Co-training LING 572 Fei Xia 02/21/06. Overview Proposed by Blum and Mitchell (1998) Important work: –(Nigam and Ghani, 2000) –(Goldman and Zhou, 2000)
Exploration and Apprenticeship Learning in Reinforcement Learning Pieter Abbeel and Andrew Y. Ng Stanford University.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Online Learning Algorithms
1 AUTOMATIC TRANSCRIPTION OF PIANO MUSIC - SARA CORFINI LANGUAGE AND INTELLIGENCE U N I V E R S I T Y O F P I S A DEPARTMENT OF COMPUTER SCIENCE Automatic.
Operations Research Models
Predicting Secondary Structure of All-Helical Proteins Using Hidden Markov Support Vector Machines Blaise Gassend, Charles W. O'Donnell, William Thies,
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
ANTs PI Meeting, Nov. 29, 2000W. Zhang, Washington University1 Flexible Methods for Multi-agent distributed resource Allocation by Exploiting Phase Transitions.
Polyphonic Music Transcription Using A Dynamic Graphical Model Barry Rafkind E6820 Speech and Audio Signal Processing Wednesday, March 9th, 2005.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Unsupervised Constraint Driven Learning for Transliteration Discovery M. Chang, D. Goldwasser, D. Roth, and Y. Tu.
Rhythmic Transcription of MIDI Signals Carmine Casciato MUMT 611 Thursday, February 10, 2005.
ICML2004, Banff, Alberta, Canada Learning Larger Margin Machine Locally and Globally Kaizhu Huang Haiqin Yang, Irwin King, Michael.
Online Passive-Aggressive Algorithms Shai Shalev-Shwartz joint work with Koby Crammer, Ofer Dekel & Yoram Singer The Hebrew University Jerusalem, Israel.
CSE 446 Logistic Regression Winter 2012 Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer.
Stefan Mutter, Mark Hall, Eibe Frank University of Freiburg, Germany University of Waikato, New Zealand The 17th Australian Joint Conference on Artificial.
Polyphonic Transcription Bruno Angeles McGill University - Schulich School of Music MUMT-621 Fall /14.
Nonlinear Data Discrimination via Generalized Support Vector Machines David R. Musicant and Olvi L. Mangasarian University of Wisconsin - Madison
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
Feature Vector Selection and Use With Hidden Markov Models to Identify Frequency-Modulated Bioacoustic Signals Amidst Noise T. Scott Brandes IEEE Transactions.
BLAST: Basic Local Alignment Search Tool Altschul et al. J. Mol Bio CS 466 Saurabh Sinha.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
Presented by: Fang-Hui Chu Discriminative Models for Speech Recognition M.J.F. Gales Cambridge University Engineering Department 2007.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
NTU & MSRA Ming-Feng Tsai
A Brief Maximum Entropy Tutorial Presenter: Davidson Date: 2009/02/04 Original Author: Adam Berger, 1996/07/05
1 Hidden Markov Model: Overview and Applications in MIR MUMT 611, March 2005 Paul Kolesnik MUMT 611, March 2005 Paul Kolesnik.
Hybrid Classifiers for Object Classification with a Rich Background M. Osadchy, D. Keren, and B. Fadida-Specktor, ECCV 2012 Computer Vision and Video Analysis.
Automated Speach Recognotion Automated Speach Recognition By: Amichai Painsky.
Piano Music Transcription Wes “Crusher” Hatch MUMT-614 Thurs., Feb.13.
A PAC-Bayesian Approach to Formulation of Clustering Objectives Yevgeny Seldin Joint work with Naftali Tishby.
Smooth ε -Insensitive Regression by Loss Symmetrization Ofer Dekel, Shai Shalev-Shwartz, Yoram Singer School of Computer Science and Engineering The Hebrew.
Learning by Loss Minimization. Machine learning: Learn a Function from Examples Function: Examples: – Supervised: – Unsupervised: – Semisuprvised:
Discriminative n-gram language modeling Brian Roark, Murat Saraclar, Michael Collins Presented by Patty Liu.
Melody Characterization by a Fuzzy Rule System Pedro J. Ponce de León, David Rizo, José M. Iñesta (DLSI, Univ. Alicante) Rafael Ramírez (MTG, Univ. Pompeu.
Journal of Computational and Applied Mathematics Volume 253, 1 December 2013, Pages 14–25 Reporter : Zong-Dian Lee A hybrid quantum inspired harmony search.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
Adaboost (Adaptive boosting) Jo Yeong-Jun Schapire, Robert E., and Yoram Singer. "Improved boosting algorithms using confidence- rated predictions."
Learning to Align: a Statistical Approach
Big data classification using neural network
Rhythmic Transcription of MIDI Signals
Boosted Augmented Naive Bayes. Efficient discriminative learning of
Recursive Identification of Switched ARX Hybrid Models: Exponential Convergence and Persistence of Excitation René Vidal National ICT Australia Brian D.O.Anderson.
CH. 2: Supervised Learning
HUMAN LANGUAGE TECHNOLOGY: From Bits to Blogs
Presentation transcript:

Learning to Align Polyphonic Music. Slide 1 Learning to Align Polyphonic Music Shai Shalev-Shwartz Hebrew University, Jerusalem Joint work with Yoram Singer, Google Inc. Joseph Keshet, Hebrew University

Learning to Align Polyphonic Music. Slide 2 Motivation Symbolic representation: Acoustic representation: Two ways for representing music

Learning to Align Polyphonic Music. Slide 3 Symbolic Representation time pitch - pitch symbolic representation: - start-time

Learning to Align Polyphonic Music. Slide 4 Acoustic Representation Feature Extraction (e.g. Spectral Analysis) acoustic representation: acoustic signal:

Learning to Align Polyphonic Music. Slide 5 The Alignment Problem Setting time pitch actual start-time:

Learning to Align Polyphonic Music. Slide 6 The Alignment Problem Setting Goal: learn an alignment function alignment function actual start-times acoustic representation - pitch symbolic representation - start-times

Learning to Align Polyphonic Music. Slide 7 Previous Work Dynamic Programming (rule based) Dannenberg 1984 Soulez et al Orio & Schwarz 2001 Generative Approaches Raphael 1999 Durey & Clements 2001 Shalev-Shwartz et al. 2002

Learning to Align Polyphonic Music. Slide 8 Our Solution Discriminative Learning Algorithm Training Set Alignment function Discriminative Learning from examples

Learning to Align Polyphonic Music. Slide 9 Why Discriminative Learning? “ When Solving a given problem, try to avoid a more general problem as an intermediate step ” (Vladimir Vapnik’s principle for solving problems using a restricted amount of information) Or, if you would like to visit Barcelona, buy a ticket ! Don’t waste so much time on writing a paper for ISMIR 2004 …

Learning to Align Polyphonic Music. Slide 10 Outline of Solution 1.Define a quantitative assessment of alignments 2.Define a hypotheses class - what is the form of our alignment functions : a.Map all possible alignments into vectors in an abstract vector-space b.Find a projection in the vector-space which ranks alignments according to their quality 3.Suggest a learning algorithm

Learning to Align Polyphonic Music. Slide 11 Assessing alignments e.g.

Learning to Align Polyphonic Music. Slide 12 Feature Functions for Alignment feature function for alignment Assessing the quality of a suggested alignment acoustic and symbolic representation suggested alignment (actual start-times) e.g.

Learning to Align Polyphonic Music. Slide 13 Feature Functions for Alignment correct alignment slightly incorrect alignment grossly incorrect alignment Mapping all possible alignments into a vector space

Learning to Align Polyphonic Music. Slide 14 Main Solution Principle grossly incorrect alignment correct alignment slightly incorrect alignment Find a linear projection that ranks alignments according to their quality

Learning to Align Polyphonic Music. Slide 15 slightly incorrect alignment Main Solution Principle (cont.) An example of projection with low confidence correct alignment grossly incorrect alignment

Learning to Align Polyphonic Music. Slide 16 slightly incorrect alignment Main Solution Principle (cont.) An example of incorrect projection correct alignment grossly incorrect alignment

Learning to Align Polyphonic Music. Slide 17 Hypotheses class The form of our alignment functions: predict the alignment which attains the highest projection defines the direction of projection

Learning to Align Polyphonic Music. Slide 18 Learning algorithm Optimization Problem: Given a training set : Find: a projection and a maximal confidence scalar such that the data is ranked correctly:

Learning to Align Polyphonic Music. Slide 19 Algorithmic aspects Iterative algorithm: Works on one alignment example at a time The algorithm works in polynomial time although the number of constraints is exponentially large Simple to implement Convergence: Converges to a high confidence solution #iterations depends on the best attainable confidence Generalization: The gap between test and train error decreases with the #examples. The gap is bounded above by

Learning to Align Polyphonic Music. Slide 20 Experimental Results Task: alignment of polyphonic piano music Dataset: 12 musical pieces where sound and MIDI were both recorded + other performances of the same pieces in MIDI format Features: see in the paper Algorithms: Discriminative method Generative method: Generalized Hidden Markov Model (GHMM) Using the same features as in the discriminative method Using different number of Gaussians (1,3,5,7)

Learning to Align Polyphonic Music. Slide 21 Experimental Results (Cont.) Our discriminative method outperforms GHMM GHMM-1 GHMM-3 GHMM-5 GHMM-7 Discriminative Loss (ms)

Learning to Align Polyphonic Music. Slide 22 The End