Memory and Melodic Density : A Model for Melody Segmentation

Slides:



Advertisements
Similar presentations
Clustering Basic Concepts and Algorithms
Advertisements

Content-based retrieval of audio Francois Thibault MUMT 614B McGill University.
Pitch-spelling algorithms David Meredith Aalborg University.
Music Perception. Why music perception? 1. Found in all cultures - listening to music is a universal activity. 2. Interesting from a developmental point.
Why Do Skips Precede Reversals? The Effect of Tessitura On Melodic Structure by Paul von Hippel & David Huron - Presented by Kevin (Cheng) Zhu.
A.Diederich – International University Bremen – USC – MMM – Spring 2005 Rhythm and timing  Clarke, E.F. Rhythm and timing in music. In Deutsch, D. Chapter.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Cluster Analysis.  What is Cluster Analysis?  Types of Data in Cluster Analysis  A Categorization of Major Clustering Methods  Partitioning Methods.
T.Sharon 1 Internet Resources Discovery (IRD) Music IR.
Similar Sequence Similar Function Charles Yan Spring 2006.
A Wavelet-Based Approach to the Discovery of Themes and Motives in Melodies Gissel Velarde and David Meredith Aalborg University Department of Architecture,
Correlation and Linear Regression
Polyphonic Queries A Review of Recent Research by Cory Mckay.
A Time Based Approach to Musical Pattern Discovery in Polyphonic Music Tamar Berman Graduate School of Library and Information Science University of Illinois.
The probability that a person visits a location. ~ means that two locations are equivalent.  Grids -> close to the border of two different.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
Hierarchical Dirichlet Process (HDP) A Dirichlet process (DP) is a discrete distribution that is composed of a weighted sum of impulse functions. Weights.
1 PATTERN COMPARISON TECHNIQUES Test Pattern:Reference Pattern:
Correlation and Prediction Error The amount of prediction error is associated with the strength of the correlation between X and Y.
MUMT611: Music Information Acquisition, Preservation, and Retrieval Presentation on Timbre Similarity Alexandre Savard March 2006.
Feature Vector Selection and Use With Hidden Markov Models to Identify Frequency-Modulated Bioacoustic Signals Amidst Noise T. Scott Brandes IEEE Transactions.
New Acoustic-Phonetic Correlates Sorin Dusan and Larry Rabiner Center for Advanced Information Processing Rutgers University Piscataway,
Learning to Detect Events with Markov-Modulated Poisson Processes Ihler, Hutchins and Smyth (2007)
John Lafferty Andrew McCallum Fernando Pereira
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
GG 313 beginning Chapter 5 Sequences and Time Series Analysis Sequences and Markov Chains Lecture 21 Nov. 12, 2005.
1 Hidden Markov Model: Overview and Applications in MIR MUMT 611, March 2005 Paul Kolesnik MUMT 611, March 2005 Paul Kolesnik.
Discovering Musical Patterns through Perceptive Heuristics By Oliver Lartillot Presentation by Ananda Jacobs.
1 Tempo Induction and Beat Tracking for Audio Signals MUMT 611, February 2005 Assignment 3 Paul Kolesnik.
Chapter 8: Estimating with Confidence
David Sears MUMT November 2009
CHAPTER 8 Estimating with Confidence
Chapter 8: Estimating with Confidence
Musical Similarity: More perspectives and compound techniques

Online Multiscale Dynamic Topic Models
Bayesian Generalized Product Partition Model
Introduction to Music scales
CHAPTER 10 Estimating with Confidence
Understanding Standards Event Higher Statistics Award
Essential Statistics (a.k.a: The statistical bare minimum I should take along from STAT 101)
Chapter 4 – Part 3.
Saliency detection Donghun Yeo CV Lab..
Expectation in music 2 Day 19
EE513 Audio Signals and Systems
Integrating Segmentation and Similarity in Melodic Analysis
Department of Electrical Engineering
What Is Good Clustering?
Chapter 8: Estimating with Confidence
Automatic Segmentation of Data Sequences
CHAPTER 8 Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
CHAPTER 8 Estimating with Confidence
CHAPTER 8 Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
INF 141: Information Retrieval
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
CHAPTER 8 Estimating with Confidence
Chapter 8: Estimating with Confidence
Measuring the Similarity of Rhythmic Patterns
Automatic Prosodic Event Detection
Presentation transcript:

Memory and Melodic Density : A Model for Melody Segmentation Ferrand et al.: Memory & Melodic Density Presentation by Amit Singh Memory and Melodic Density : A Model for Melody Segmentation By Miguel Ferrand, Peter Nelson, Geraint Wiggins Presentation by Amit Singh ISE599: Spring 2004

What and Why? Segmentation is the process of partitioning a melody Gives structure to the melody Useful as a preprocessing stage for : - Pattern discovery - Music search

The LBDM This model is based on Gestalt principles The ‘strength’ of the interval is given by:

LBDM (contd.) A change rule assigns boundaries to intervals with strength proportional to the degree of change between consecutive interval pairs Proximity rule scales the previous boundary down. For each parameter k a sequence is calculated and all sequences are normalized and combined to give the overall boundary strength profile. Suggested weights are: Pitch = rest = 0.25 IOI = 0.5 Local peaks indicate boundaries.

Melodic Density Segmentation Model In contrast with the LBDM, the melodic density model calculates melodic cohesion between pitch intervals Uses a sliding window system and an attenuation function to model short-term memory More recent events have a stronger contribution to the melody than earlier ones. Size of window is fixed – the tempo is the determining factor in boundary perception.

Formulae Melodic Density in a sequence of N events: The density at event i is: F(r) returns the frequency of the interval; ri(m,n) = |pi-m – pi-m-n| is pitch interval in semitones Attenuation function where ti = onset time of ei and M is duration of memory window

Results and Comparison The MDSM approach led to fewer boundaries detected than the LBDM approach MDSM had higher Precision and higher Recall than the LBDM Author’s Claim : The MDSM is a better method than the LBDM method MDSM is a more cognitively realistic approach

Unsupervised Learning of Melodic Segmentation: A Memory Based Approach Miguel Ferrand, Peter Nelson, Geraint Wiggins

Aims To provide an automatic method of performing melodic segmentation To do the above task with no prior musical knowledge

What is Melody? Melody is seen to be a “temporal process where sound events unfold in time” From basic information like pitch, duration and inter-onset intervals we get : Pitch Step : the interval distance between consecutive notes (in semitones) Duration Ratio: the ratio between the duration of consecutive events.

Example

Markov Models and Mixed Markov Models Markov model (nth order n-gram): - The probability of occurrence of a depends on the prior occurrence of n-1 other symbols and the probability of a sequence P(s) is given by: This causes problems if any of the terms has zero probability Mixed Markov Models:

Mixed Markov Models (contd.) aμ(wi|wi-μ) is a k x k transition matrix containing probabilities of occurrence of a symbol at position i given that it has occurred at position i-μ. Mixing coefficients Φ(μ) are estimated using an iterative process

Entropy and Boundary Prediction Boundaries occur when there is a change in entropy Entropy in a context ‘c’ is given by : w = all symbols that can be successors of context c Context c is a sequence of size n-1, where n is the order of the model Entropy vectors are calculated by taking successive context sequences from the feature vectors of the target melody and calculating their means and standard deviations. All values outside the standard deviation are rejected. Of the remaining, only those that have a contiguous high to low or low to high variation with respect to the mean are considered.

Results Compared results from actual listeners to results from the computational model Listeners were made to mark boundaries in the melodies (called L-boundaries) The computational model also marked boundaries In the case of DeBussy’s Syrinx, the 11 L-boundaries were predicted correctly by the software, but it also generated 5 extra ones.

Conclusion An entropy based model was constructed to evaluate boundaries using pitch and duration features. The experiment seems to corroborate the idea that variations in entropy constitute boundaries.