Music retrieval Conventional music retrieval systems Exact queries: ”Give me all songs from J.Lo’s latest album” What about ”Give me the music that I like”?

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Text Categorization.
DECISION TREES. Decision trees  One possible representation for hypotheses.
Naïve-Bayes Classifiers Business Intelligence for Managers.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Data Mining Classification: Alternative Techniques
Multidimensional Indexing
Content-based retrieval of audio Francois Thibault MUMT 614B McGill University.
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lecture 3 Nonparametric density estimation and classification
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
Learning from Observations Chapter 18 Section 1 – 4.
Lesson 8: Machine Learning (and the Legionella as a case study) Biological Sequences Analysis, MTA.
Modern Information Retrieval Chapter 2 Modeling. Can keywords be used to represent a document or a query? keywords as query and matching as query processing.
Principle of Locality for Statistical Shape Analysis Paul Yushkevich.
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
Non Parametric Classifiers Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Modern Information Retrieval Chapter 2 Modeling. Can keywords be used to represent a document or a query? keywords as query and matching as query processing.
Support Vector Machines
IR Models: Review Vector Model and Probabilistic.
05/06/2005CSIS © M. Gibbons On Evaluating Open Biometric Identification Systems Spring 2005 Michael Gibbons School of Computer Science & Information Systems.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Indexing Techniques Mei-Chen Yeh.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
A Simple Method to Extract Fuzzy Rules by Measure of Fuzziness Jieh-Ren Chang Nai-Jian Wang.
APPLICATIONS OF DATA MINING IN INFORMATION RETRIEVAL.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Text mining.
CS591k - 20th November - Fall Content-Based Retrieval of Music and Audio Seminar : CS591k Multimedia Systems By Rahul Parthe Anirudha Vaidya.
ArrayCluster: an analytic tool for clustering, data visualization and module finder on gene expression profiles 組員:李祥豪 謝紹陽 江建霖.
CSC321: Neural Networks Lecture 12: Clustering Geoffrey Hinton.
Improving Web Spam Classification using Rank-time Features September 25, 2008 TaeSeob,Yun KAIST DATABASE & MULTIMEDIA LAB.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Computational Intelligence: Methods and Applications Lecture 30 Neurofuzzy system FSM and covering algorithms. Włodzisław Duch Dept. of Informatics, UMK.
1 SUPPORT VECTOR MACHINES İsmail GÜNEŞ. 2 What is SVM? A new generation learning system. A new generation learning system. Based on recent advances in.
Text mining. The Standard Data Mining process Text Mining Machine learning on text data Text Data mining Text analysis Part of Web mining Typical tasks.
Beyond Sliding Windows: Object Localization by Efficient Subwindow Search The best paper prize at CVPR 2008.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
Information Retrieval Lecture 6 Introduction to Information Retrieval (Manning et al. 2007) Chapter 16 For the MSc Computer Science Programme Dell Zhang.
Similarity Searching in High Dimensions via Hashing Paper by: Aristides Gionis, Poitr Indyk, Rajeev Motwani.
Query Sensitive Embeddings Vassilis Athitsos, Marios Hadjieleftheriou, George Kollios, Stan Sclaroff.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Externally growing self-organizing maps and its application to database visualization and exploration.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
An Approximate Nearest Neighbor Retrieval Scheme for Computationally Intensive Distance Measures Pratyush Bhatt MS by Research(CVIT)
Information Retrieval and Organisation Chapter 16 Flat Clustering Dell Zhang Birkbeck, University of London.
USE RECIPE INGREDIENTS TO PREDICT THE CATEGORY OF CUISINE Group 7 – MEI, Yan & HUANG, Chenyu.
Optimization by Model Fitting Chapter 9 Luke, Essentials of Metaheuristics, 2011 Byung-Hyun Ha R1.
V. Clustering 인공지능 연구실 이승희 Text: Text mining Page:82-93.
Language Modeling Putting a curve to the bag of words Courtesy of Chris Jordan.
Text Categorization With Support Vector Machines: Learning With Many Relevant Features By Thornsten Joachims Presented By Meghneel Gore.
Vector Quantization CAP5015 Fall 2005.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
SIMILARITY SEARCH The Metric Space Approach
ARTIFICIAL NEURAL NETWORKS
Instance Based Learning
Christian Böhm, Bernhard Braunmüller, Florian Krebs, and Hans-Peter Kriegel, University of Munich Epsilon Grid Order: An Algorithm for the Similarity.
K Nearest Neighbor Classification
CS4670: Intro to Computer Vision
Machine Learning: UNIT-4 CHAPTER-1
Machine Learning with Clinical Data
Presentation transcript:

Music retrieval Conventional music retrieval systems Exact queries: ”Give me all songs from J.Lo’s latest album” What about ”Give me the music that I like”?  New methods are needed: sophisticated similarity measures Increasing importance: MP3 players (10 3 songs) Personal music collections (10 4 songs) Music on demand many songs, huge market value…

Proposal Try a classifier method –Similarity measure  enables matching of fuzzy data  always returns results Implement relevance feedback –User feedback Improves retrieval performance

Classifier systems Genetic programming Neural networks Curve fitting algorithms Vector quantizers

Tree structured Vector Quantization Audio parameterization Feature space: MFCC coefficients Quantization tree A supervised learning algorithm, TreeQ: Attempts to partition feature space for maximum class separation

Features: MFCC coefficients waveform DFTLogMelIDFT MFCCs: A 13-dimensional vector per window 5 minutes song  30  10 3 windows 100 Hamming windows/second

Classifying feature space

Nearest neighbor Discrimination line in feature space Problems: –Curse of dimensionality –Distribution assumptions –Complicated distributions

Vector Quantization: Adding decision surfaces Each surface is added such that It cuts only one dimension (speed) the mutual information is maximized:

Until further splits are not worthwile – according to certain stop conditions

Decision tree Tree partitions features space –L regions (cells/leaves) –Based on class belonging of training data

Template generation Generate templates for –Training data –Test data Each MFCC vector is routed through the tree

Template generation With a series of feature vectors, each vector will end up in one of the leaves. This results in a histogram, or template, for each series of feature vectors.

Template comparison Corpus templates – one per training class ABn X Query template Compute similarity sim(X,A), sim(X,B), sim(X,C), …sim(X,n) Augmented similarity measure, e.g. DiffSim(X) = sim(X,A) – sim(X,C)

Template comparison Corpus templates – one per training class ABn Query templates Compute similarity DiffSim(X) Sort Result list

Preliminary experiments Test subjects listened to 107 songs Rated them: good, fair, poor (class belonging C g, C f, C p ) Training process: –For each user Select randomly a subset (N songs) from each class Construct a tree based on class belonging Generate histogram templates for C g, C f, C p For each song X –Generate histogram template –Compute DiffSim(X) = sim(X,C g ) – sim(X,C p ) Sort the list of songs according to DiffSim

Results N13579 random,236,234,246,240,234 cos,305,364,370,388,389

Relevance feedback Result list user classifier

Implementation Adjust histogram profiles based on user feedback For each user –Select the top M songs from the result list –Add the contents of the songs to the histogram profile based on the user rating (class belonging C g, C f, C p ) –For each song X Generate histogram template Compute DiffSim(X) = sim(X,C g ) – sim(X,C p ) –Sort the list of songs according to DiffSim

Improvement Amount of training data N M ,685,8810,222,524,74 340,9419,7023,8017,2027,50 552,1532,1434,0827,9940,59 762,8943,4543,7636,4552,89