Speed improvements to information retrieval-based dynamic time warping using hierarchical K-MEANS clustering Presenter: Kai-Wun Shih Gautam Mantena 1,2.

Slides:



Advertisements
Similar presentations
Hierarchical Cellular Tree: An Efficient Indexing Scheme for Content-Based Retrieval on Multimedia Databases Serkan Kiranyaz and Moncef Gabbouj.
Advertisements

Indexing DNA Sequences Using q-Grams
Learning Trajectory Patterns by Clustering: Comparative Evaluation Group D.
Word Spotting DTW.
Effective Keyword Based Selection of Relational Databases Bei Yu, Guoliang Li, Karen Sollins, Anthony K.H Tung.
Assessment. Schedule graph may be of help for selecting the best solution Best solution corresponds to a plateau before a high jump Solutions with very.
Image Indexing and Retrieval using Moment Invariants Imran Ahmad School of Computer Science University of Windsor – Canada.
SASH Spatial Approximation Sample Hierarchy
Using Structure Indices for Efficient Approximation of Network Properties Matthew J. Rattigan, Marc Maier, and David Jensen University of Massachusetts.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
1 Abstract This paper presents a novel modification to the classical Competitive Learning (CL) by adding a dynamic branching mechanism to neural networks.
Whole Genome Alignment using Multithreaded Parallel Implementation Hyma S Murthy CMSC 838 Presentation.
Basic Data Mining Techniques Chapter Decision Trees.
Basic Data Mining Techniques
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
Efficient Estimation of Emission Probabilities in profile HMM By Virpi Ahola et al Reviewed By Alok Datar.
Decision Trees (2). Numerical attributes Tests in nodes are of the form f i > constant.
1 Indexing Structures for Files. 2 Basic Concepts  Indexing mechanisms used to speed up access to desired data without having to scan entire.
Efficient Distance Computation between Non-Convex Objects By Sean Quinlan Presented by Sean Augenstein and Nicolas Lee.
FLANN Fast Library for Approximate Nearest Neighbors
Birch: An efficient data clustering method for very large databases
Fast Subsequence Matching in Time-Series Databases Christos Faloutsos M. Ranganathan Yannis Manolopoulos Department of Computer Science and ISR University.
ICS 220 – Data Structures and Algorithms Week 7 Dr. Ken Cosh.
Zero Resource Spoken Term Detection on STD 06 dataset Justin Chiu Carnegie Mellon University 07/24/2012, JHU.
1 TEMPLATE MATCHING  The Goal: Given a set of reference patterns known as TEMPLATES, find to which one an unknown pattern matches best. That is, each.
S DTW: COMPUTING DTW DISTANCES USING LOCALLY RELEVANT CONSTRAINTS BASED ON SALIENT FEATURE ALIGNMENTS K. Selçuk Candan Arizona State University Maria Luisa.
CSCE350 Algorithms and Data Structure Lecture 17 Jianjun Hu Department of Computer Science and Engineering University of South Carolina
The BIRCH Algorithm Davitkov Miroslav, 2011/3116
The X-Tree An Index Structure for High Dimensional Data Stefan Berchtold, Daniel A Keim, Hans Peter Kriegel Institute of Computer Science Munich, Germany.
K. Selçuk Candan, Maria Luisa Sapino Xiaolan Wang, Rosaria Rossini
Copyright © Curt Hill Query Evaluation Translating a query into action.
Clustering Supervised vs. Unsupervised Learning Examples of clustering in Web IR Characteristics of clustering Clustering algorithms Cluster Labeling 1.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
A Query Adaptive Data Structure for Efficient Indexing of Time Series Databases Presented by Stavros Papadopoulos.
Video Google: A Text Retrieval Approach to Object Matching in Videos Josef Sivic and Andrew Zisserman.
Chapter 9 DTW and VQ Algorithm  9.1 Basic idea of DTW  9.2 DTW algorithm  9.3 Basic idea of VQ  9.4 LBG algorithm  9.5 Improvement of VQ.
Indexing and hashing Azita Keshmiri CS 157B. Basic concept An index for a file in a database system works the same way as the index in text book. For.
Sequence Comparison Algorithms Ellen Walker Bioinformatics Hiram College.
March 23 & 28, Csci 2111: Data and File Structures Week 10, Lectures 1 & 2 Hashing.
Event retrieval in large video collections with circulant temporal encoding CVPR 2013 Oral.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Levels of Image Data Representation 4.2. Traditional Image Data Structures 4.3. Hierarchical Data Structures Chapter 4 – Data structures for.
Exact indexing of Dynamic Time Warping
BIRCH: Balanced Iterative Reducing and Clustering Using Hierarchies A hierarchical clustering method. It introduces two concepts : Clustering feature Clustering.
Marwan Al-Namari Hassan Al-Mathami. Indexing What is Indexing? Indexing is a mechanisms. Why we need to use Indexing? We used indexing to speed up access.
Voice Activity Detection based on OptimallyWeighted Combination of Multiple Features Yusuke Kida and Tatsuya Kawahara School of Informatics, Kyoto University,
UNIT 5.  The related activities of sorting, searching and merging are central to many computer applications.  Sorting and merging provide us with a.
Social-Aware Stateless Forwarding in Pocket Switched Networks Soo-Jin SHIN
A Multiresolution Symbolic Representation of Time Series Vasileios Megalooikonomou Qiang Wang Guo Li Christos Faloutsos Presented by Rui Li.
File Systems cs550 Operating Systems David Monismith.
Jen-Tzung Chien, Meng-Sung Wu Minimum Rank Error Language Modeling.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Basic Machine Learning: Clustering CS 315 – Web Search and Data Mining 1.
Database Management Systems, R. Ramakrishnan 1 Algorithms for clustering large datasets in arbitrary metric spaces.
Predicting the Location and Time of Mobile Phone Users by Using Sequential Pattern Mining Techniques Mert Özer, Ilkcan Keles, Ismail Hakki Toroslu, Pinar.
DYNAMIC TIME WARPING IN KEY WORD SPOTTING. OUTLINE KWS and role of DTW in it. Brief outline of DTW What is training and why is it needed? DTW training.
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
File Processing : Query Processing 2008, Spring Pusan National University Ki-Joune Li.
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Azita Keshmiri CS 157B Ch 12 indexing and hashing
Database Management System
Instance Based Learning
핵심어 검출을 위한 단일 끝점 DTW 알고리즘 Yong-Sun Choi and Soo-Young Lee
Subject Name: File Structures
The Functional Space of an Activity Ashok Veeraraghavan , Rama Chellappa, Amit Roy-Chowdhury Avinash Ravichandran.
Dynamic Time Warping and training methods
BIRCH: Balanced Iterative Reducing and Clustering Using Hierarchies
Unit 12 Index in Database 大量資料存取方法之研究 Approaches to Access/Store Large Data 楊維邦 博士 國立東華大學 資訊管理系教授.
Presentation transcript:

Speed improvements to information retrieval-based dynamic time warping using hierarchical K-MEANS clustering Presenter: Kai-Wun Shih Gautam Mantena 1,2 ∗,Xavier Anguera 1 1 Telefonica Research,Edicio Telefonica-Diagonal 00,08019 Barcelona, Spain 2 International Institute of Information Technology-Hyderabad, India

Outline  Information retrieval-based DTW(IR-DTW)  Indexing and retrieval using hierarchical K-MEANS  Experiment and results

Most DTW-based algorithms build the matrices of size N x M, where N, M are the length of the query and the reference audio. This is a big limitation when requiring to search on big databases. Instead, IR-DTW performs the matching just using a vector memory structure, thus reducing the amount of total memory required, regardless of the size of the query and the reference data. In addition, by indexing the reference data a considerable speedup is obtained. Information retrieval-based DTW(IR-DTW)

Next briefly review the IR-DTW algorithm. Let Q = {q 1,q 2,…,q i,…,q N } be a query audio file containing N feature vectors, and R = {r 1,r 2,…,r j,…,r M } be a reference audio file containing M feature vectors. The main steps of the IR-DTW are shown in Algorithm1. Information retrieval-based DTW(IR-DTW)

IR-DTW performs a sequential scan of all the query points q i in Q. In STEP 1 of Algorithm 1 a search is performed for all the reference points R’ ⊆ R that are similar to a given query q i. Information retrieval-based DTW(IR-DTW)

This can be done using an exhaustive comparison of all query- reference pairs or, as proposed in this paper, using a fast retrieval of the reference points (STEP 1 in algorithm 1) using a hierarchical K- Means clustering based indexing and retrieval approach. Information retrieval-based DTW(IR-DTW)

From this point on, only matching pairs are used to find final matching subsequences. For this reason, a system parameter maxQDist is introduced as the maximum allowed distance between any two matching points within a good matching subsequence. Information retrieval-based DTW(IR-DTW)

From the retrieved reference points R’, a match point element is defined as a triplet composed of tq i,tr j and d(q i, r j ). tq i and tr j are the time offsets of the query and the reference data measured from the start of their respective temporal sequences. Information retrieval-based DTW(IR-DTW)

d(i,j) is the distance measure between query q i and reference r j points, given by a cosine similarity in logarithmic scale. Information retrieval-based DTW(IR-DTW)

The core of the IR-DTW algorithm (in STEP 2 of Algorithm 1) combines all match point elements into matching subsequences of low overall distance between query and reference that follow some local and global warping constraints. Information retrieval-based DTW(IR-DTW)

In STEP 2 of Algorithm 1 a vector data structure called △ T is used to store the matching subsequences formed by concatenating the match point elements. Information retrieval-based DTW(IR-DTW)

The location in △ T where subsequences are created and updated is based on the offset calculated from the time stamps of query and reference points in each match point element. For every query-reference matching point t offset = tr j − tq i. Information retrieval-based DTW(IR-DTW)

Would there be a diagonal match between the query and some portion in the reference, t offset would be constant across that region in the reference. Due to the variation in the speaking rate and acoustic conditions this is seldom true, even for the same speaker. Information retrieval-based DTW(IR-DTW)

Following what is done in a standard DTW implementation, in the IR-DTW algorithm a range of values around t offset are checked to find the best matching subsequence to contain each match point element. Information retrieval-based DTW(IR-DTW)

The input match point element is appended to the subsequence found to be the best among those within range. This decision can be taken based on the normalized score (similar to DTW) or on the subsequence length. Information retrieval-based DTW(IR-DTW)

Instead, if there is no existing subsequence at any location in the range defined by t’ offset then the match point element is used to create a new subsequence at location t offset. Information retrieval-based DTW(IR-DTW)

Once a sequential scan of all the query points is complete, △ T contains all the possible matching subsequences between the query and the reference data. At this point we perform STEP 3 to select those subsequences with a minimum length and the net score below a defined threshold. Such a threshold is applied to remove any spurious paths that might have been detected. Information retrieval-based DTW(IR-DTW)

Although the IR-DTW is already finished and all matching subsequences have been found, in practice a standard DTW algorithm is then applied in the detected regions to get an accurate start and end time stamps and a score dependent on all frame pairs (not only those selected as match point elements). Information retrieval-based DTW(IR-DTW)

In this paper, we focus on an indexing and retrieval technique for obtaining the reference points for a given query. In standard IR-DTW we compute distances for a given query q i with all the reference points and compare the scores to a given threshold θ to select the closest reference points. When dealing with large scale databases, computing distances for all the reference points is time consuming and is a hindrance for real time systems. INDEXING AND RETRIEVAL USING HIERARCHICAL K-MEANS

In this paper, we propose an indexing type of approach to overcome the problems with the exhaustive search. We use a hierarchical K- Means clustering based approach to cluster all the reference data points. A given query point q i traverses the tree to retrieve the closest reference points. In this approach the query point q i is no longer compared with each of the reference points but only with a few cluster centers. INDEXING AND RETRIEVAL USING HIERARCHICAL K-MEANS

Cluster centers which are closest to the query point q i, based on a threshold, are selected and all the reference points belonging to those clusters are retrieved. A detailed description of the indexing and retrieval is given in sections 3.1 and 3.2. INDEXING AND RETRIEVAL USING HIERARCHICAL K-MEANS

Indexing the reference vectors The reference points are indexed using hierarchical K-Means. Each node in the tree corresponds to a cluster. Each cluster can be further divided into sub-clusters based on some criteria. All the leaf nodes contain the indices of the reference points. The goal is to index the data in the form of a tree where similar reference points lie within the same node (or cluster).

Indexing the reference vectors Splitting each of the nodes in the tree is based on the number of points present in the given node. This is to make sure that the nodes do not contain too many reference points, which would result in increased false alarms. In building the hierarchical K-Means tree, we define K which is the maximum number of children a given node can have. Each of the inner nodes will have a minimum of 2 children and a maximum of K. For simplicity K-Means initialization was done using random selection of the data points.

Indexing the reference vectors At times, a given node has a large number of data points but the cluster is very dense to make any further partitions. For such clusters, the clustering algorithm is performed again by choosing K=2. For these nodes, the first center initialization is randomly selected and the second center is chosen by selecting a point farthest from the first. This is to ensure that the clusters are as small as possible.

Fast retrieval of reference vectors All the reference vectors are indexed using a hierarchical K-Means tree. In the tree data structure the leaf nodes contain the indices for all the reference data vectors. Tree traversal is based on the same thresholding condition, θ, used in algorithm 1.

Fast retrieval of reference vectors

where β is called the overlapping factor. The smaller the value of the overlapping factor the more likely the selection of the node will be. β plays a role in maintaining a balance between speed and accuracy. As shown in tables 1 and 3, smaller values of β indicate a strict selection process of the node and increases the speed of IR-DTW while larger values of β show better performance.

Experiment and results For the K-Means clustering the maximum value of K was set to 8. To avoid clusters with very few points, the minimum number of points in a cluster is set to 200 data points. A constant frame threshold of θ = 4 was considered for all the experiments. The results reported in tables 1 and 2.

In table 1, MTWV scores indicate that the IR-DTW with exhaustive search performs better than the S-DTW, and that the proposed tree- based IR-DTW shows a similar performance to that of exhaustive IR-DTW for β = 0.5. In standard IR-DTW we do an exhaustive search for all the reference points whose similarity measure is less than θ. ExpS-DTWIR-DTWTree-based IR-DTW β = 0.4 β = 0.5 β = 0.6 dev eval Table 1. Maximum term weighted values (MTWV) of S-DTW, exhaustive IR- DTW and tree-based IR-DTW for various values of β. Experiment and results

Table 1 scores show that the tree-based IR-DTW for β = 0.5 performs better than for β = 0.6 because for β = 0.6 there are a lot of reference points which are not similar to a given query point. Experiment and results

To validate the performance of the system we need to define the terms coverage, relevance and miss percentage: Coverage: Percentage of reference points retrieved from the complete reference data set. Relevance: Percentage of reference points retrieved which are relevant, i.e. the percentage of retrieved reference points whose distance from a given query q i is below a defined threshold θ. Miss: Percentage of reference points which are closer to the query but have not been selected. This occurs when a particular cluster fails the node selection criteria but has reference points close to the given query point. Experiment and results

Table 2 shows the average coverage, relevance and miss percentages computed for the dev and eval data for the various β values. Experiment and results Table 2. Average coverage and relevance scores computed for the dev and eval data sets. β CoverageRelevanceMiss Percentage %96.12%60.97% %84.06%44.91% %69.53%34.85%

Experiment and results A lower coverage percentage indicates that fewer reference points have been retrieved which results in a speedup as there are less points to be inserted into △ T. Low coverage reflects a high relevance percentage because many points retrieved from the leaf node cluster are closest to the query and subsequently giving a high miss percentage and resulting in a poor performance.

From tables 1 and 2, we conclude that tree-based search performs better for β = 0.5. Even though the proposed approach is missing 44.9% of the points for β = 0.5 it is still able to find the matching paths. Experiment and results β = 0.6β = 0.5β = 0.4 ExpR’Alg.1R’Alg.1R’Alg.1 Dev Eval Table 3. Speed improvements for retrieval and in Algorithm 1 using tree-based over exhaustive search IR-DTW

R’is the ratio of computation time in retrieving the reference points(STEP 1) and Alg.1 is the ratio of total computation in retrieving the reference points and updating the matching paths in △ T (STEP1 and STEP 2 in Algorithm 1). There is a very big improvement in the computation time as we are only comparing the query q i with the cluster centers to retrieve all the relevant points. In table 3, lower values of β indicate a lower coverage percentage resulting in the speedup in the retrieval of the reference points and subsequently a speedup in the algorithm 1. On the other hand,low values of beta have a higher miss percentage resulting in a poor performance and is shown in table 1. Experiment and results

CONCLUSIONS AND FUTURE WORK With the increase in multi-media data over the Internet, QbE-STD has become important in providing a search mechanism to find spoken queries in a spoken audio. Due to the limitations of training Hidden Markov Models for automatic speech recognition for lowresource languages, zero resource pattern matching techniques were proposed. In general, approaches derived from the well known DTW algorithms suffer from scalability problems. The proposed method has shown an estimated speedup of 2400X.