Download presentation
Presentation is loading. Please wait.
Published byCurtis Gray Modified over 8 years ago
1
Semi-Supervised Time Series Classification & DTW-D REPORTED BY WANG YAWEN
2
Outline Semi-Supervised Time Series Classification(KDD_06) ◦Background ◦Time Series Classification ◦Semi-Supervised Time Series Classification ◦Empirical Evaluation DTW-D: Time Series Semi-Supervised Learning from a single sample(KDD_13) ◦Introduction ◦DTW-D ◦Two Key Assumptions ◦Algorithm ◦Experiment
3
Semi-Supervised Time Series Classification LI WEI, EAMONN KEOGH DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING UNIVERSITY OF CALIFORNIA, RIVERSIDE {WLI, EAMONN}@CS.UCR.EDU
4
Semi-Supervised Time Series Classification Background ◦Labeled training data difficult or expensive to obtain ◦Use the value of unlabeled data
5
Semi-Supervised Time Series Classification Background ◦Learning from both labeled and unlabeled data SSL ◦Semi-supervised technique for building time series classifiers that take advantage of the large collections of unlabeled data ◦Only a handful of labeled examples to construct accurate classifiers ◦Note: the usefulness of unlabeled data depends on the critical assumption that the underlying models / features / kernels / similarity functions match well with the problem at hand
6
Semi-Supervised Time Series Classification Time Series Classification
7
Semi-Supervised Time Series Classification Time Series Classification
8
Semi-Supervised Time Series Classification Time Series Classification under realistic situation ◦It is typically not the case that we have two or more well defined classes ◦Positive class with some structure ◦Negative examples have little or no common structure ◦Can not in general assume that subsequences not belonging to the positive class look similar to each other ◦It is typically the cade that positive labeled examples are rare, but unlabeled data is abundant ◦ building binary time series classifiers for extremely imbalanced class distributions, with only a small number of labeled examples from the positive class
9
Semi-Supervised Time Series Classification ◦One-nearest-neighbor with Euclidean distance classifier
10
Semi-Supervised Time Series Classification ◦Training the classifier(train itself) ◦Step 1: train on the initial training set, where all labeled instances are positive and all unlabeled instances are regarded as negative ◦Note: the size of the training set never changes during the training process, but the labeled set is augmented ◦Step 2: classify the unlabeled data in the training set ◦Step 3: among all the unlabeled instances, the one we can most confidently classify as positive is the instance which is closet to the labeled positive examples, which will be added into the positive set. Go back to Step 1. ◦Look BackLook Back
11
Semi-Supervised Time Series Classification
13
◦Stooping criterion ◦Precision-recall breakeven point ◦Distance between the closest pair in the labeled positive set ◦Decrease gradually: space gets denser ◦Stabilizing phase: incorporate closest pair ◦Drop: negative example add
14
Semi-Supervised Time Series Classification Empirical Evaluation ◦Precision-recall breakeven point ◦The value at which precision and recall are equal
15
Semi-Supervised Time Series Classification Empirical Evaluation ◦ECG dataset
16
Semi-Supervised Time Series Classification Empirical Evaluation ◦Word Spotting dataset
17
Semi-Supervised Time Series Classification Empirical Evaluation ◦Gun dataset
18
Semi-Supervised Time Series Classification Empirical Evaluation ◦Wafer dataset
19
Semi-Supervised Time Series Classification Empirical Evaluation ◦Yoga dataset
20
DTW-D: Time Series Semi-Supervised Learning from a Single Example YANPING CHEN, BING HU, EAMONN KEOGH, GUSTAVO E.A.P.A BATISTA1 UNIVERSITY OF CALIFORNIA, RIVERSIDE 1UNIVERSIDADE DE SÃO PAULO-USP {YCHEN053, BHU002, EAMONN}@CS.UCR.EDU GBATISTA@ICMC.USP.BR
21
DTW-D Introduction ◦Unlabeled members of a circumscribed positive class may be closer to some unlabeled members of a diverse negative class than to the labeled positive class
22
DTW-D ED is an upper bound to DTW
23
DTW-D
24
Why did DTW not solve the problem? ◦There are other differences between P1 and U2, including the fact that the first and last peaks have different heights, DTW cannot mitigate this ◦Simple shape tend to be close to everything ◦Smooth, flat or least very slowly changing time series tend to be surprisingly close to other objects ◦Observation: if a class is characterized by the existence of intra-class warping(possibly among other distortions), then we should expect that moving from ED to DTW reduces distances more for intra-class comparisons than interclass caparisons.
25
DTW-D
26
Two Key Assumptions ◦Assumption 1: the positive class(the target concept) contains time warped versions of some platonic ideal(some prototypical shape), possibly with other types of noise/distortions. ◦Assumption 2: the negative class may be very diverse, and occasionally by chance produces objects close to a member of the positive class, even under DTW.
27
DTW-D Assumption 1 is mitigated by large amounts of labeled data ◦Our noted weakness of semi-supervised learning happens when the nearest instance to a labeled positive exemplar is a negative instance. With more labeled positive instances this becomes less and less likely to happen.
28
DTW-D Assumption 2 is compounded by a large negative dataset ◦If the negative class is random and/or diverse, then the larger the negative class is, the more likely it is that it will produce an instance that just happens to be close to a labeled positive item
29
DTW-D Assumption 2 is compounded by low complexity negative data ◦A complex time series is one that is not well approximated by few DFT coefficients or by a low degree polynomial. ◦Low complexity data close to everything
30
DTW-D Algorithm Details ◦One-class classifier with no training examples from the negative class
31
DTW-D Why is DTW-D better? ◦Training process(choosing) ◦DTW-D selects better labeled objects than DTW ◦Evaluation process(classification) ◦DTW-D is better at selecting the top K nearest neighbors during evaluation process
32
DTW-D Comparison to rival methods ◦We favor our rival approaches by offering them with fifty more initial labeled examples ◦We start with a single labeled example
33
DTW-D Experiment ◦Learning dataset ◦Labeled dataset P: a single positive example ◦Unlabeled dataset U: the rest of objects in the learning dataset ◦Holdout dataset ◦Test the accuracy of the learned classifier
34
DTW-D Experiment ◦Insect Wingbeat Sound Detection
35
DTW-D Experiment ◦Historical Manuscript Mining
36
DTW-D Experiment ◦Activity Recognition
37
Conclusion Semi-Supervised learning framework when only a small set of labeled examples is available Simple idea that dramatically improves the quality of SSL in time series domain. Future work ◦Revisit the stopping criteria issue in light of DTW-D ◦Consider other avenues where DTW-D may be useful
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.