Behavior Recognition Based on Machine Learning Algorithms for a Wireless Canine Machine Interface Students: Avichay Ben Naim Lucie Levy 14 May, 2014 Ort.

Slides:



Advertisements
Similar presentations
1 Gesture recognition Using HMMs and size functions.
Advertisements

Pattern Finding and Pattern Discovery in Time Series
Character Recognition using Hidden Markov Models Anthony DiPirro Ji Mei Sponsor:Prof. William Sverdlik.
Large Vocabulary Unconstrained Handwriting Recognition J Subrahmonia Pen Technologies IBM T J Watson Research Center.
Decision Tree Approach in Data Mining
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Variants, improvements etc. of activity recognition with wearable accelerometers Mitja Luštrek Jožef Stefan Institute Department of Intelligent Systems.
2004/11/161 A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition LAWRENCE R. RABINER, FELLOW, IEEE Presented by: Chi-Chun.
數據分析 David Shiuan Department of Life Science Institute of Biotechnology Interdisciplinary Program of Bioinformatics National Dong Hwa University.
Hidden Markov Models Ellen Walker Bioinformatics Hiram College, 2008.
Ch 9. Markov Models 고려대학교 자연어처리연구실 한 경 수
Statistical NLP: Lecture 11
Chapter 6: HIDDEN MARKOV AND MAXIMUM ENTROPY Heshaam Faili University of Tehran.
Hidden Markov Models Theory By Johan Walters (SR 2003)
Foundations of Statistical NLP Chapter 9. Markov Models 한 기 덕한 기 덕.
1 Hidden Markov Models (HMMs) Probabilistic Automata Ubiquitous in Speech/Speaker Recognition/Verification Suitable for modelling phenomena which are dynamic.
Hidden Markov Models Fundamentals and applications to bioinformatics.
Lecture 15 Hidden Markov Models Dr. Jianjun Hu mleg.cse.sc.edu/edu/csce833 CSCE833 Machine Learning University of South Carolina Department of Computer.
Accelerometer-based Transportation Mode Detection on Smartphones
PatReco: Hidden Markov Models Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
ID3 Algorithm Abbas Rizvi CS157 B Spring What is the ID3 algorithm? ID3 stands for Iterative Dichotomiser 3 Algorithm used to generate a decision.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
Decision Tree Algorithm
Learning, Uncertainty, and Information Big Ideas November 8, 2004.
Induction of Decision Trees
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Hidden Markov Model: Extension of Markov Chains
Chapter 3 (part 3): Maximum-Likelihood and Bayesian Parameter Estimation Hidden Markov Model: Extension of Markov Chains All materials used in this course.
Scenario Generation for the Asset Allocation Problem Diana Roman Gautam Mitra EURO XXII Prague July 9, 2007.
Hidden Markov Models. Hidden Markov Model In some Markov processes, we may not be able to observe the states directly.
Ensemble Learning (2), Tree and Forest
1 A Network Traffic Classification based on Coupled Hidden Markov Models Fei Zhang, Wenjun Wu National Lab of Software Development.
CS344 : Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 21- Forward Probabilities and Robotic Action Sequences.
THE HIDDEN MARKOV MODEL (HMM)
Hidden Markov Models for Sequence Analysis 4
Fundamentals of Hidden Markov Model Mehmet Yunus Dönmez.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Hidden Markov Models CBB 231 / COMPSCI 261 part 2.
Feature Vector Selection and Use With Hidden Markov Models to Identify Frequency-Modulated Bioacoustic Signals Amidst Noise T. Scott Brandes IEEE Transactions.
S. Salzberg CMSC 828N 1 Three classic HMM problems 2.Decoding: given a model and an output sequence, what is the most likely state sequence through the.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
CS Statistical Machine learning Lecture 24
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
Hidden Markovian Model. Some Definitions Finite automation is defined by a set of states, and a set of transitions between states that are taken based.
California Pacific Medical Center
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
1 Hidden Markov Models Hsin-min Wang References: 1.L. R. Rabiner and B. H. Juang, (1993) Fundamentals of Speech Recognition, Chapter.
Automated Speach Recognotion Automated Speach Recognition By: Amichai Painsky.
DECISION TREES Asher Moody, CS 157B. Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Machine Learning: A Brief Introduction Fu Chang Institute of Information Science Academia Sinica ext. 1819
A Hybrid Model of HMM and RBFN Model of Speech Recognition 길이만, 김수연, 김성호, 원윤정, 윤아림 한국과학기술원 응용수학전공.
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Other Models for Time Series. The Hidden Markov Model (HMM)
HW7: Evolutionarily conserved segments ENCODE region 009 (beta-globin locus) Multiple alignment of human, dog, and mouse 2 states: neutral (fast-evolving),
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Data Mining CH6 Implementation: Real machine learning schemes(2) Reporter: H.C. Tsai.
Graphical Models for Segmenting and Labeling Sequence Data Manoj Kumar Chinnakotla NLP-AI Seminar.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
10. Decision Trees and Markov Chains for Gene Finding.
Learning, Uncertainty, and Information: Learning Parameters
DECISION TREES An internal node represents a test on an attribute.
Privacy-Preserving Data Mining
Decision Trees.
Hidden Markov Models Part 2: Algorithms
1.
Handwritten Characters Recognition Based on an HMM Model
CS639: Data Management for Data Science
Presentation transcript:

Behavior Recognition Based on Machine Learning Algorithms for a Wireless Canine Machine Interface Students: Avichay Ben Naim Lucie Levy 14 May, 2014 Ort Braude College – SE Dept.

Motivation Definitions Description of the research Useful algorithms The classification flow ConclusionsAgenda

It’s all about a canine training Bloodhound Watchdogs Guide dog Police dog Search and rescue dog

Definitions Inertial Measurement Unit (IMU) The gyroscope The accumulator moving average filter decision tree

Inertial Measurement Units

The gyroscope Etymology Description and diagram   Modern Uses

moving average filter Statistical tool. analyze data points. A moving average is commonly used with time series.

decision tree Decision support tool. Decision trees are commonly used in decision analysis.

Useful algorithms Forward algorithm Viterbi algorithm Baum-Welch algorithm C4.5 algorithm Hidden Markov Model (HMM)

The state is directly visible to the observer The state transition probabilities are the only parameters H C Markov chain

Hidden Markov Model The state is not directly visible The output, dependent on the state, is visible. Each state has a probability distribution over the possible output tokens.

Hidden Markov Model

Three problems in HMM and their solutions  Problem 1: Given a certain Markov model what is the probability of a certain sequence of observation.  Solution 1: Forward algorithm.  Problem 2: Given a certain sequence of observation and a certain Markov model what is the most possible sequence of states that create this sequence of observation.  solution 2: Viterbi algorithm.

Forward algorithm Solve the first problem of HDD. Algorithm input: HMM. Algorithm output: probability of a certain sequence of observation. Equations:

Forward algorithm- Example Calculate the probability of :

Viterbi algorithm Dynamic programming algorithm. Find the most likely sequence of states from observed events. Equations: animated_demo.gif animated_demo.gif

Baum-Welch algorithm

Baum-Welch algorithm- Example To start we first guess the transition and emission matrices: The next step is to estimate a new transition matrix:

Baum-Welch algorithm- Example The new estimate for the S1 to S2 transition is now Then calculate the S2 to S1, S2 to S2 and S1 to S1probabilities.

Baum-Welch algorithm- Example Next, we want to estimate a new emission matrix : The new estimate for the E coming from S1 emission is now

Baum-Welch algorithm- Example calculate all the emission matrix. estimate the initial probabilities

C4.5 Algorithm C4.5 is an extension of ID3 algorithm The algorithm used to generate a decision tree greedy approach selecting the best attribute to split the dataset on each iteration Local worst splitBest split Good split

Entropy Normalized Information Gain Gini coefficient splitting criterions

Entropy H(S) is a measure of the amount of uncertainty in the set S. Where: S-The current set for which entropy is being calculated. X-set of classes in S. P(t)-The proportion of the number of elements in class X to the number of elements in set S. Where H(S)=0 the set S is perfectly classified. splitting criterion-Entropy

Example

IG(A) is the measure of the difference in entropy from before to after the set S is split on the attribute A. Where: H(S) is the entropy of set S. T-The subsets created from splitting set S by attribute A such that P(t)-The proportion of the number of elements in t to the number of elements in set S H(t)-entropy of the subset t. splitting criterion-Information Gain

splitting criterion-Gini coefficient The Gini coefficient measures the inequality among values of a frequency distribution. A Gini coefficient of zero expresses perfect equality, where all values are the same

It iterates through every unused attribute of the set S. calculates the entropy- H(S) or the Information gain- IG(A) of the attribute. Selects the attribute which has the smallest entropy or largest information gain value. pruning C4.5 Algorithm

The c’BAN Wireless communication device Wireless sensor platform Remote computational node

Data Collection protocol Five Labrador Retrievers Four different sensor sites:  rump  chest  abdomen  back

Description of the research CommentsThe BehaviorsNumber of repetitions Type of activity The dogs returned to a standing position between repetitions sitting, standing, lying down, eating off ground and standing on two legs 5static The dogs walk back to the starting position between repetitions walk up the stairs, walk across a platform to the ramp, walk down the ramp 3dynamic

The findings representative sample of acceleration data from the x-axis of the accelerometer in each of the four locations

The classification flow

stage 1 1. HMMs, Viterbi algorithm and Baum-Welch algorithm was used to identify each of the dynamic activities.

stage 2 2. decides if the activity is Dynamic or not.

stage 3 and 4 and 5 3. moving average filter. 4. Decision tree was used to distinguish between transitions and postures. 5. decides if the activity is a posture or a transition.

stage 6 and 7 6. Decision tree was used to classify the specific postures. 7. Finally the algorithm find the specific postures.

Conclusions

THANK YOU!