ECE 8443 – Pattern Recognition ECE 3163 – Signals and Systems Objectives: Pattern Recognition Feature Generation Linear Prediction Gaussian Mixture Models.

Slides:



Advertisements
Similar presentations
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Advertisements

ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Periodograms Bartlett Windows Data Windowing Blackman-Tukey Resources:
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
ECE 8443 – Pattern Recognition Objectives: Course Introduction Typical Applications Resources: Syllabus Internet Books and Notes D.H.S: Chapter 1 Glossary.
Machine Learning Neural Networks
Pattern Recognition and Machine Learning
1 Speech Parametrisation Compact encoding of information in speech Accentuates important info –Attempts to eliminate irrelevant information Accentuates.
Speaker Adaptation for Vowel Classification
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Feature vs. Model Based Vocal Tract Length Normalization for a Speech Recognition-based Interactive Toy Jacky CHAU Department of Computer Science and Engineering.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
The Social Web: A laboratory for studying s ocial networks, tagging and beyond Kristina Lerman USC Information Sciences Institute.
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Data Mining Techniques
ECE 8443 – Pattern Recognition EE 3512 – Signals: Continuous and Discrete Objectives: Definition of a System Examples Causality Linearity Time Invariance.
Isolated-Word Speech Recognition Using Hidden Markov Models
Machine Learning CUNY Graduate Center Lecture 1: Introduction.
EVENT MANAGEMENT IN MULTIVARIATE STREAMING SENSOR DATA National and Kapodistrian University of Athens.
Image recognition using analysis of the frequency domain features 1.
Multimodal Interaction Dr. Mike Spann
1 CS 551/651: Structure of Spoken Language Lecture 8: Mathematical Descriptions of the Speech Signal John-Paul Hosom Fall 2008.
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 02: BAYESIAN DECISION THEORY Objectives: Bayes.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Whitening.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
1 Linear Prediction. 2 Linear Prediction (Introduction) : The object of linear prediction is to estimate the output sequence from a linear combination.
Dan Rosenbaum Nir Muchtar Yoav Yosipovich Faculty member : Prof. Daniel LehmannIndustry Representative : Music Genome.
1 PATTERN COMPARISON TECHNIQUES Test Pattern:Reference Pattern:
Overview Particle filtering is a sequential Monte Carlo methodology in which the relevant probability distributions are iteratively estimated using the.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
LML Speech Recognition Speech Recognition Introduction I E.M. Bakker.
ECE 8443 – Pattern Recognition ECE 3163 – Signals and Systems Objectives: Definition of a System Examples Causality Linearity Time Invariance Resources:
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Introduction to Digital Signals
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
Lecture 2: Statistical learning primer for biologists
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
MSc Project Musical Instrument Identification System MIIS Xiang LI ee05m216 Supervisor: Mark Plumbley.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Objectives: Terminology Components The Design Cycle Resources: DHS Slides – Chapter 1 Glossary Java Applet URL:.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Autoregressive (AR) Spectral Estimation
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 04: GAUSSIAN CLASSIFIERS Objectives: Whitening.
ECE 8443 – Pattern Recognition EE 3512 – Signals: Continuous and Discrete Objectives: Eigenfunctions Fourier Series of CT Signals Trigonometric Fourier.
Linear Constant-Coefficient Difference Equations
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
ECE 8443 – Pattern Recognition ECE 3163 – Signals and Systems Objectives: Eigenfunctions Fourier Series of CT Signals Trigonometric Fourier Series Dirichlet.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Mixture Densities Maximum Likelihood Estimates.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
CSE 4705 Artificial Intelligence
Data Transformation: Normalization
Computing and Compressive Sensing in Wireless Sensor Networks
Presentation on Artificial Neural Network Based Pathological Voice Classification Using MFCC Features Presenter: Subash Chandra Pakhrin 072MSI616 MSC in.
LECTURE 01: COURSE OVERVIEW
Linear Prediction.
Machine Learning Math Essentials Part 2
EE513 Audio Signals and Systems
LECTURE 01: COURSE OVERVIEW
Presentation transcript:

ECE 8443 – Pattern Recognition ECE 3163 – Signals and Systems Objectives: Pattern Recognition Feature Generation Linear Prediction Gaussian Mixture Models Emerging Applications Resources: Introduction to DSP Fundamentals of ASR Pattern Recognition Information Theory Introduction to DSP Fundamentals of ASR Pattern Recognition Information Theory LECTURE 42: SELECTED APPLICATIONS Audio: URL:

ECE 3163: Lecture 42, Slide 1 Pattern Recognition Pattern Recognition: “the act of taking raw data and taking an action based on the category of the pattern.” Common Applications: speech recognition, fingerprint identification (biometrics), DNA sequence identification Related Terminology:  Machine Learning: The ability of a machine to improve its performance based on previous results.  Machine Understanding: acting on the intentions of the user generating the data. Related Fields: artificial intelligence, signal processing and discipline-specific research (e.g., speech and image recognition, biometrics, DNA sequencing). Classification is typically performed using advanced statistical methods (e.g., maximum likelihood, Gaussian models, Bayes’ Rule). Feature extraction is the process of converting a signal to a set of vectors that contain relevant information. Feature Extraction Post-Processing Classification Segmentation Sensing Input Decision

ECE 3163: Lecture 42, Slide 2 Common Frequency Domain Representations Salient information in the signal is often manifests itself in both the time and frequency domain. There are many ways to derive a frequency domain representation of a signal: The (fast) Fourier transform (or Discrete Cosine Transform) are often used in audio and image processing because of their linear properties and computationally efficient structure.

ECE 3163: Lecture 42, Slide 3 Example: Modeling Energy in a Speech Signal Salient information in the speech signal is often manifests itself in both the time-domain (e.g., energy as a function of time): and the frequency domain (e.g., spectral magnitude as a function of time): Modeling the change in energy as a function of time and frequency (and even space) forms the basis of most pattern recognition systems.

ECE 3163: Lecture 42, Slide 4 Example: Feature Extraction in Speech Recognition A popular approach for capturing these dynamics is the Mel-Frequency Cepstral Coefficients (MFCC) “front-end:”

ECE 3163: Lecture 42, Slide 5 Linear Prediction Another popular approach for capturing information in a signal integrates concepts of signal processing and statistics, and is known as linear prediction. In this approach, we attempt to model the signal as a linear combination of its previous samples. Why? Consider a p th -order linear prediction model: This model is based on notions of regression and minimum mean-squared error (MMSE): We can adjust the coefficients to minimize the mean-squared error: + – Regression or MMSE estimate

ECE 3163: Lecture 42, Slide 6 Linear Prediction and Filtering The solution to this optimization problem has a familiar look: where: Using the z-Transform, we can show that the error equation can be computed using a linear filter:

ECE 3163: Lecture 42, Slide 7 Linear Prediction and the Spectrum To the left are some examples of the linear prediction error for voiced speech signals. The points where the prediction error peaks are points in the signal where the signal is least predictable by a linear prediction model. In the case of voiced speech, this relates to the manner in which the signal is produced. The LP filter becomes increasingly more accurate if you increase the order of the model. We can interpret this as a spectral matching process, as shown to the right. As the order increases, the LP model better models the envelope of the spectrum of the original signal. The LP parameters can be used to recognize patterns in a signal.

ECE 3163: Lecture 42, Slide 8 A multivariate normal, or Gaussian, distribution is defined as: where the vector, , represents the mean and the matrix, , represents the covariance. Note the exponent term is really a dot product or weighted Euclidean distance. The covariance is always symmetric and positive semidefinite. How does the shape vary as a function of the covariance? It is common in many signal processing system to model the statistics of a feature vector as a weighted sum of Gaussian distributions: This is known as a Gaussian mixture model. Multivariate Normal Distributions

ECE 3163: Lecture 42, Slide 9 Capturing the Time-Frequency Evolution

ECE 3163: Lecture 42, Slide 10 Analytics Definition: A tool or process that allows an entity (i.e., business) arrive at an optimal or realistic decision based on existing data. (Wiki).Wiki Google is building a highly profitable business around analytics derived from people using its search engine. Any time you access a web page, you are leaving a footprint of yourself, particularly with respect to what you like to look at. This has allows advertisers to tailor their ads to your personal interests by adapting web pages to your habits. Web sites such as amazon.com, netflix.com and pandora.com have taken this concept of personalization to the next level.amazon.comnetflix.compandora.com As people do more browsing from their telephones, which are now GPS enabled, an entirely new class of applications is emerging that can track your location, your interests and your network of “friends.”

ECE 3163: Lecture 42, Slide 11 Social Networking Sites such as Facebook, MySpace, and Linkedin.com are building a business based on your desire to network with other people. By monitoring what sites/people you visit and how frequently you visit them, a graph can be built describing your social network. These graphs can be aggregated, or averaged, over groups to better understand group dynamics. The intersection of these graphs with other people’s graphs can help determine who or what is at the center of an activity. These graphs can often be clustered into three levels: immediate, near and distant contacts. Graph theory plays a prominent role in the analysis of your “connectedness.” Other types of information, such as the time of day and the physical location of your edge device, coupled with who/what you are viewing, can be useful in constructing a timeline of an event and understanding your habitual behavior. Deviation from this behavior can set off alerts (e.g., credit cards).

ECE 3163: Lecture 42, Slide 12 Predicting User Preferences These models can be used to generate alternatives for you that are consistent with your previous choices (or the choices of people like you). Such models are referred to as generative models because they can generate new data spontaneously that is statistically consistent with previously collected data. Alternately, you can build graphs in which movies are nodes and links represent connections between movies judged to be similar. Some sites, such as Pandora, allow you to continuously rate choices, and adapt the mathematical models of your preferences in real time. This area of science is known as adaptive systems, dealing with algorithms for rapidly adjusting to new data.

ECE 3163: Lecture 42, Slide 13 Summary Introduced the concept of a pattern recognition system. Discussed how it integrates signal processing and statistics. Introduced conventional techniques for converting a discrete-time signal to a sequence of feature vectors. Discussed linear prediction as a statistically-based method for modeling a signal as an all-pole or autoregressive process. Introduced Gaussian mixture models as a statistical technique for modeling systematic variations in parameters, such as linear prediction coefficients. Discussed several emerging applications that attempt to model and predict human behavior using statistics. Signal processing remains a vital and growing field. In upper-level and graduate studies, you can investigate more powerful statistical methods for modeling and controlling signals that might appear to be random.