B. RAMAMURTHY EAP#2: Data Mining, Statistical Analysis and Predictive Analytics for Automotive Domain CSE651C, B. Ramamurthy 1 6/28/2014.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Pattern Recognition and Machine Learning
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
The General Linear Model. The Simple Linear Model Linear Regression.
Visual Recognition Tutorial
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Machine Learning CMPT 726 Simon Fraser University
Part III: Inference Topic 6 Sampling and Sampling Distributions
Modeling Gene Interactions in Disease CS 686 Bioinformatics.
Thanks to Nir Friedman, HU
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Maximum likelihood (ML)
Statistics for Marketing & Consumer Research Copyright © Mario Mazzocchi 1 Further advanced methods Chapter 17.
Modern Navigation Thomas Herring
Elec471 Embedded Computer Systems Chapter 4, Probability and Statistics By Prof. Tim Johnson, PE Wentworth Institute of Technology Boston, MA Theory and.
Crash Course on Machine Learning
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
PATTERN RECOGNITION AND MACHINE LEARNING
Anomaly detection with Bayesian networks Website: John Sandiford.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Statistical Decision Theory
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Chapter 7 Point Estimation
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
CHAPTER 2 Statistical Inference, Exploratory Data Analysis and Data Science Process cse4/587-Sprint
1 A Bayesian statistical method for particle identification in shower counters IX International Workshop on Advanced Computing and Analysis Techniques.
Machine Learning CUNY Graduate Center Lecture 4: Logistic Regression.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
INTRODUCTION TO Machine Learning 3rd Edition
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
Overview G. Jogesh Babu. Overview of Astrostatistics A brief description of modern astronomy & astrophysics. Many statistical concepts have their roots.
Sampling and estimation Petter Mostad
Linear Methods for Classification Based on Chapter 4 of Hastie, Tibshirani, and Friedman David Madigan.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 Statistics & R, TiP, 2011/12 Neural Networks  Technique for discrimination & regression problems  More mathematical theoretical foundation  Works.
Data Mining and Decision Support
Machine Learning 5. Parametric Methods.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Stats Term Test 4 Solutions. c) d) An alternative solution is to use the probability mass function and.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Overview G. Jogesh Babu. R Programming environment Introduction to R programming language R is an integrated suite of software facilities for data manipulation,
Lecture 1.31 Criteria for optimal reception of radio signals.
Ch3: Model Building through Regression
Parameter Estimation 主講人:虞台文.
Chapter 7: Sampling Distributions
Statistical Inference, Exploratory Data Analysis and Data Science Process Chapter 2 CSE4/587 B. Ramamurthy 9/21/2018.
Special Topics In Scientific Computing
Data Mining Lecture 11.
Data Science Process Chapter 2 Rich's Training 11/13/2018.
Course Outline MODEL INFORMATION COMPLETE INCOMPLETE
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Pattern Recognition and Machine Learning
INNOvation in TRAINING BUSINESS ANALYSTS HAO HElEN Zhang UniVERSITY of ARIZONA
LECTURE 07: BAYESIAN ESTIMATION
Parametric Methods Berlin Chen, 2005 References:
Mathematical Foundations of BME Reza Shadmehr
Applied Statistics and Probability for Engineers
Presentation transcript:

B. RAMAMURTHY EAP#2: Data Mining, Statistical Analysis and Predictive Analytics for Automotive Domain CSE651C, B. Ramamurthy 1 6/28/2014

Data Collection in Automobiles CSE651C, B. Ramamurthy 2 Large volumes of data is being collected from the increasing number of sensors that are being added to modern automobiles. Traditionally this data is used for diagnostics purposes, after a certain incident look for the causes… How else can you use this data? How about predictive analytics?  For example, predict the failure of a part based on the historical data and on-board data collected?  Discover unusual pattern, in say, fuel consumption. Traditionally, 55mph the optimal speed for fuel consumption...may be not so today. How can we do this? 6/28/2014

Introduction Data represents the traces of the real-world processes.  What traces we collect depends on the sampling methods Two sources of randomness and uncertainty:  The process that generates data is random  The sampling process itself is random Your mind-set should be “statistical thinking in the age of big-data”  Combine statistical approach with big-data Our goal for this emerging application area: understand the statistical process of dealing with automobile data and practice it using R How can you use this idea in your term project/capstone project? CSE651C, B. Ramamurthy 3 6/28/2014

Transforming data into analytics  Strategies/decisions 6/28/2014 CSE651C, B. Ramamurthy 4 Automotive (sensor) data Social/ media data/ web data Probability- Statistics- Stochastic Randomness, Uncertainty Machines learning algorithms Results Decisions Diagnosis Strategies Vertical domain Horizontal domain

Uncertainty and Randomness A mathematical model for uncertainty and randomness is offered by probability theory. A world/process is defined by one or more variables. The model of the world is defined by a function: Model == f(w) or f(x,y,z) (A multivariate function) The function is unknown  model is unclear, at least initially. Typically our task is to come up with the model, given the data. Uncertainty: is due to lack of knowledge: GM’s faulty ignition switch; Toyota’s faulty acceleration pedal; Randomness: is due lack of predictability: 1-6 when rolling a die Both can be expressed by probability theory CSE651C, B. Ramamurthy 5 6/28/2014

Statistical Inference CSE651C, B. Ramamurthy 6 World  Collect Data  Capture the understanding/meaning of data through models or functions  statistical estimators for predicting things about  The same world Development of procedures, methods, and theorems that allow us to extract meaning and information from data that has been generated by stochastic (random/non-deterministic) processes 6/28/2014

Population and Sample CSE651C, B. Ramamurthy 7 Population is complete set of traces/data points  US population 314 Million, world population is 7 billion for example  All voters, all things Sample is a subset of the complete set (or population): how we select the sample introduces biases into the data See an example in Here out of the 314 Million US population, households are form the sample (monthly) Population  mathematical model  sample Lets look at a automobile data collection example: complaints in India and about India is that there are very few studies about these accidents… 6/28/2014

Population and Sample (contd.) CSE651C, B. Ramamurthy 8 Example: s sent by people in the Bosch in a year. Method 1: 1/10 of all s over the year randomly chosen Method 2: 1/10 of people randomly chosen; all their over the year Both are reasonable sample selection method for analysis. However estimations pdfs (probability distribution functions) of the s sent by a person for the two samples will be different. 6/28/2014

Big Data vs statistical inference CSE651C, B. Ramamurthy 9 Sample size N For statistical inference N < All For big data N == All 6/28/2014

What is you model? 6/28/2014 CSE651C, B. Ramamurthy 10 What is your data model? 1. Linear regression (lm): Understand the concept. Use Simpler package to explore lm. 2. Naïve Bayes and Bayesian classification 3. Classification vs clustering 4. Logistic regression: Computing the odds.

From the nutshell book 6/28/2014 CSE651C, B. Ramamurthy 11 A model is a concise way to describe a set of data, usually with a mathematical formula. Sometimes, the goal is to build a predictive model with training data to predict values based on other data. Other times, the goal is to build a descriptive model that helps you understand the data better.

Modeling CSE651C, B. Ramamurthy 12 6/28/2014

Probability Distributions CSE651C, B. Ramamurthy 13 Normal, uniform, Cauchy, t-, F-, Chi-square, exponential, Weibull, lognormal,.. They are know as continuous density functions Any random variable x or y can be assumed to have probability distribution p(x), if it maps it to a positive real number. For a probability density function, if we integrate the function to find the area under the curve it is 1, allowing it to be interpreted as probability. Further, joint distributions, conditional distribution.. 6/28/2014

Fitting a Model CSE651C, B. Ramamurthy 14 Fitting a model means estimating the parameters of the model: what distribution, what are the values of min, max, mean, stddev, etc. Don’t worry a statistical language R has built-in optimization algorithms that readily offer all these functionalities It involves algorithms such as maximum likelihood estimation (MLE) and optimization methods… Example: y = β1+β2 ∗  y = *x 6/28/2014

What if? 6/28/2014 CSE651C, B. Ramamurthy 15 The variable is not a continuous one as in linear regression? What if you want to determine the probability of an event (e.g. ABS activation ) happening given some “prior” probabilities? Ans: Naïve Bayes and Bayesian approaches What if you want to find the “odds” of an event (say, an engine failure) happening over not happening given their probabilities: Logistic regression There are many models for various situations… we will look into just these two above.

Summary CSE651C, B. Ramamurthy 16 An excellent tool supporting statistical inference is R R statistical language and the environment supporting it will be the second emerging technology and platform we consider in this course. We will examine R next We will also look into some machine learning (ML) approaches (algorithms) for clustering and classification. Then we will look into Naïve Bayes and logistic regression as two of the many approaches for analytics. 6/28/2014