A presentation to El Paso del Norte Software Association

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Slides from: Doug Gray, David Poole
Introduction to Neural Networks Computing
Artificial Neural Networks (1)
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Simple Neural Nets For Pattern Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
CIS 678 Artificial Intelligence problems deduction, reasoning knowledge representation planning learning natural language processing motion and manipulation.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Lecture 5: Automatic cluster detection Lecture 6: Artificial neural networks Lecture 7: Evaluation of discovered knowledge Brief introduction to lectures.
Introduction to Neural Network Justin Jansen December 9 th 2002.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Chapter 6: Multilayer Neural Networks
Part I: Classification and Bayesian Learning
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Chapter 5 Data mining : A Closer Look.
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Least-Mean-Square Training of Cluster-Weighted-Modeling National Taiwan University Department of Computer Science and Information Engineering.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Using Neural Networks to Predict Claim Duration in the Presence of Right Censoring and Covariates David Speights Senior Research Statistician HNC Insurance.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Artificial Neural Networks An Introduction. What is a Neural Network? A human Brain A porpoise brain The brain in a living creature A computer program.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
An Introduction to Support Vector Machine (SVM)
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
EEE502 Pattern Recognition
Data Mining and Decision Support
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
NEURONAL NETWORKS AND CONNECTIONIST (PDP) MODELS Thorndike’s “Law of Effect” (1920’s) –Reward strengthens connections for operant response Hebb’s “reverberatory.
INTRODUCTION TO NEURAL NETWORKS 2 A new sort of computer What are (everyday) computer systems good at... and not so good at? Good at..Not so good at..
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Big data classification using neural network
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Semi-Supervised Clustering
Multilayer Perceptrons
an introduction to: Deep Learning
DATA MINING © Prentice Hall.
第 3 章 神经网络.
Introductory Seminar on Research: Fall 2017
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
CH. 1: Introduction 1.1 What is Machine Learning Example:
LECTURE 28: NEURAL NETWORKS
Dipartimento di Ingegneria «Enzo Ferrari»,
Machine Learning Today: Reading: Maria Florina Balcan
Artificial Intelligence Chapter 3 Neural Networks
Perceptron as one Type of Linear Discriminants
Introduction to Neural Networks And Their Applications - Basics
Artificial Intelligence Chapter 3 Neural Networks
LECTURE 28: NEURAL NETWORKS
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Support Vector Machines and Kernels
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Artificial Intelligence Chapter 3 Neural Networks
Neural networks (1) Traditional multi-layer perceptrons
Artificial Intelligence Chapter 3 Neural Networks
III. Introduction to Neural Networks And Their Applications - Basics
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
Introduction to Neural Network
Machine Learning in Business John C. Hull
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

Business Data Solution Using Clustering, Linear Programming, and Neural Net A presentation to El Paso del Norte Software Association Somnath (Shom) Mukhopadhyay Information and Decision Sciences Department The University of Texas at El Paso August 27th 2003

Outline of Presentation Data Mining Definition Introduction of Neural Net - Physiological flavor - General framework - Classes of PDP models - Sigma-PI units - Conclusion

Outline of Presentation (Continued) Examples of real-world application problems Organization of theoretical concepts - Three methods used for classification - A new LP based method for classification problem. - Application to a fictitious problem with four classes. - Comparing LP method results with the results from a neural network method - Q&A

Data Mining - definition - Exploring relationships in large amount of data - Should generalize - Should be empirically validated Examples - Customer Relationship Management (CRM) - Credit Scoring - Clinical decision support

PDP Models and Brain Physiological Flavor Representation and Learning in PDP models Origins of PDP - Jackson (1869) and Luria (1966) - Hebb (1950) - Rosenblatt (1959) - Grossberg (1970) - Rumelhart (1977)

General Framework for PDP A set of processing units A state of activation An output function of each unit A pattern of connectivity among units A propagation rule An activation rule A learning rule An operating environment

The Basic Components of a PDP system

Classes of PDP models Simple Linear Models Linear Threshold Units Brain State in a Box (BSB) by J. A. Anderson Thermodynamic models Grossberg Connectionist modeling

Sigma-PI Units

A few real-world applications of interest to organizations and individuals Breast cancer detection Heart disease diagnosis Enemy sub-marine detection Mortgage delinquency prediction Stock market prediction Japanese Character recognition and conversion

What is classification? Identification of a set of certain mutually exclusive classes Identify a set of meaningful attributes that discriminate among the classes Illustrations Using a meaningful set of attributes, can we differentiate between frequent and infrequent occurrence?

Decision Boundaries of a typical classification problem

Three Methods for Classification Identifying decision boundaries for each class region Linear discriminant (Glover at al., 1988) Linear programming (Roy and Mukhopadhyay, 1991) Neural Networks (Rumelhart, 1986)

A new LP based method for classification problem Step 1. Identify and discard outliers using Clustering Step 2. Form decision boundaries for each class region by using LP

Step 2: Form Decision Boundaries Development of Boundary Functions Use convex functions to calibrate the boundary. One example function: f(x) = ai Xi + bi Xi2 +  cij Xi Xj + d   where j = i + 1

Step 2: Form Decision Boundaries (Contd.) One instance of the general function. fA(x) = a1 X1 + a2 X2 + b1 X12 + b2 X22 + d

Step 2: Form Decision Boundaries (Contd.) LP formulation of the previous problem instance Minimize e s.t. fA(x1) >= e … fA(x8) >= e fA(x9) <= -e ... fA(x18) <= -e e>= a small positive constant. Minimize e s.t. a2 + b2 + d >= e for pattern x1 a1 + b1 + d >= e for pattern x2 - a2 + b2 + d >= e for pattern x3 - a1 + b1 + d >= e for pattern x4 …. a1 + a2 + b1 + b2 + d <= - e for pattern x15 a1 - a2 + b1 + b2 + d <= - e for pattern x16 - a1 - a2 + b1 + b2 + d <= - e for pattern x17 - a1 + a2 + b1 + b2 + d <= - e for pattern x18 e>= a small positive constant.

Step 2: Form Decision Boundaries (Contd.) Solution of this LP formulation gives decision boundaries. Specifically we get, a1 = 0, a2 = 0, b1 = -1, b2 = -1, d = 1+e   Therefore, the boundary function fA(x) = a1 X1 + a2 X2 + b1 X12 + b2 X22 + d translates into: fA(x) = 1 - X12 - X22 + e

Step 2: Form Decision Boundaries (Contd.) Putting this result into picture we have the following decision boundary:

Step 2: Form Multiple Decision Boundaries A class does not have to be neatly packed within one boundary. For problems requiring multiple decision boundaries, the algorithm can find multiple disjointed regions for the same class. For example, a class called “corner seats” in a soccer stadium is scattered into four disjointed regions.

An example of a decision space of a fictitious problem (It has four classes: A, B, C, D)

Decision Boundary Identification Process for Class D only

Six Decision Boundaries found for Class B

Constructing MLP from masks Masking functions put on a network to exploit parallelism.

Neural Networks Method for Classification develops non-linear functions to associate inputs with outputs no assumptions about distribution of data handles missing data well (graceful degradation) Supervised neural networks Estimating and testing the model Construct a training sample and a holdout sample Estimate model parameters using training sample Test the estimated model’s classification ability using holdout sample

Comparison between LP and NN performance for three real-world problem   Problem Test Error Rate (%) Total Number of Parameters Trained LP NN 1. Breast Cancer 1.7 2.96 19 990 2. Heart Disease 18.38 36.36 27 900 3. Submarine Detection 9.62 N/A 61  

Future Research - Autonomous Learning: learn without outside interventions does class dependent feature selection derives simple if-then type classification rules that humans can understand develops non-linear functions to associate inputs with outputs

Q & A Thank you.