Data Mining 2 (ex Análisis Inteligente de Datos y Data Mining) Lluís A. Belanche.

Slides:



Advertisements
Similar presentations
Introduction to Support Vector Machines (SVM)
Advertisements

ECG Signal processing (2)
Rule extraction in neural networks. A survey. Krzysztof Mossakowski Faculty of Mathematics and Information Science Warsaw University of Technology.
Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
Imbalanced data David Kauchak CS 451 – Fall 2013.
An Introduction of Support Vector Machine
Support Vector Machines and Kernels Adapted from slides by Tim Oates Cognition, Robotics, and Learning (CORAL) Lab University of Maryland Baltimore County.
DEVELOPMENT OF A METHOD FOR RELIABLE AND LOW COST PREDICTIVE MAINTENANCE Jacopo Cassina.
Huge Raw Data Cleaning Data Condensation Dimensionality Reduction Data Wrapping/ Description Machine Learning Classification Clustering Rule Generation.
WRSTA, 13 August, 2006 Rough Sets in Hybrid Intelligent Systems For Breast Cancer Detection By Aboul Ella Hassanien Cairo University, Faculty of Computer.
Chapter 4 DECISION SUPPORT AND ARTIFICIAL INTELLIGENCE
Evaluating Hypotheses
AI – CS364 Hybrid Intelligent Systems Overview of Hybrid Intelligent Systems 07 th November 2005 Dr Bogdan L. Vrusias
Intelligent Systems Group Emmanuel Fernandez Larry Mazlack Ali Minai (coordinator) Carla Purdy William Wee.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Part I: Classification and Bayesian Learning
Introduction to machine learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Tennessee Technological University1 The Scientific Importance of Big Data Xia Li Tennessee Technological University.
COMP3503 Intro to Inductive Modeling
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
10/6/2015 1Intelligent Systems and Soft Computing Lecture 0 What is Soft Computing.
CSC321: Neural Networks Lecture 12: Clustering Geoffrey Hinton.
Artificial Neural Networks. Applied Problems: Image, Sound, and Pattern recognition Decision making  Knowledge discovery  Context-Dependent Analysis.
Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos VC 14/15 – TP19 Neural Networks & SVMs Miguel Tavares.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
Constraint Satisfaction Problems (CSPs) CPSC 322 – CSP 1 Poole & Mackworth textbook: Sections § Lecturer: Alan Mackworth September 28, 2012.
Computational Intelligence II Lecturer: Professor Pekka Toivanen Exercises: Nina Rogelj
An Overview of Intrusion Detection Using Soft Computing Archana Sapkota Palden Lama CS591 Fall 2009.
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Learning from Observations Chapter 18 Through
Machine Learning.
Particle Filters.
Soft Computing A Gentle introduction Richard P. Simpson.
CS 478 – Tools for Machine Learning and Data Mining Linear and Logistic Regression (Adapted from various sources) (e.g., Luiz Pessoa PY 206 class at Brown.
Introduction to Machine Learning Supervised Learning 姓名 : 李政軒.
Week 1 - An Introduction to Machine Learning & Soft Computing
START OF DAY 5 Reading: Chap. 8. Support Vector Machine.
CS 478 – Tools for Machine Learning and Data Mining SVM.
Concept learning, Regression Adapted from slides from Alpaydin’s book and slides by Professor Doina Precup, Mcgill University.
Bing LiuCS Department, UIC1 Chapter 8: Semi-supervised learning.
Intelligent System Ming-Feng Yeh Department of Electrical Engineering Lunghwa University of Science and Technology Website:
Team Members Ming-Chun Chang Lungisa Matshoba Steven Preston Supervisors Dr James Gain Dr Patrick Marais.
What is Artificial Intelligence?
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall Perceptron Rule and Convergence Proof Capacity.
Data Mining and Decision Support
A field of study that encompasses computational techniques for performing tasks that require intelligence when performed by humans. Simulation of human.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 2 Nanjing University of Science & Technology.
Computational Intelligence: Methods and Applications Lecture 26 Density estimation, Expectation Maximization. Włodzisław Duch Dept. of Informatics, UMK.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Overfitting, Bias/Variance tradeoff. 2 Content of the presentation Bias and variance definitions Parameters that influence bias and variance Bias and.
FNA/Spring CENG 562 – Machine Learning. FNA/Spring Contact information Instructor: Dr. Ferda N. Alpaslan
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Inexact Reasoning 2 Session 10
Fuzzy Systems Simulation Session 5
CS 9633 Machine Learning Support Vector Machines
Soft Computing Basics Ms. Parminder Kaur.
Inexact Reasoning 2 Session 10
Soft Computing Introduction.
SOFT COMPUTING.
Introduction to Soft Computing
Overview of Machine Learning
Intelligent Systems and
Chapter 11 Practical Methodology
Genetic Algorithm Soft Computing: use of inexact t solution to compute hard task problems. Soft computing tolerant of imprecision, uncertainty, partial.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Presentation transcript:

Data Mining 2 (ex Análisis Inteligente de Datos y Data Mining) Lluís A. Belanche

/~belanche/docencia/aiddm/aiddm.html /~avellido/teaching/data_mining.htm

Contents of the course (hopefully) 1. Introduction & methodologies 2. Exploratory DM through visualization 3. Pattern recognition: introduction 4. Pattern recognition: the Gaussian case 5. Feature extraction 6. Feature selection & weighing 7. Error estimation 8. Linear methods are nice! 9. Probability in Data Mining 10. Latency, generativity, manifolds and all that 11. Application of GTM: from medicine to ecology 12. DM Case studies Sorry guys! … no fuzzy systems …

Error estimation

Feature extraction, selection and weighing have many uses

Linear classifiers are nice! (I)

Linear classifiers are nice! (II)  Transformation  (x) = [   (x),   (x), …  m (x) ] with x = [ x 1, x 2, …, x n ] Useful for “ascending” (m>n) or “descending” (m>n) with 0 < m,n < oo (integers) … an example?

Linear classifiers are nice! (III)  Nets  (x) = [   (x),   (x), …  m (x) ] with x = [ x 1, x 2, …, x n ] x     (x)

Utility This is a very powerful setting Let us suppose: r>s  increase in dimension increase in expressive power, ease the task for almost any learning machine r<s  decrease in dimension visualization, compactation, noise reduction, removal of useless information Contradictory !?

On intelligence … What is Intelligence? What is the function of Intelligence?  to ensure survival in nature What are the ingredients of intelligence? –Perceive in a changing world –Reason under partial truth –Plan & prioritize under uncertainty –Coordinate different simultaneous tasks –Learn under noisy experiences

“Generally, a car can be parked rather easily because the final position of the car is not specified exactly. It it were specified to within, say, a fraction of a millimeter and a few seconds of arc, it would take hours of maneuvering and precise measurements of distance and angular position to solve the problem.” Highhigh  High precision carries a high cost. Parking a Car (difficult or easy?)

Soft Computing Rough Sets Fuzzy Logic Neural Networks Evolutionary Algorithms Chaos & Fractals Belief Networks The primordial soup

What could MACHINE LEARNING possibly be? In the beginning, there was a set of examples … To exploit imprecision, uncertainty, robustness, data dependencies, learning and/or optimization ability, to achieve a working solution to a problem which is hard to solve. To find an exact (approximate) solution to an imprecisely (precisely) formulated problem.

 The challenge is to put these capabilities into use by devising methods of computation which lead to an acceptable solution at the lowest possible cost.  This should be the guiding principle So what is the aim?

Fuzzy Logic : the algorithms for dealing with imprecision and uncertainty Neural Networks : the machinery for learning and function approximation with noise Evolutionary Algorithms : the algorithms for adaptive search and optimization RS Rough Sets uncertainty arising from the granularity in the domain of discourse Different methods = different roles

Examples of soft computing TSP: 10 5 cities, –accuracy within 0.75%, 7 months –accuracy within 1%, 2 days Compare –“absoulute best for sure” with “very good with very high probability”

Are you one of the top guns? Consider … –Search space of size s –Draw N random samples –What is the probability p that at least one of them is in the top t ? Answer: p = 1 – (1-t/s) N Example: s= 10 12, N= , t=1.000  1 in !

On Algorithms what is worth? Problems Efficiency P Specialized algorithms: best performance for special problems Generic algorithms: good performance over a wide range of problems Specialized Algo. Generic Algorithms

Words are important ! What is a theory ? What is an algorithm ? What is an implementation ? What is a model ? What does “non-linear” mean ? What does “non-parametric” mean ?

Learning “Foreignia” (Poggio & Girosi’93) Can a machine learn to pronounce? 1. Do nothing and wait 2. Learn all the pronunciation rules 3. Memorize pronunciation examples 4. Pick a subset of pronunciation pairs and learn/memorize them 5. Pick subsets of pronunciation examples and develop a model explaining them

The problem of induction Classical problem in Philosophy Example: 1,2,3,4,5,? A more through example: JT

What are the conditions for successful learning? Training data (sufficiently) representative Principle of similarity Target function within capacity of the learner Non-dull learning algorithm Enough computational resources A correct (or close to) learning bias

And the Oscar goes to … The real problem is not whether machines think, but whether men do. B.F. Skinner, Contingencies of Reinforcement