Data Visualization STAT 890, STAT 442, CM 462

Slides:



Advertisements
Similar presentations
Introduction to Machine Learning BITS C464/BITS F464
Advertisements

Adjusting Active Basis Model by Regularized Logistic Regression
Godfather to the Singularity
SoLSTiCe Similarity of locally structured data in computer vision Université-Jean Monnet (Saint-Etienne) LIRIS (Lyon) (1/02/ ) Elisa Fromont,
Machine learning continued Image source:
Computer and Robot Vision I
Soft computing Lecture 6 Introduction to neural networks.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Introduction to Machine Learning Anjeli Singh Computer Science and Software Engineering April 28 th 2008.
Graph Based Semi- Supervised Learning Fei Wang Department of Statistical Science Cornell University.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
INTRODUCTION TO Machine Learning 3rd Edition
1. Introduction to Pattern Recognition and Machine Learning. Prof. A.L. Yuille. Dept. Statistics. UCLA. Stat 231. Fall 2004.
Introduction to machine learning
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
LLNL-PRES This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
AbstractAbstract Adjusting Active Basis Model by Regularized Logistic Regression Ruixun Zhang Peking University, Department of Statistics and Probability.
Data Mining. 2 Models Created by Data Mining Linear Equations Rules Clusters Graphs Tree Structures Recurrent Patterns.
Introduction Mohammad Beigi Department of Biomedical Engineering Isfahan University
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
TINONS1 Nonlinear SP and Pattern recognition
MACHINE LEARNING 張銘軒 譚恆力 1. OUTLINE OVERVIEW HOW DOSE THE MACHINE “ LEARN ” ? ADVANTAGE OF MACHINE LEARNING ALGORITHM TYPES  SUPERVISED.
Machine learning & category recognition Cordelia Schmid Jakob Verbeek.
© Negnevitsky, Pearson Education, Will neural network work for my problem? Will neural network work for my problem? Character recognition neural.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Machine Learning An Introduction. What is Learning?  Herbert Simon: “Learning is any process by which a system improves performance from experience.”
Data Mining Chapter 1 Introduction -- Basic Data Mining Tasks -- Related Concepts -- Data Mining Techniques.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Yazd University, Electrical and Computer Engineering Department Course Title: Advanced Software Engineering By: Mohammad Ali Zare Chahooki 1 Introduction.
Source-Selection-Free Transfer Learning
1 Learning from Shadows Dimensionality Reduction and its Application in Artificial Intelligence, Signal Processing and Robotics Ali Ghodsi Department of.
Different Features. Glasses vs. No Glasses Beard vs. No Beard.
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
Copyright © 2012, SAS Institute Inc. All rights reserved. ANALYTICS IN BIG DATA ERA ANALYTICS TECHNOLOGY AND ARCHITECTURE TO MANAGE VELOCITY AND VARIETY,
Machine Learning Extract from various presentations: University of Nebraska, Scott, Freund, Domingo, Hong,
An Efficient Greedy Method for Unsupervised Feature Selection
The Interplay Between Mathematics/Computation and Analytics Haesun Park Division of Computational Science and Engineering Georgia Institute of Technology.
Machine Learning Overview Tamara Berg CS 560 Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
Foundational Issues Machine Learning 726 Simon Fraser University.
Introduction to Machine Learning August, 2014 Vũ Việt Vũ Computer Engineering Division, Electronics Faculty Thai Nguyen University of Technology.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Machine learning & object recognition Cordelia Schmid Jakob Verbeek.
Introduction to Machine Learning, its potential usage in network area,
Brief Intro to Machine Learning CS539
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Who am I? Work in Probabilistic Machine Learning Like to teach 
Machine Learning overview Chapter 18, 21
Deep Learning Amin Sobhani.
Machine Learning overview Chapter 18, 21
Eick: Introduction Machine Learning
Intro to Machine Learning
Neural Networks for Machine Learning Lecture 1e Three types of learning Geoffrey Hinton with Nitish Srivastava Kevin Swersky.
Introductory Seminar on Research: Fall 2017
CH. 1: Introduction 1.1 What is Machine Learning Example:
Machine Learning Ali Ghodsi Department of Statistics
Unsupervised Learning and Autoencoders
What is Pattern Recognition?
Basic Intro Tutorial on Machine Learning and Data Mining
Finding Clusters within a Class to Improve Classification Accuracy
3.1.1 Introduction to Machine Learning
ITEC323 Lecture 1.
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Christoph F. Eick: A Gentle Introduction to Machine Learning
Machine Learning – a Probabilistic Perspective
Lecture 21: Machine Learning Overview AP Computer Science Principles
Azure Machine Learning
Lecture 9: Machine Learning Overview AP Computer Science Principles
Presentation transcript:

Data Visualization STAT 890, STAT 442, CM 462 Ali Ghodsi Department of Statistics School of Computer Science University of Waterloo aghodsib @uwaterloo.ca September 2006

Two Problems Classical Statistics Infer information from small data sets (Not enough data) Machine Learning Infer information from large data sets (Too many data)

are all closely related to the second problem. Other Names for ML Data mining, Applied statistics Adaptive (stochastic) signal processing Probabilistic planning or reasoning are all closely related to the second problem.

Applications Machine Learning is most useful when the structure of the task is not well understood but can be characterized by a dataset with strong statistical regularity. Search and recommendation (e.g. Google, Amazon) Automatic speech recognition and speaker verification Text parsing Face identification Tracking objects in video Financial prediction, fraud detection (e.g. credit cards) Medical diagnosis

Tasks Supervised Learning: given examples of inputs and corresponding desired outputs, predict outputs on future inputs. e.g.: classification, regression Unsupervised Learning: given only inputs, automatically discover representations, features, structure, etc. e.g.: clustering, dimensionality reduction, Feature extraction

Dimensionality Reduction Dimensionality: The number of measurements available for each item in a data set. The dimensionality of real world items is very high. For example: The dimensionality of a 600 by 600 image is 360,000. The Key to analyzing data is comparing these measurements to find relationships among this plethora of data points. Usually these measurements are highly redundant, and relationships among data points are predictable.

Dimensionality Reduction Knowing the value of a pixel in an image, it is easy to predict the value of nearby pixels since they tend to be similar. Knowing that the word “corporation” occurs often in articles about economics, but not very often in articles about art and poetry then it is easy to predict that it will not occur very often in articles about love. Although there are lots of measurements per item, there are far fewer that are likely to vary. Using a data set that only includes the items likely to vary allows humans to quickly and easily recognize changes in high dimensionality data.

Data Representation

Data Representation

Data Representation 1 0.5

2 by 103 644 by 103 644 by 2 -2.19 -0.02 -3.19 1.02 23 by 28 23 by 28 2 by 1 2 by 1

Arranging words: Each word was initially represented by a high-dimensional vector that counted the number of times it appeared in different encyclopedia articles. Words with similar contexts are collocated

Different Features

Glasses vs. No Glasses

Beard vs. No Beard

Beard Distinction

Glasses Distinction

Multiple-Attribute Metric

Embedding of sparse music similarity graph Platt, 2004

Reinforcement learning Mahadevan and Maggioini, 2005

Semi-supervised learning Use graph-based discretization of manifold to infer missing labels. Belkin & Niyogi, 2004; Zien et al, Eds., 2005 Build classifiers from bottom eigenvectors of graph Laplacian.

Learning correspondences How can we learn manifold structure that is shared across multiple data sets? c et al, 2003, 2005

Mapping and robot localization Bowling, Ghodsi, Wilkinson 2005 Ham, Lin, D.D. 2005

The Big Picture

Manifold and Hidden Variables

Reading Journals: Neural Computation, JMLR, ML, IEEE PAMI Conferences: NIPS, UAI, ICML, AI-STATS, IJCAI, IJCNN Vision: CVPR, ECCV, SIGGRAPH Speech: EuroSpeech, ICSLP, ICASSP Online: citesser, google Books: Elements of Statistical Learning, Hastie, Tibshirani, Friedman Learning from Data, Cherkassky, Mulier Machine Learning, Mitchell Neural Networks for pattern Recognition, Bishop Introduction to Graphical Models, Jordan et. al