1 CSC 9010 Spring 2011. Paula Matuszek CSC 9010 ANN Lab Paula Matuszek Spring, 2011.

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
Neural networks Introduction Fitting neural networks
Introduction to Neural Networks Computing
Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
CS0007: Introduction to Computer Programming Console Output, Variables, Literals, and Introduction to Type.
Chapter 3 Producing Data 1. During most of this semester we go about statistics as if we already have data to work with. This is okay, but a little misleading.
Artificial Neural Networks
Common Factor Analysis “World View” of PC vs. CF Choosing between PC and CF PAF -- most common kind of CF Communality & Communality Estimation Common Factor.
Data Mining Techniques Outline
Connectionist models. Connectionist Models Motivated by Brain rather than Mind –A large number of very simple processing elements –A large number of weighted.
Neural Networks. R & G Chapter Feed-Forward Neural Networks otherwise known as The Multi-layer Perceptron or The Back-Propagation Neural Network.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
October 5, 2010Neural Networks Lecture 9: Applying Backpropagation 1 K-Class Classification Problem Let us denote the k-th class by C k, with n k exemplars.
October 7, 2010Neural Networks Lecture 10: Setting Backpropagation Parameters 1 Creating Data Representations On the other hand, sets of orthogonal vectors.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Chapter 5 Data mining : A Closer Look.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Chapter 7 Artificial Neural Networks
Artificial Intelligence (AI) Addition to the lecture 11.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Classification Part 3: Artificial Neural Networks
Integrating Neural Network and Genetic Algorithm to Solve Function Approximation Combined with Optimization Problem Term presentation for CSC7333 Machine.
Artificial Neural Network Theory and Application Ashish Venugopal Sriram Gollapalli Ulas Bardak.
An Introduction to Artificial Intelligence and Knowledge Engineering N. Kasabov, Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering,
Instructor: Chris Trenkov Hands-on Course Python for Absolute Beginners (Spring 2015) Class #002 (January 17, 2015)
Generic Approaches to Model Validation Presented at Growth Model User’s Group August 10, 2005 David K. Walters.
Universit at Dortmund, LS VIII
1 CSC 8520 Spring Paula Matuszek Kinds of Machine Learning Machine learning techniques can be grouped into several categories, in several ways: –What.
Chapter Twenty-ThreeModern Programming Languages1 Formal Semantics.
©2003 Paula Matuszek CSC 9010: Text Mining Applications Document Summarization Dr. Paula Matuszek (610)
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 9: Ways of speeding up the learning and preventing overfitting Geoffrey Hinton.
Applying Neural Networks Michael J. Watts
CSC Intro. to Computing Lecture 22: Artificial Intelligence.
Java Data Types Assignment and Simple Arithmetic.
Soft Computing Lecture 19 Part 2 Hybrid Intelligent Systems.
1 CSC 4510, Spring © Paula Matuszek CSC 4510 Support Vector Machines (SVMs)
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Perceptrons Gary Cottrell. Cognitive Science Summer School 2 Perceptrons: A bit of history Frank Rosenblatt studied a simple version of a neural net called.
Copyright Paula Matuszek Kinds of Machine Learning.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Chapter 13 Artificial Intelligence. Artificial Intelligence – Figure 13.1 The Turing Test.
Computable Functions.
Applying Neural Networks
Formatting Output.
A Simple Artificial Neuron
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
K Nearest Neighbors and Instance-based methods
Neural Networks A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Data Mining Lecture 11.
FUNDAMENTAL CONCEPT OF ARTIFICIAL NETWORKS
Simple Learning: Hebbian Learning and the Delta Rule
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Artificial Intelligence 12. Two Layer ANNs
Logistic Regression Chapter 7.
Evaluating Classifiers
Arrays.
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Outline Announcement Neural networks Perceptrons - continued
Data Mining CSCI 307, Spring 2019 Lecture 6
Presentation transcript:

1 CSC 9010 Spring Paula Matuszek CSC 9010 ANN Lab Paula Matuszek Spring, 2011

2 CSC 9010 Spring Paula Matuszek Knowledge-Based Systems and Artificial Neural Nets Knowledge Representation means: –Capturing human knowledge –In a form computer can reason about Is this relevant to ANNs? Consider the question we’ve been modeling all semester: –what class should I take? What would be involved in representing this as a neural net?

3 CSC 9010 Spring Paula Matuszek Knowledge for ANNs What human knowledge do we still need to capture? –Input variables –Output variables –Training and test cases Let’s look at some. –

4 CSC 9010 Spring Paula Matuszek Example from AISpace Mail –Examples –Set properties –Initialize the parameters –Solve –How do we use it? Calculate output

5 CSC 9010 Spring Paula Matuszek Our “Which class to take” problem Inputs? Outputs? Sample data

6 CSC 9010 Spring Paula Matuszek Some Examples Example 1:Example 1 –3 inputs, 1 output, all binary Example 2:Example 2 –same inputs, output inverted

7 CSC 9010 Spring Paula Matuszek Getting the right inputs Example 3 –Same inputs as 1 and 2 –Same output as 1 –Outcomes reversed for half the cases

8 CSC 9010 Spring Paula Matuszek Getting the right inputs Example 3 –Same inputs as 1 and 2 –Same output as 1 –Outcomes reversed for half the cases Network is not converging The output here cannot be predicted from these inputs. Whatever is determining whether to take the class, we haven’t captured it

9 CSC 9010 Spring Paula Matuszek Representing non-numeric values Example 4 –Required is represented as “yes” or “no”

10 CSC 9010 Spring Paula Matuszek Representing non-numeric values Example 4 –Required is represented as “yes” or “no” Actual model still uses 1 and 0; transformation is done by applet.

11 CSC 9010 Spring Paula Matuszek More non-numeric values Example 5 Workload is low, med, high: text values but they can be ordered.

12 CSC 9010 Spring Paula Matuszek More non-numeric values Example 5 Workload is low, med, high: text values but they can be ordered. Applet asks us to assign values. 1, 0.5, 0 is typical.

13 CSC 9010 Spring Paula Matuszek Unordered values Example 6 –Input variables here include professor –Non-numeric, can’t be ordered.

14 CSC 9010 Spring Paula Matuszek Unordered values Example 6 –Input variables here include professor –Non-numeric, can’t be ordered. Still need numeric values Solution is to treat n possible values as n separate binary values Again, applet does this for us

15 CSC 9010 Spring Paula Matuszek Variables with more values Example 7 –GPA and number of classes taken are integer values Takes considerably longer to solve Looks for a while like it’s not converging

16 CSC 9010 Spring Paula Matuszek And Reals Example 8 –GPA is a real. Takes about 20,000 steps to converge

17 CSC 9010 Spring Paula Matuszek And multiple outputs Small Car database from AIspace For any given input case, you will get a value for each possible outcome. Typical for, for instance, character recognition.

18 CSC 9010 Spring Paula Matuszek Training and Test Cases The basic training approach will fit the training data as closely as possible. But we really want something that will generalize to other cases This is why we have test cases. –The training cases are used to compute the weights –The test cases tell us how well they generalize Both training and test cases should represent the overall population as well as possible.

19 CSC 9010 Spring Paula Matuszek Representative Training Cases Example 9 –Training cases and test cases are similar

20 CSC 9010 Spring Paula Matuszek Representative Training Cases Example 9 –Training cases and test cases are similar –(actually identical...) Training error and test error are comparable

21 CSC 9010 Spring Paula Matuszek Non-Representative Training Cases Example 10 –Training cases and test cases represent different circumstances –We’ve missed including any cases involving Lee in the training Training error goes down, but test error goes up. In reality these are probably bad training AND test cases; neither seems representative.

22 CSC 9010 Spring Paula Matuszek So: Getting a good ANN still involves understanding your domain and capturing knowledge about it –choosing the right inputs and outputs –choosing representative training and test sets –Beware “convenient” training sets You can represent any kind of variable: numeric or not, ordered or not. Not every set of variables and training cases will produce something that can be trained.

23 CSC 9010 Spring Paula Matuszek Once it’s trained... When your ANN is trained, you can feed it a specific set of inputs and get one or more outputs. These outputs are typically interpreted as some decision: –take the class –this is probably a “5” The network itself is black box. If the situation changes the ANN should be retrained –new variables –new values for some variables –new patterns of cases

24 CSC 9010 Spring Paula Matuszek One last note These have all been simple cases, as examples Most of my examples could in fact be predicted much more easily and cleanly with a decision tree, or even a couple of IF statements A more typical use for any connectionist system has many more inputs and many more training cases