For Wednesday Read ch. 20, sections 1, 2, 5, and 7 No homework.

Slides:



Advertisements
Similar presentations
Slides from: Doug Gray, David Poole
Advertisements

Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
For Wednesday Read chapter 19, sections 1-3 No homework.
Kostas Kontogiannis E&CE
Artificial Neural Networks
Machine Learning Neural Networks
Simple Neural Nets For Pattern Classification
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 21 Jim Martin.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
PERCEPTRON. Chapter 3: The Basic Neuron  The structure of the brain can be viewed as a highly interconnected network of relatively simple processing.
20.5 Nerual Networks Thanks: Professors Frank Hoffmann and Jiawei Han, and Russell and Norvig.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Artificial Neural Networks
Lecture 4 Neural Networks ICS 273A UC Irvine Instructor: Max Welling Read chapter 4.
Data Mining with Neural Networks (HK: Chapter 7.5)
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 22 Jim Martin.
CS 4700: Foundations of Artificial Intelligence
INTRODUCTION TO ARTIFICIAL INTELLIGENCE
For Monday Read chapter 22, sections 1-3 No homework.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
CISC 4631 Data Mining Lecture 11: Neural Networks.
For Friday Read chapter 22 Program 4 due. Program 4 Any questions?
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Neural Networks Chapter 18.7
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Networks
1 CS 343: Artificial Intelligence Neural Networks Raymond J. Mooney University of Texas at Austin.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
1 Machine Learning Neural Networks Neural Networks NN belong to the category of trained classification models The learned classification model is an.
Artificial Neural Network Yalong Li Some slides are from _24_2011_ann.pdf.
Theory Revision Chris Murphy. The Problem Sometimes we: – Have theories for existing data that do not match new data – Do not want to repeat learning.
For Friday Read chapter 18, section 7 Program 3 due.
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
For Thursday No reading or homework Exam 2 Take-home due Tuesday Read chapter 22 for next Tuesday.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
For Monday Read Chapter 22, sections 4-6 No written homework.
For Tuesday Read chapter 18, section 7 Read chapter 19, sections 1-2 and 5 Homework: –Chapter 18, exercises 6 and 11.
For Wednesday No reading No homework. Exam 2 Friday. Will cover material through chapter 18. Take home is due Friday.
For Wednesday No reading Homework: –Chapter 18, exercise 25 (a & b)
For Monday Finish chapter 19 No homework. Program 4 Any questions?
For Wednesday Read 20.4 Lots of interesting stuff in chapter 20, but we don’t have time to cover it all.
For Monday Finish chapter 19 Take-home exam due. Program 4 Any questions?
CS 478 – Tools for Machine Learning and Data Mining Perceptron.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 CS 391L: Machine Learning Neural Networks Raymond J. Mooney University of Texas at Austin.
For Friday No reading Take home exam due Exam 2. For Monday Read chapter 22, sections 1-3 FOIL exercise due.
1 Machine Learning Neural Networks Tomorrow no lesson! (will recover on Wednesday 11, WEKA !!)
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
For Friday Read ch. 20, sections 1-3 Program 4 due.
For Monday Read Chapter 18, sections Homework: –Chapter 18, exercise 6.
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Perceptrons Michael J. Watts
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Chap. 10 Learning Sets of Rules 박성배 서울대학교 컴퓨터공학과.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
CS 388: Natural Language Processing: Neural Networks
Fall 2004 Perceptron CS478 - Machine Learning.
Learning with Perceptrons and Neural Networks
Artificial Neural Networks
Real Neurons Cell structures Cell body Dendrites Axon
Perceptron as one Type of Linear Discriminants
Rule Learning Hankui Zhuo April 28, 2018.
Artificial Intelligence Lecture No. 28
David Kauchak CS158 – Spring 2019
Presentation transcript:

For Wednesday Read ch. 20, sections 1, 2, 5, and 7 No homework

Program 4 Any questions?

Learning mini-project Worth 2 homeworks Due next Monday Foil6 is available in /home/mecalif/public/itk340/foil A manual and sample data files are there as well. Create a data file that will allow FOIL to learn rules for a sister/2 relation from background relations of parent/2, male/1, and female/1. You can look in the prolog folder of my 327 folder for sample data if you like. Electronically submit your data file—which should be named sister.d, and turn in a hard copy of the rules FOIL learns.

Inductive Logic Programming Representation is Horn clauses Builds rules using background predicates Rules are potentially much more expressive than attribute-value representations

Example Results Rules for family relations from data of primitive or related predicates. uncle(A,B) :­ brother(A,C), parent(C,B). uncle(A,B) :­ husband(A,C), sister(C,D), parent(D,B). Recursive list programs. member(X,[X | Y]). member(X, [Y | Z]) :­ member(X, Z).

ILP Goal is to induce a Horn­clause definition for some target predicate P given definitions of background predicates Q i. Goal is to find a syntactically simple definition D for P such that given background predicate definitions B –For every positive example p i : D  B |= p –For every negative example n i : D  B |/= n Background definitions are either provided –Extensionally: List of ground tuples satisfying the predicate. –Intensionally: Prolog definition of the predicate.

Sequential Covering Algorithm Let P be the set of positive examples Until P is empty do Learn a rule R that covers a large number of positives without covering any negatives. Add R to the list of learned rules. Remove positives covered by R from P

This is just an instance of the greedy algorithm for minimum set covering and does not guarantee that a minimum number of rules is learned but tends to learn a reasonably small rule set. Minimum set covering is an NP­hard problem and the greedy algorithm is a common approximation algorithm. There are several ways to learn a single rule used in various methods.

Strategies for Learning a Single Rule Top­Down (General to Specific): –Start with the most general (empty) rule. –Repeatedly add feature constraints that eliminate negatives while retaining positives. –Stop when only positives are covered. Bottom­Up (Specific to General): –Start with a most specific rule (complete description of a single instance). –Repeatedly eliminate feature constraints in order to cover more positive examples. –Stop when further generalization results in covering negatives.

FOIL Basic top­down sequential covering algorithm adapted for Prolog clauses. Background provided extensionally. Initialize clause for target predicate P to P(X 1,...X r ) :­. Possible specializations of a clause include adding all possible literals: –Q i (V 1,...V r ) –not(Q i (V 1,...V r )) –X i = X j –not(X i = X ) where X's are variables in the existing clause, at least one of V 1,...V r is an existing variable, others can be new. Allow recursive literals if not cause infinite regress.

Foil Input Data Consider example of finding a path in a directed acyclic graph. Intended Clause: path(X,Y) :­ edge(X,Y). path(X,Y) :­ edge(X,Z), path (Z,Y). Examples edge: {,,,,, } path: {,,,,,,,,, } Negative examples of the target predicate can be provided directly or indirectly produced using a closed world assumption. Every pair not in positive tuples for path.

Example Induction + : {,,,,,,,,, } - : {,,,,,,,,,,,,,,,, } Start with empty rule: path(X,Y) :­. Among others, consider adding literal edge(X,Y) (also consider edge(Y,X), edge(X,Z), edge(Z,X), path(Y,X), path(X,Z), path(Z,X), X=Y, and negations) 6 positive tuples and NO negative tuples covered. Create “base case” and remove covered examples: path(X,Y) :­ edge(X,Y).

+ : {,,, } - : {,,,,,,,,,,,,,,,,, } Start with new empty rule: path(X,Y) :­. Consider literal edge(X,Z) (among others...) 4 remaining positives satisfy it but so do 10 of 20 negatives Current rule: path(x,y) :­ edge(X,Z). Consider literal path(Z,Y) (as well as edge(X,Y), edge(Y,Z), edge(X,Z), path(Z,X), etc....) No negatives covered, complete clause. path(X,Y) :­ edge(X,Z), path(Z,Y). New clause actually covers all remaining positive tuples of path, so definition is complete.

Picking the Best Literal Based on information gain (similar to ID3). |p|*(log 2 (|p| /(|p|+|n|)) - log 2 (|P| /(|P|+|N|))) P is number of positives before adding literal L N is number of negatives before adding literal L p is number of positives after adding literal L n is number of negatives after adding literal L Given n predicates of arity m there are O(n2 m ) possible literals to chose from, so branching factor can be quite large.

Other Approaches Golem CHILL Foidl Bufoidl

Domains Any kind of concept learning where background knowledge is useful. Natural Language Processing Planning Chemistry and biology –DNA –Protein structure

Why Neural Networks?

Analogy to biological systems, the best examples we have of robust learning systems. Models of biological systems allowing us to understand how they learn and adapt. Massive parallelism that allows for computational efficiency. Graceful degradation due to distributed represent- ations that spread knowledge representation over large numbers of computational units. Intelligent behavior is an emergent property from large numbers of simple units rather than resulting from explicit symbolically encoded rules.

Neural Speed Constraints Neuron “switching time” is on the order of milliseconds compared to nanoseconds for current transistors. A factor of a million difference in speed. However, biological systems can perform significant cognitive tasks (vision, language understanding) in seconds or tenths of seconds.

What That Means Therefore, there is only time for about a hundred serial steps needed to perform such tasks. Even with limited abilties, current AI systems require orders of magnitude more serial steps. Human brain has approximately neurons each connected on average to 10 4 others, therefore must exploit massive parallelism.

Real Neurons Cells forming the basis of neural tissue –Cell body –Dendrites –Axon –Syntaptic terminals The electrical potential across the cell membrane exhibits spikes called action potentials. Originating in the cell body, this spike travels down the axon and causes chemical neuro- transmitters to be released at syntaptic terminals. This chemical difuses across the synapse into dendrites of neighboring cells.

Real Neurons (cont) Synapses can be excitory or inhibitory. Size of synaptic terminal influences strength of connection. Cells “add up” the incoming chemical messages from all neighboring cells and if the net positive influence exceeds a threshold, they “fire” and emit an action potential.

Model Neuron (Linear Threshold Unit) Neuron modelled by a unit (j) connected by weights, w ji, to other units (i): Net input to a unit is defined as: net j =  w ji * o i Output of a unit is a threshold function on the net input: –1 if net j > T j –0 otherwise

Neural Computation McCollough and Pitts (1943) show how linear threshold units can be used to compute logical functions. Can build basic logic gates –AND: Let all w ji be (T j /n)+  where n = number of inputs –OR: Let all w ji be T j +  –NOT: Let one input be a constant 1 with weight T j +e and the input to be inverted have weight ­T j

Neural Computation (cont) Can build arbitrary logic circuits, finite­state machines, and computers given these basis gates. Given negated inputs, two layers of linear threshold units can specify any boolean function using a two­layer AND­OR network.

Learning Hebb (1949) suggested if two units are both active (firing) then the weight between them should increase: w ji = w ji +  o j o i –  is a constant called the learning rate –Supported by physiological evidence

Alternate Learning Rule Rosenblatt (1959) suggested that if a target output value is provided for a single neuron with fixed inputs, can incrementally change weights to learn to produce these outputs using the perceptron learning rule. –Assumes binary valued input/outputs –Assumes a single linear threshold unit. –Assumes input features are detected by fixed networks.

Perceptron Learning Rule If the target output for output unit j is t j w ji = w ji +  (t j - o j )o i Equivalent to the intuitive rules: –If output is correct, don't change the weights –If output is low (o j = 0, t j =1), increment weights for all inputs which are 1. –If output is high (o j = 1, t j =0), decrement weights for all inputs which are 1. Must also adjust threshold: T j = T j +  (t j - o j ) or equivalently assume there is a weight w j0 = -T j for an extra input unit 0 that has constant output o 0 =1 and that the threshold is always 0.

Perceptron Learning Algorithm Repeatedly iterate through examples adjusting weights according to the perceptron learning rule until all outputs are correct Initialize the weights to all zero (or randomly) Until outputs for all training examples are correct For each training example, e, do Compute the current output o j Compare it to the target t j and update the weights according to the perceptron learning rule.

Algorithm Notes Each execution of the outer loop is called an epoch. If the output is considered as concept membership and inputs as binary input features, then easily applied to concept learning problems. For multiple category problems, learn a separate perceptron for each category and assign to the class whose perceptron most exceeds its threshold. When will this algorithm terminate (converge) ??

Representational Limitations Perceptrons can only represent linear threshold functions and can therefore only learn data which is linearly separable (positive and negative examples are separable by a hyperplane in n­dimensional space) Cannot represent exclusive­or (xor)

Perceptron Learnability System obviously cannot learn what it cannot represent. Minsky and Papert(1969) demonstrated that many functions like parity (n­input generalization of xor) could not be represented. In visual pattern recognition, assumed that input features are local and extract feature within a fixed radius. In which case no input features support learning –Symmetry –Connectivity These limitations discouraged subsequent research on neural networks.

Perceptron Convergence and Cycling Theorems Perceptron Convergence Theorem: If there are a set of weights that are consistent with the training data (i.e. the data is linearly separable), the perceptron learning algorithm will converge (Minsky & Papert, 1969). Perceptron Cycling Theorem: If the training data is not linearly separable, the Perceptron learning algorithm will eventually repeat the same set of weights and threshold at the end of some epoch and therefore enter an infinite loop.

Perceptron Learning as Hill Climbing The search space for Perceptron learning is the space of possible values for the weights (and threshold). The evaluation metric is the error these weights produce when used to classify the training examples. The perceptron learning algorithm performs a form of hill­ climbing (gradient descent), at each point altering the weights slightly in a direction to help minimize this error. Perceptron convergence theorem guarantees that for the linearly separable case there is only one local minimum and the space is well behaved.

Perceptron Performance Can represent and learn conjunctive concepts and M­of­N concepts (true if any M of a set of N selected binary features are true). Although simple and restrictive, this high­ bias algorithm performs quite well on many realistic problems. However, the representational restriction is limiting in many applications.