Computational Intelligence

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

Computational Intelligence Winter Term 2009/10 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Computational Intelligence Winter Term 2013/14 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Perceptron Learning Rule
NEURAL NETWORKS Perceptron
Simple Neural Nets For Pattern Classification
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
The McCulloch-Pitts Neuron. Characteristics The activation of a McCulloch Pitts neuron is binary. Neurons are connected by directed weighted paths. A.
Artificial Neural Network
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
1 Introduction to Artificial Neural Networks Andrew L. Nelson Visiting Research Faculty University of South Florida.
Neurons, Neural Networks, and Learning 1. Human brain contains a massively interconnected net of (10 billion) neurons (cortical cells) Biological.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
2101INT – Principles of Intelligent Systems Lecture 10.
Establishing the Equivalence between Recurrent Neural Networks and Turing Machines. Ritesh Kumar Sinha(02d05005) Kumar Gaurav Bijay( )
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Intelligence Lecture No. 29 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Introduction to Artificial Intelligence (G51IAI) Dr Rong Qu Neural Networks.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
Artificial Neural Networks Chapter 4 Perceptron Gradient Descent Multilayer Networks Backpropagation Algorithm 1.
IE 585 History of Neural Networks & Introduction to Simple Learning Rules.
COSC 4426 AJ Boulay Julia Johnson Artificial Neural Networks: Introduction to Soft Computing (Textbook)
1 Perceptron as one Type of Linear Discriminants IntroductionIntroduction Design of Primitive UnitsDesign of Primitive Units PerceptronsPerceptrons.
Artificial Intelligence Methods Neural Networks Lecture 1 Rakesh K. Bissoondeeal Rakesh K.
Perceptrons Michael J. Watts
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Where are we? What’s left? HW 7 due on Wednesday Finish learning this week. Exam #4 next Monday Final Exam is a take-home handed out next Friday in class.
Computational Intelligence Winter Term 2015/16 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Neural networks.
Artificial Neural Networks
Learning with Perceptrons and Neural Networks
Artificial neural networks:
CS344: Introduction to Artificial Intelligence (associated lab: CS386)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Computational Intelligence
Computational Intelligence
Computational Intelligence
Computational Intelligence
CS623: Introduction to Computing with Neural Nets (lecture-2)
Chapter 3. Artificial Neural Networks - Introduction -
CSE (c) S. Tanimoto, 2004 Neural Networks
of the Artificial Neural Networks.
Perceptron as one Type of Linear Discriminants
Neural Network - 2 Mayank Vatsa
Computational Intelligence
Perceptrons Introduced in1957 by Rosenblatt
Computational Intelligence
Artificial Intelligence Lecture No. 28
CSE (c) S. Tanimoto, 2001 Neural Networks
CSE (c) S. Tanimoto, 2002 Neural Networks
Artificial Intelligence 12. Two Layer ANNs
Computational Intelligence
Computational Intelligence
ARTIFICIAL NEURAL networks.
Computational Intelligence
Artificial Neural Networks
CSE (c) S. Tanimoto, 2007 Neural Nets
Computational Intelligence
Introduction to Neural Network
David Kauchak CS158 – Spring 2019
Computational Intelligence
PYTHON Deep Learning Prof. Muhammad Saeed.
Computational Intelligence
Artificial Neural Network
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Computational Intelligence Winter Term 2016/17 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund

McCulloch Pitts Neuron (MCP) Minsky / Papert Perceptron (MPP) Plan for Today Organization (Lectures / Tutorials) Overview CI Introduction to ANN McCulloch Pitts Neuron (MCP) Minsky / Papert Perceptron (MPP) G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 2

Organizational Issues Who are you? either studying “Automation and Robotics” (Master of Science) Module “Optimization” or studying “Informatik” BSc-Modul “Einführung in die Computational Intelligence” Hauptdiplom-Wahlvorlesung (SPG 6 & 7) or … let me know! G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 3

Organizational Issues Who am I ? Günter Rudolph Fakultät für Informatik, LS 11 Guenter.Rudolph@tu-dortmund.de ← best way to contact me OH-14, R. 232 ← if you want to see me office hours: Tuesday, 10:30–11:30am and by appointment G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 4

Organizational Issues Lectures Wednesday 10:15-11:45 OH12, R. E.003, weekly Tutorials Thursday 14:15-15:45 OH16, R. 2.05, bi-weekly Friday 14:15-15:45 OH14, R. 1.04, bi-weekly Tutor Vanessa Volz, MSc, LS 11 Information http://ls11-www.cs.tu-dortmund.de/people/rudolph/ teaching/lectures/CI/WS2016-17/lecture.jsp Slides see web page Literature see web page either or G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 5

Organizational Issues Exams Effective since winter term 2016/17: written exam (not oral) ● Informatik, Diplom: Leistungsnachweis → Übungsschein ● Informatik, Diplom: Fachprüfung → written exam (90 min) ● Informatik, Bachelor: Module → written exam (90 min) ● Automation & Robotics, Master: Module → written exam (90 min) mandatory for registration to written exam: must pass tutorial G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 6

But what if something is unknown to me? • covered in the lecture Prerequisites Knowledge about • mathematics, • programming, • logic is helpful. But what if something is unknown to me? • covered in the lecture • pointers to literature ... and don‘t hesitate to ask! G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 7

Overview “Computational Intelligence“ What is CI ?  umbrella term for computational methods inspired by nature artifical neural networks evolutionary algorithms fuzzy systems swarm intelligence artificial immune systems growth processes in trees ... backbone new developments G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 8

Overview “Computational Intelligence“ term „computational intelligence“ made popular by John Bezdek (FL, USA) originally intended as a demarcation line  establish border between artificial and computational intelligence nowadays: blurring border our goals: know what CI methods are good for! know when refrain from CI methods! know why they work at all! know how to apply and adjust CI methods to your problem! G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 9

Introduction to Artificial Neural Networks Biological Prototype Neuron Information gathering (D) Information processing (C) Information propagation (A / S) human being: 1012 neurons electricity in mV range speed: 120 m / s axon (A) cell body (C) nucleus dendrite (D) synapse (S) G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 10

signal input signal processing signal output Introduction to Artificial Neural Networks Abstraction axon nucleus / cell body dendrites synapse … signal input signal processing signal output G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 11

Introduction to Artificial Neural Networks Model x1 function f x2 f(x1, x2, …, xn) … xn McCulloch-Pitts-Neuron 1943: xi  { 0, 1 } =: B f: Bn → B G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 12

Introduction to Artificial Neural Networks 1943: Warren McCulloch / Walter Pitts description of neurological networks → modell: McCulloch-Pitts-Neuron (MCP) basic idea: neuron is either active or inactive skills result from connecting neurons considered static networks (i.e. connections had been constructed and not learnt) G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 13

Introduction to Artificial Neural Networks McCulloch-Pitts-Neuron n binary input signals x1, …, xn threshold  > 0 ≥ 1 ... x1 x2 xn  = 1 boolean OR ≥ n ... x1 x2 xn  = n boolean AND  can be realized: G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 14

Introduction to Artificial Neural Networks McCulloch-Pitts-Neuron x1 y1 ≥ 0 NOT n binary input signals x1, …, xn threshold  > 0 in addition: m binary inhibitory signals y1, …, ym if at least one yj = 1, then output = 0 otherwise: sum of inputs ≥ threshold, then output = 1 else output = 0 G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 15

Introduction to Artificial Neural Networks Assumption: inputs also available in inverted form, i.e.  inverted inputs.  x1 + x2 ≥  x1 x2 ≥  Theorem: Every logical function F: Bn → B can be simulated with a two-layered McCulloch/Pitts net. Example: x1 x2 x3 x4 ≥ 3 ≥ 2 ≥ 1 G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 16

Introduction to Artificial Neural Networks Proof: (by construction) Every boolean function F can be transformed in disjunctive normal form  2 layers (AND - OR) Every clause gets a decoding neuron with  = n  output = 1 only if clause satisfied (AND gate) All outputs of decoding neurons are inputs of a neuron with  = 1 (OR gate) q.e.d. G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 17

Introduction to Artificial Neural Networks Generalization: inputs with weights ≥ 0,7 0,2 0,4 0,3 x1 x2 x3 fires 1 if 0,2 x1 + 0,4 x2 + 0,3 x3 ≥ 0,7 · 10 2 x1 + 4 x2 + 3 x3 ≥ 7  duplicate inputs! ≥ 7 x2 x3 x1  equivalent! G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 18

Introduction to Artificial Neural Networks Theorem: Weighted and unweighted MCP-nets are equivalent for weights  Q+. Proof: „“ N Let Multiplication with yields inequality with coefficients in N Duplicate input xi, such that we get ai b1 b2  bi-1 bi+1  bn inputs. Threshold  = a0 b1  bn „“ Set all weights to 1. q.e.d. G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 19

Introduction to Artificial Neural Networks Conclusion for MCP nets + feed-forward: able to compute any Boolean function + recursive: able to simulate DFA − very similar to conventional logical circuits − difficult to construct − no good learning algorithm available G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 20

Introduction to Artificial Neural Networks Perceptron (Rosenblatt 1958) → complex model → reduced by Minsky & Papert to what is „necessary“ → Minsky-Papert perceptron (MPP), 1969 → essential difference: x  [0,1]  R What can a single MPP do? Y N 1 isolation of x2 yields: Y N 1 Example:  1 Y N separating line separates R2 in 2 classes G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 21

→ MPP at least as powerful as MCP neuron! Introduction to Artificial Neural Networks = 0 = 1 AND 1 OR NAND NOR → MPP at least as powerful as MCP neuron! XOR 1 x1 x2 xor 1  0 <  w1, w2 ≥  > 0  w2 ≥  ?  w1 + w2 ≥ 2  w1 ≥   w1 + w2 <  contradiction! w1 x1 + w2 x2 ≥  G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 22

Introduction to Artificial Neural Networks 1969: Marvin Minsky / Seymor Papert book Perceptrons → analysis math. properties of perceptrons disillusioning result: perceptions fail to solve a number of trivial problems! XOR-Problem Parity-Problem Connectivity-Problem „conclusion“: All artificial neurons have this kind of weakness!  research in this field is a scientific dead end! consequence: research funding for ANN cut down extremely (~ 15 years) G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 23

Introduction to Artificial Neural Networks how to leave the „dead end“: Multilayer Perceptrons: x1 x2 2 1  realizes XOR Nonlinear separating functions: XOR 1 g(x1, x2) = 2x1 + 2x2 – 4x1x2 -1 with  = 0 g(0,0) = –1 g(0,1) = +1 g(1,0) = +1 g(1,1) = –1 G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 24

Introduction to Artificial Neural Networks How to obtain weights wi and threshold  ? as yet: by construction example: NAND-gate x1 x2 NAND 1  0 ≥   w2 ≥   w1 ≥   w1 + w2 <  requires solution of a system of linear inequalities ( P) (e.g.: w1 = w2 = -2,  = -3) now: by „learning“ / training G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 25

Introduction to Artificial Neural Networks Perceptron Learning Assumption: test examples with correct I/O behavior available Principle: choose initial weights in arbitrary manner feed in test pattern if output of perceptron wrong, then change weights goto (2) until correct output for all test paterns graphically: → translation and rotation of separating lines G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 26

Introduction to Artificial Neural Networks Example ≥0 x2 x1 1 w2 w1 - threshold as a weight: w = (, w1, w2)‘  suppose initial vector of weights is w(0) = (1, -1, 1)‘ G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 27

Introduction to Artificial Neural Networks P: set of positive examples → output 1 N: set of negative examples → output 0 Perceptron Learning threshold µ integrated in weights choose w0 at random, t = 0 choose arbitrary x  P  N if x  P and wt‘x > 0 then goto 2 if x  N and wt‘x ≤ 0 then goto 2 if x  P and wt‘x ≤ 0 then wt+1 = wt + x; t++; goto 2 if x  N and wt‘x > 0 then wt+1 = wt – x; t++; goto 2 stop? If I/O correct for all examples! I/O correct! let w‘x ≤ 0, should be > 0! (w+x)‘x = w‘x + x‘x > w‘ x let w‘x > 0, should be ≤ 0! (w–x)‘x = w‘x – x‘x < w‘ x remark: algorithm converges, is finite, worst case: exponential runtime G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 28

Introduction to Artificial Neural Networks We know what a single MPP can do. What can be achieved with many MPPs? Single MPP  separates plane in two half planes Many MPPs in 2 layers  can identify convex sets How?  2 layers! Convex? A B   a,b  X:  a + (1-) b  X for   (0,1) G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 29

Introduction to Artificial Neural Networks Single MPP  separates plane in two half planes Many MPPs in 2 layers  can identify convex sets Many MPPs in 3 layers  can identify arbitrary sets Many MPPs in > 3 layers  not really necessary! arbitrary sets: partitioning of nonconvex set in several convex sets two-layered subnet for each convex set feed outputs of two-layered subnets in OR gate (third layer) G. Rudolph: Computational Intelligence ▪ Winter Term 2016/17 30