1 Algorithmic Networks & Optimization Maastricht, October 2009 Ronald L. Westra, Department of Mathematics Maastricht University.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

NEU Neural Computing MSc Natural Computation Department of Computer Science University of York.
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Artificial Neural Networks - Introduction -
Artificial Neural Networks - Introduction -
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Artificial Intelligence (CS 461D)
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Simple Neural Nets For Pattern Classification
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
September 30, 2010Neural Networks Lecture 8: Backpropagation Learning 1 Sigmoidal Neurons In backpropagation networks, we typically choose  = 1 and 
Neural Networks Marco Loog.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Chapter Seven The Network Approach: Mind as a Web.
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Dr.-Ing. Erwin Sitompul President University Lecture 1 Introduction to Neural Networks and Fuzzy Logic President UniversityErwin SitompulNNFL 1/1
Artificial Neural Network
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Neural Networks
Functional Brain Signal Processing: EEG & fMRI Lesson 8 Kaushik Majumdar Indian Statistical Institute Bangalore Center M.Tech.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
1 The Identification of Scale-Free Gene-Protein Networks Ronald Westra Department of Mathematics Maastricht University.
Artificial Neural Networks An Overview and Analysis.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
1 Pattern storage in gene-protein networks Ronald Westra Department of Mathematics Maastricht University.
NEURAL NETWORKS FOR DATA MINING
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Gap filling of eddy fluxes with artificial neural networks
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Akram Bitar and Larry Manevitz Department of Computer Science
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Lecture 5 Neural Control
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
Chapter 8: Adaptive Networks
Hazırlayan NEURAL NETWORKS Backpropagation Network PROF. DR. YUSUF OYSAL.
CHEE825 Fall 2005J. McLellan1 Nonlinear Empirical Models.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Chapter 6 Neural Network.
Neural Networks Si Wu Dept. of Informatics PEV III 5c7 Spring 2008.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Neural Network Architecture Session 2
Artificial Intelligence (CS 370D)
Dr. Unnikrishnan P.C. Professor, EEE
Artificial Neural Network & Backpropagation Algorithm
of the Artificial Neural Networks.
Introduction to Neural Networks and Fuzzy Logic
Artificial Intelligence Lecture No. 28
Introduction to Neural Networks and Fuzzy Logic
Ch4: Backpropagation (BP)
Computer Vision Lecture 19: Object Recognition III
The Network Approach: Mind as a Web
Akram Bitar and Larry Manevitz Department of Computer Science
Presentation transcript:

1 Algorithmic Networks & Optimization Maastricht, October 2009 Ronald L. Westra, Department of Mathematics Maastricht University

2 Course Objectives This course gives an introduction to basic neural network architectures and learning rules. Emphasis is placed on the mathematical analysis of these networks, on methods of training them and on their application to practical engineering problems in such areas as pattern recognition, signal processing and control systems.

3 COURSE MATERIAL 1.Neural Networks Design (Hagan et al) 2.Network Theory (C. Gross) 3.Lecture Notes and Slides 4.Lecture - handouts 5.MATLAB Neural Networks Toolbox

4 Course Material 1. NEURAL NETWORKS Course book : Neural Network Design, Martin T. Hagan, Howard B. Demuth, Mark H. Beale, ISBN: Neural Network Design 2. NETWORK THEORY Course book : Complex and Adaptive Dynamical Systems, a Primer, Claudius Gros (Institute for Theoretical Physics, Frankfurt University), Springer (2008), ISBN-10: This document is freely available under: Complex and Adaptive Dynamical Systems, a Primer

5 GENERAL INFORMATION Course Methodology The course consists of the following components; i. a series of 10 lectures and 10 mini-exams, ii. 7 skills classes, each with one programming task, iii. one final written exam. In the lectures the main theoretical aspects will be presented. Each lecture starts with a "mini-exam" with three short questions belonging to the previous lecture. In the skills classes (SCs) several programming tasks are performed, one of which has to be submitted until next SC. Finally,the course terminates with an exam.

6 GENERAL INFORMATION mini-exams * First 10 minutes of the lecture * Closed Book * Three short questions on the previous lecture * Counts as bonus points for the final mark … * There is *NO* resit.

7 Introduction Computational dynamical networks are all around us, ranging from the metabolic and gene regulated networks in our body, via the neural networks in our brain making sense of these words, to the internet and the world wide web on which we rely so much in modern society.

8 Introduction In this course we will study networks consisting of computational components - the 'nodes', in which nodes communicate via a specific but potentially dynamic and flexible network structure and react based on all incoming information - from the other nodes and from specific 'inputs'. The subject of network is vast and extensive, more than one single course can capture.

9 Introduction In this course we will emphasize three essential aspects of dynamical networks: 1. their ability to translate incoming 'information' into a suitable 'reaction', 2. their abilities and limitations in accurately mapping a collection of specific inputs to a collection of associated outputs, and 3. the relation of network topology to their efficiency to process and store data. The first aspect relates to the 'Algorithmic' in the name of the course.

10 Introduction The second to the 'Optimization', because learning in our context is regarded as the optimization of the network parameters to map the input to the desired output. The third topic covers the subject of what nowadays is called 'Network Theory' and thus relates to the remaining word in the title of the course.

11 Introduction In this course major attention is devoted to artificial neural networks (ANNs or simply NNs). These were among the first research topics in AI, and are therefore of importance to knowledge engineering. At present, many other networks are modeled to NNs, for instance Gene Regulated networks. We will consider simple single-layer feed-forward networks like the perceptron and the ADALINE networks, multi-layered feed-forward networks with back-propagation, and competitive networks with associative learning including Kohonen networks.

12 Introduction Furthermore, we will shortly discuss the topic of gene networks as example of potentially fully connected networks. Finally, we will extensively study network structures and network topology, and their relation to the efficiency of storing and processing information.

13 1. Background Do natural networks exhibit characteristic architectural and structural properties that may act as a format for reconstruction? Some observations...

14 Mycoplasma genitalium 500 nm 580 Kbp 477 genes 74% coding DNA Obligatory parasitic endosymbiont Mycoplasma genitalium Metabolic Network Metabolic Network Nodes are genes, edges are gene co-expressions Degree distribution Horizontally log of degree (= number of connections), vertically log of number of genes with this degree

15 Protein complex network and connected complexes in yeast S. cerevisiae, Gavin et al., Nature Cumulative degree distributions of Saccharomyces cerevisiae, Jeong et al, Nature 2001 Yeast

16 Functional modules of the kinome network [Hee, Hak, 2004]

17 Degree distributions in human gene coexpression network. Coexpressed genes are linked for different values of the correlation r, King et al, Molecular Biology and Evolution, 2004

18 Statistical properties of the human gene coexpression network. (a)Node degree distribution. (b)Clustering coefficient plotted against the node degree King et al, Molecular Biology and Evolution, 2004

19 Cumulative degree distributions for six different networks. These kind of networks are all around us …

20 Special Network Architectures

21 Network Architecture and Degree Distribution

22 Neuronal Networks Metabolic Network Nodes are genes, edges are gene co-expressions Neurons and Synapses LEFT: in vivo prepared image BELOW: Fluorescent dye indicates neuron activity

23 The fundamental component of a NN

24

25

26 Active Neurons exhibiting a “Small World” Network

27 Small world networks key to memory Philip Cohen, New Scientist, 26 May 2004 If you recall this sentence a few seconds from now, you can thank a simple network of neurons for the experience. That is the conclusions of researchers who have built a computer model that can reproduce an important aspect of short-term memory. The key, they say, is that the neurons form a "small world" network. Small-world networks are surprisingly common. Human social networks, for example, famously connect any two people on Earth - or any actor to Kevin Bacon - in six steps or less.

28 Network Structures Fully Connected Networks : all nodes are connected to all other nodes Random Networks : the nodes are connected randomly to other nodes Sparse Networks : Nodes are connected to only a few other nodes Regular Networks : there is a regular pattern like in lattices Layered Networks : the nodes are grouped in layers Small world Networks: to be defined later … Scale Free Networks : to be defined later …

29 So, what is a COMPUTATIONAL NETWORK ? NODES:process information: input→output CONNECTIONS: information transport lines between nodes with specific weight/ impedance EXTERNAL INPUTS:interact with specific nodes OUTPUTS:specific nodes that are observed

30 external inputs input-coupling nodes interaction- coupling Example of an general dynamics network topology output

31 General state space dynamics The evolution of the n-dimensional state space vector x (gene expressions/neuron activity) depend on p-dim inputs u, system parameters θ and Gaussian white noise ξ.

32 Computational Networks continuously process “information” Not considered in this course: Material flow networks, Petri nets Not considered in this course: static networks, graph, relation networks, Bayesian belief networks

33 Computational Networks We will study the influence of: * network architecture * the way the node processes information * processing of information through the network * storage of information in the network … and their computational efficiency.

34 Computational Networks Learning, pattern matching, memory, and optimization = change the network parameters such that a given input pattern is mapped to a desired output pattern

35 PART I Linear Networks

36 Let us assume a computational network (with n nodes) with linear state space dynamics Suppose we want to store M patterns in the network Memory storage in linear computational networks

37 Linearized form of a subsystem First order linear approximation of system separates state vector x and inputs u. Note that matrices A and B can be directly translated to the network topology

38 Connectivity Matrix and Network Topology

39 input-output pattern: The system has ‘learned’ to react to an external input u (e.g. toxic agent, viral infection) with a specific pattern x(t). This combination (x,u) is the input-output PATTERN

40 Memory Storage = Network Reconstruction Using these definitions it is possible to map the problem of pattern storage to the * solved * problem of gene network reconstruction with sparse estimation

41 Data from experiments & observations Empirical data

42 Information Pattern Alternatively, suppose that we have M patterns we want to store in the network:

43 The relation between the desired patterns (state derivatives, states and inputs) defines constraints on the data matrices A and B, which have to be computed. Pattern Storage:

44 Row-by-row relation Set of N decoupled linear systems of size Mx(N+m) Rich data: there is enough experimental data to reconstruct the matrix Poor data: there isn’t …

45 Reformulation: A: data matrices X and U, x: rows of A and B, b: row of state derivatives

46 With this approach we can optimally find the most suitable matrices A and B for a sufficiently large set of experimental data D. However, the matrices are very large – imagine 10 5 neurons interacting: then A contains variables! Therefore, we need a large amount of experimental data!!! Pattern Storage:

47 Pattern Storage:

48 One way to avoid this dimensionality problem is to use specific characteristics of the underlying network – so of the matrices A and B. Pattern Storage:

49 Computing the optimal A and B for storing the Patterns USE Network characteristics: In many systems the matrices A and B, are sparse (most elements are zero): Using optimization techniques from robust/sparse optimization, this problem can be defined as: Pattern Storage:

50 Solution to partial sparse case Primal problem

51 Partial sparse case – dual approach Dual problem

52 Performance of dual partial approach Artificially produced data reconstructed with this approach Compare reconstructed and original data

53 The influence of increasing intrinsic noise on the identifiability.

54 a: CPU-time Tc as a function of the problem size N, b: Number of errors as a function of the number of nonzero entries k, M = 150, m = 5, N =

55 a: Number of errors versus M, b: Computation time versus M N = 50000, k = 10, m = 0.

56 a: Minimal number of measurements Mmin required to compute A free of error versus the problem size N, b: Number of errors as a function of the intrinsic noise level σ A N = 10000, k = 10, m = 5, M = 150, measuring noise B = 0.

57 For linear networks with enough empirical data we can reconstruct the network structure. Conclusions:

58 MOST networks are not linear Moreover, a linear combination cann’t handle the AND function: IS: fire = 0.5*match + 0.5*wood ? NO: fire = 1*match *wood ! Problem: AND

59 A quasi-linear network is defined as: where f (.) is a sigmoid function QUASI-LINEAR NETWORKS:

60 Notice that as f (.) is a monotonic increasing function that we can write: However, if the argument in f is outside the linear domain this will provide huge identfiability problems. QUASI-LINEAR NETWORKS:

61 PART II Neural Networks

62 Literature: Neural Network Design Martin T. Hagan, Howard B. Demuth, Mark H. Beale ISBN: See: with chapter 1 to 14