Download presentation
Presentation is loading. Please wait.
Published byBritney Ray Modified over 8 years ago
1
CHEE825 Fall 2005J. McLellan1 Nonlinear Empirical Models
2
CHEE825 Fall 2005J. McLellan2 Neural Network Models of Process Behaviour generally modeling input-output behaviour empirical models - no attempt to model physical structure estimated from plant data
3
CHEE825 Fall 2005J. McLellan3 Neural Networks... structure motivated by physiological structure of brain individual nodes or cells - “neurons” -sometimes called “perceptrons” neuron characteristics - notion of “firing” or threshold behaviour
4
CHEE825 Fall 2005J. McLellan4 Stages of Neural Network Model Development data collection - training set, validation set specification / initialization - structure of network, initial values “learning” or training - estimation of parameters validation - ability to predict new data set collected under same conditions
5
CHEE825 Fall 2005J. McLellan5 Data Collection expected range and point of operation size of input perturbation signal type of input perturbation signal -random input sequence? -number of levels (two or more?) validation data set
6
CHEE825 Fall 2005J. McLellan6 Model Structure numbers and types of nodes input, “hidden”, output depends on type of neural network -e.g., Feedforward Neural Network -e.g., Recurrent Neural Network types of neuron functions - threshold behaviour - e.g., sigmoid function, ordinary differential equation
7
CHEE825 Fall 2005J. McLellan7 “Learning” (Training) estimation of network parameters - weights, thresholds and bias terms nonlinear optimization problem objective function - typically sum of squares of output prediction error optimization algorithm - gradient-based method or variation
8
CHEE825 Fall 2005J. McLellan8 Validation use estimated NN model to predict outputs for new data set if prediction unacceptable, “re-train” NN model with modifications - e.g., number of neurons diagnostics -sum of squares of prediction error -R 2 - coefficient of determination
9
CHEE825 Fall 2005J. McLellan9 Feedforward Neural Networks signals flow forward from input through hidden nodes to output -no internal feedback input nodes - receive external inputs (e.g., controls) and scale to [0,1] range hidden nodes - collect weighted sums of inputs from other nodes and act on the sum with a nonlinear function
10
CHEE825 Fall 2005J. McLellan10 Feedforward Neural Networks (FNN) output nodes - similar to hidden nodes BUT they produce signals leaving the network (outputs) FNN has one input layer, one output layer, and can have many hidden layers
11
CHEE825 Fall 2005J. McLellan11 FNN - Neuron Model ith neuron in layer l+1 threshold value weight activation function state of neuron
12
CHEE825 Fall 2005J. McLellan12 FNN parameters weights w l+1 ij - weight on output from jth neuron in layer l entering neuron i in layer l+1 threshold - determines value of function when inputs to neuron are zero bias - provision for additional constants to be added
13
CHEE825 Fall 2005J. McLellan13 FNN Activation Function typically sigmoidal function
14
CHEE825 Fall 2005J. McLellan14 FNN Structure input layer hidden layer output layer
15
CHEE825 Fall 2005J. McLellan15 Mathematical Basis approximation of functions e.g., Cybenko, 1989 - J. of Mathematics of Control, Signals and Systems approximation to arbitrary degree given sufficiently large number of nodes - sigmoidal
16
CHEE825 Fall 2005J. McLellan16 Training FNN’s calculate sum of squares of output prediction error take current iterates of parameters, calculate forward and calculate E update estimates of weights working backwards - “backpropagation”
17
CHEE825 Fall 2005J. McLellan17 Estimation typically using a gradient-based optimization method make adjustments proportional to issues - highly over-parameterized models - potential for singularity e.g., Levenberg-Marquardt algo.
18
CHEE825 Fall 2005J. McLellan18 How to use FNN for modeling dynamic behaviour? structure of FNN suggests static model model dynamic model as nonlinear difference equation essentially a NARMAX model
19
CHEE825 Fall 2005J. McLellan19 Linear discrete time transfer function transfer function equivalent difference equation
20
CHEE825 Fall 2005J. McLellan20 FNN Structure - 1st order linear example input layer hidden layer output layer ykyk ukuk u k-1 y k+1
21
CHEE825 Fall 2005J. McLellan21 FNN model for 1st order linear example essentially modelling algebraic relationship between past and present inputs and outputs nonlinear activation function not required weights required - correspond to coefficients in discrete transfer function
22
CHEE825 Fall 2005J. McLellan22 Applications of FNN’s process modeling - bioreactors, pulp and paper, nonlinear control data reconciliation fault detection some industrial applications - many academic (simulation) studies
23
CHEE825 Fall 2005J. McLellan23 “Typical dimensions” Dayal et al., 1994 - 3-state jacketted CSTR as a basis 700 data points in training set 6 inputs, 1 hidden layer with 6 nodes, 1 output
24
CHEE825 Fall 2005J. McLellan24 Advantages of Neural Net Models limited process knowledge required - but be careful (e.g., Dayal et al. paper) flexible - can model difficult relationships directly (e.g., inverse of a nonlinear control problem)
25
CHEE825 Fall 2005J. McLellan25 Disadvantages potential for large computational requirements - implications for real-time application highly over-parameterized limited insight into process structure amount of data required limited to range of data collection
26
CHEE825 Fall 2005J. McLellan26 Recurrent Neural Networks neurons contain differential equation model - 1st order linear + nonlinearity contain feedback and feedforward components can represent continuous dynamics e.g., You and Nikolaou, 1993
27
CHEE825 Fall 2005J. McLellan27 Nonlinear Empirical Model Representations Volterra Series (continuous and discrete) Nonlinear Auto-Regressive Moving Average with Exogenous Inputs (NARMAX) Cascade Models
28
CHEE825 Fall 2005J. McLellan28 Volterra Series Models higher-order convolution models continuous
29
CHEE825 Fall 2005J. McLellan29 Volterra Series Model discrete time
30
CHEE825 Fall 2005J. McLellan30 Volterra Series models... can be estimated directly from data or derived from state space models causality - limits of sum or integration functions h i - referred to as the ith order kernel applications - typically second-order (e.g., Pearson et al., 1994 - binder)
31
CHEE825 Fall 2005J. McLellan31 NARMAX models nonlinear difference equation models typical form dependence on lagged y’s - autoregressive dependence on lagged u’s - moving average
32
CHEE825 Fall 2005J. McLellan32 NARMAX examples with products, cross-products 2nd order Volterra model –as NARMAX model in u only, with second order terms
33
CHEE825 Fall 2005J. McLellan33 Nonlinear Cascade Models made from serial and parallel arrangements of static nonlinear and linear dynamic elements e.g., 1st order linear dynamic element fed into a “squaring” element –obtain products of lagged inputs –cf. second order Volterra term
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.