Self-organizing Learning Array based Value System — SOLAR-V Yinyin Liu EE690 Ohio University Spring 2005.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Neural Networks and Kernel Methods
EE 690 Design of Embodied Intelligence
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Uncertainty in fall time surrogate Prediction variance vs. data sensitivity – Non-uniform noise – Example Uncertainty in fall time data Bootstrapping.
Pattern Recognition and Machine Learning
Support Vector Machines
Ch11 Curve Fitting Dr. Deshi Ye
Artificial Neural Networks
Machine Learning Neural Networks
Optimizing number of hidden neurons in neural networks
Evolutionary Feature Extraction for SAR Air to Ground Moving Target Recognition – a Statistical Approach Evolving Hardware Dr. Janusz Starzyk Ohio University.
Reduced Support Vector Machine
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
Three kinds of learning
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Associative Learning in Hierarchical Self Organizing Learning Arrays Janusz A. Starzyk, Zhen Zhu, and Yue Li School of Electrical Engineering and Computer.
MACHINE LEARNING 12. Multilayer Perceptrons. Neural Networks Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Fourth International Symposium on Neural Networks (ISNN) June 3-7, 2007, Nanjing, China Online Dynamic Value System for Machine Learning Haibo He, Stevens.
Information Fusion Yu Cai. Research Article “Comparative Analysis of Some Neural Network Architectures for Data Fusion”, Authors: Juan Cires, PA Romo,
Dan Simon Cleveland State University
SOMTIME: AN ARTIFICIAL NEURAL NETWORK FOR TOPOLOGICAL AND TEMPORAL CORRELATION FOR SPATIOTEMPORAL PATTERN LEARNING.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
Radial Basis Function Networks
An Introduction to Support Vector Machines Martin Law.
Radial Basis Function Networks
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
PATTERN RECOGNITION AND MACHINE LEARNING
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Particle Filtering in Network Tomography
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Classification Part 3: Artificial Neural Networks
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
Chapter 12 Multiple Linear Regression Doing it with more variables! More is better. Chapter 12A.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
WB1440 Engineering Optimization – Concepts and Applications Engineering Optimization Concepts and Applications Fred van Keulen Matthijs Langelaar CLA H21.1.
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression.
CSC321: Neural Networks Lecture 2: Learning with linear neurons Geoffrey Hinton.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Overview of Supervised Learning Overview of Supervised Learning2 Outline Linear Regression and Nearest Neighbors method Statistical Decision.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
CHECKERS: TD(Λ) LEARNING APPLIED FOR DETERMINISTIC GAME Presented By: Presented To: Amna Khan Mis Saleha Raza.
Non-Bayes classifiers. Linear discriminants, neural networks.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Mutation Operator Evolution for EA-Based Neural Networks By Ryan Meuth.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.
Neural Networks Lecture 11: Learning in recurrent networks Geoffrey Hinton.
Information Processing by Neuronal Populations Chapter 6: Single-neuron and ensemble contributions to decoding simultaneously recoded spike trains Information.
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Deep Learning and Deep Reinforcement Learning. Topics 1.Deep learning with convolutional neural networks 2.Learning to play Atari video games with Deep.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
RiskTeam/ Zürich, 6 July 1998 Andreas S. Weigend, Data Mining Group, Information Systems Department, Stern School of Business, NYU 2: 1 Nonlinear Models.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Machine Learning Supervised Learning Classification and Regression
Robert Anderson SAS JMP
Deterministic Dynamics
Dr. Unnikrishnan P.C. Professor, EEE
Neural Networks Advantages Criticism
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Biointelligence Laboratory, Seoul National University
Neural networks (3) Regularization Autoencoder
What is Artificial Intelligence?
Presentation transcript:

Self-organizing Learning Array based Value System — SOLAR-V Yinyin Liu EE690 Ohio University Spring 2005

2 Outline What is Value System What is Value System Basics of SOLAR Basics of SOLAR Least Square Method Least Square Method Value Learning in SOLAR Value Learning in SOLAR Application Application How to use in SOLAR-V Project How to use in SOLAR-V Project

3 Value System In value system, terms such as 'good', 'bad', 'better', and 'worse' are quantified. In value system, terms such as 'good', 'bad', 'better', and 'worse' are quantified. Reinforcement learning Reinforcement learning Data analysis Data analysis Model interpreting and decision making Model interpreting and decision making Value prediction Value prediction With the right value system, one will make good short term decisions and have good long term results. With the right value system, one will make good short term decisions and have good long term results.

4 Value System Reinforcement Learning (RL) Reinforcement Learning (RL) a computational approach to learning a computational approach to learning An agent tries to maximize the total reward when interacting with complex uncertain environment An agent tries to maximize the total reward when interacting with complex uncertain environment Future expected reward — value Future expected reward — value t... s t a r t +1 s a r t +2 s a r t +3 s... t +3 a

5 Value System Value functions in RL Value functions in RL Functions of state-action pair: how good it is to perform a given action in a given state Functions of state-action pair: how good it is to perform a given action in a given state Value functions can be estimated from experience Value functions can be estimated from experience θ Trial X- initial X over limit r=1 1 0 x1, x2, x3, x4, a value

6 Basics of SOLAR Basics of SOLAR Training data: Training data: SOLAR have N inputs and reads the information in parallel from M feature vectors. SOLAR have N inputs and reads the information in parallel from M feature vectors. Prewiring procedure. Prewiring procedure. SOLAR is a feed forward structure. SOLAR is a feed forward structure. Interconnections and neuron operations are dynamic based on data during the interaction with environment. Interconnections and neuron operations are dynamic based on data during the interaction with environment. M N input samples

7 Basics of SOLAR An Deviation-based Selection (DBS) — determine a proper operation and inputs for each neuron. An Deviation-based Selection (DBS) — determine a proper operation and inputs for each neuron. Each neuron is a value estimator. Final value approximation is the global voting from all the neurons Each neuron is a value estimator. Final value approximation is the global voting from all the neurons

8 Least Square Method Least Square Method: least square fit to obtain least sum of squared errors between the data and approximation Least Square Method: least square fit to obtain least sum of squared errors between the data and approximation Function — linear combination of k basis functions Function — linear combination of k basis functions W is a set of weights — needs to be found out W is a set of weights — needs to be found out Projection to the space spanned by the basis function Projection to the space spanned by the basis function Easy to implement and debug — quantify the importance of each basis feature, engineer the features for better performance. Easy to implement and debug — quantify the importance of each basis feature, engineer the features for better performance.

9 Least Square Method Signal-to-noise Ratio (SNR)-controlled LSF Signal-to-noise Ratio (SNR)-controlled LSF How many basis functions should be considered? How many basis functions should be considered? Polynomials as basis functions — up to which order? Polynomials as basis functions — up to which order? Information may be corrupted by noise — over-fitting should be avoided Information may be corrupted by noise — over-fitting should be avoided we need to determine when the difference between the approximated and the measured data has characteristics of the noise signal. we need to determine when the difference between the approximated and the measured data has characteristics of the noise signal.

10 Least Square Method SNR-controlled LSF SNR-controlled LSF approximation error signal e(x)=s(x)-a(x) approximation error signal e(x)=s(x)-a(x) Determine the signal to noise ratio of e(x) by using signal correlation Determine the signal to noise ratio of e(x) by using signal correlation This S/N of the Gaussian noise is a random variable and its statistics can be directly estimated This S/N of the Gaussian noise is a random variable and its statistics can be directly estimated If this means that, most likely, not all the information was extracted from sampled data. In such case, increase basis function order by 1 If this means that, most likely, not all the information was extracted from sampled data. In such case, increase basis function order by 1

11 Least Square Method SNR-controlled LSF SNR-controlled LSF

12 Least Square Method Weighted LSF Weighted LSF knowledge is accumulated through the learning process, should relate more to recent information knowledge is accumulated through the learning process, should relate more to recent information Recent data weights more in the learning Recent data weights more in the learning Apply weights to data: exponentially declining going back from the most recent data Apply weights to data: exponentially declining going back from the most recent data

13 Least Square Method Weighted LSF Weighted LSF

14 Least Square Method Weighted LSF Weighted LSF error signal has to be weighted as well. error signal has to be weighted as well. The SNR of Gaussian noise The SNR of Gaussian noise Comparison Comparison

15 Value Learning Training data: Training data: Neuron Neuron initial wiring Trial X- initial X over limit r=1 1 0 x1, x2, x3, x4, a value

16 Value Learning Non-linear Scaling: Non-linear Scaling:

17 Value Learning Deviation-based Selection in neurons Deviation-based Selection in neurons Inputs: 1 or 2 inputs Inputs: 1 or 2 inputs Operation: Operation: ident, half, exp, log, add, sub Self-organized Self-organizedstructure

18 Value Learning Value information is stored distributed in neurons. Value information is stored distributed in neurons. Value approximation is collecting the approximation result from all the neurons. Value approximation is collecting the approximation result from all the neurons.

19 Value Learning Information fusion: use information from many sources. Information fusion: use information from many sources. Value approximation: global voting Value approximation: global voting

20 Value Learning Learning performance Learning performance

21 Application On-line Learning control by Reinforcement and Association On-line Learning control by Reinforcement and Association Action Network Value Network System X(t) X(t+1) X(t) u(t) J(t)

22 Application Financial data analysis Financial data analysis Data:103 features from 52 companies. Value: 52 gain values given by the one-year gain on investment. Prediction Prediction Prediction future gain based on current features.

23 SOLAR-V Project Prepare the data Prepare the data Data sample along the row: N samples Data sample along the row: N samples Features along the column: M features Features along the column: M features Given values in a row vector: N values Given values in a row vector: N values Save “ features ” and “ values ” in a training MAT file Save “ features ” and “ values ” in a training MAT file Save “ features_test ” and “ values_test ” in testing MAT file Save “ features_test ” and “ values_test ” in testing MAT file How to recall the function How to recall the function Run “ solar_v_main.m ” Run “ solar_v_main.m ” Input MAT file name and number of layers in command window. Input MAT file name and number of layers in command window. Input if you think any feature is more significant and like to repeat how many times. Input if you think any feature is more significant and like to repeat how many times. In the function, data will be scaled to 0~255, values are kept unchanged In the function, data will be scaled to 0~255, values are kept unchanged Function will determine how many neurons per layer, you can decide how many layers Function will determine how many neurons per layer, you can decide how many layers Several figures will be generated Several figures will be generated Prewired network structure Prewired network structure Self-organized network structure Self-organized network structure Learning performance compared with Neural Network Learning performance compared with Neural Network Testing results from SOLAR-V compared with Neural Network Testing results from SOLAR-V compared with Neural Network M x N matrix: features 1 x N vector: values

24 SOLAR-V Project Example: data from on-line control model Example: data from on-line control model … … … … … …. 5 neurons per layer, 3 layers 5 neurons per layer, 3 layers Data 4 states 1 action Value