A Modified Meta-controlled Boltzmann Machine Tran Duc Minh, Le Hai Khoi (*), Junzo Watada (**), Teruyuki Watanabe (***) (*) Institute Of Information Technology-Viet.

Slides:



Advertisements
Similar presentations
The simplex algorithm The simplex algorithm is the classical method for solving linear programs. Its running time is not polynomial in the worst case.
Advertisements

Problems and Their Classes
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
Perceptron Learning Rule
H EURISTIC S OLVER  Builds and tests alternative fuel treatment schedules (solutions) at each iteration  In each iteration:  Evaluates the effects of.
O(N 1.5 ) divide-and-conquer technique for Minimum Spanning Tree problem Step 1: Divide the graph into  N sub-graph by clustering. Step 2: Solve each.
Simulated Annealing Premchand Akella. Agenda Motivation The algorithm Its applications Examples Conclusion.
1 Using ANN to solve the Navier Stoke Equations Motivation: Solving the complete Navier Stokes equations using direct numerical simulation is computationally.
CS 678 –Boltzmann Machines1 Boltzmann Machine Relaxation net with visible and hidden units Learning algorithm Avoids local minima (and speeds up learning)
CHAPTER 8 A NNEALING- T YPE A LGORITHMS Organization of chapter in ISSO –Introduction to simulated annealing –Simulated annealing algorithm Basic algorithm.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
1 Simulation Modeling and Analysis Session 13 Simulation Optimization.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Prof. Carla P. Gomes Module: Neural Networks: Concepts (Reading:
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
Network Optimization Models: Maximum Flow Problems In this handout: The problem statement Solving by linear programming Augmenting path algorithm.
A Comparative Study Of Deterministic And Stochastic Optimization Methods For Integrated Design Of Processes Mario Francisco a, Silvana Revollar b, Pastora.
Multicast Routing in ATM Networks with Multiple Classes of QoS Ren-Hung Hwang, Min-Xiou Chen, and Youn-Chen Sun Department of Computer Science & Information.
Simulated Annealing Van Laarhoven, Aarts Version 1, October 2000.
D Nagesh Kumar, IIScOptimization Methods: M1L4 1 Introduction and Basic Concepts Classical and Advanced Techniques for Optimization.
Linear Programming Applications
1 IOE/MFG 543 Chapter 7: Job shops Sections 7.1 and 7.2 (skip section 7.3)
October 28, 2010Neural Networks Lecture 13: Adaptive Networks 1 Adaptive Networks As you know, there is no equation that would tell you the ideal number.
Application of Back-Propagation neural network in data forecasting Le Hai Khoi, Tran Duc Minh Institute Of Information Technology – VAST Ha Noi – Viet.
Simulated Annealing G.Anuradha. What is it? Simulated Annealing is a stochastic optimization method that derives its name from the annealing process used.
Elements of the Heuristic Approach
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
JingTao Yao Growing Hierarchical Self-Organizing Maps for Web Mining Joseph P. Herbert and JingTao Yao Department of Computer Science, University or Regina.
Topology Design for Service Overlay Networks with Bandwidth Guarantees Sibelius Vieira* Jorg Liebeherr** *Department of Computer Science Catholic University.
Biointelligence Laboratory, Seoul National University
Hybrid AI & Machine Learning Systems Using Ne ural Network and Subsumption Architecture Libraries By Logan Kearsley.
0 Portfolio Managment Albert Lee Chun Construction of Portfolios: Introduction to Modern Portfolio Theory Lecture 3 16 Sept 2008.
EXPLORATION AND PRODUCTION (E&P) How to Choose and Manage Exploration and Production Projects Supat Kietnithiamorn Kumpol Trivisvavet May 9, 2001 Term.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
1 Nasser Alsaedi. The ultimate goal for any computer system design are reliable execution of task and on time delivery of service. To increase system.
Simulated Annealing.
CSC321: Introduction to Neural Networks and machine Learning Lecture 16: Hopfield nets and simulated annealing Geoffrey Hinton.
Method of Hooke and Jeeves
Exact and heuristics algorithms
Soft Computing Lecture 8 Using of perceptron for image recognition and forecasting.
The Greedy Method. The Greedy Method Technique The greedy method is a general algorithm design paradigm, built on the following elements: configurations:
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
Simulated Annealing G.Anuradha.
1 Network Models Transportation Problem (TP) Distributing any commodity from any group of supply centers, called sources, to any group of receiving.
Optimizing Pheromone Modification for Dynamic Ant Algorithms Ryan Ward TJHSST Computer Systems Lab 2006/2007 Testing To test the relative effectiveness.
Tetris Agent Optimization Using Harmony Search Algorithm
Introduction to Neural Networks. Biological neural activity –Each neuron has a body, an axon, and many dendrites Can be in one of the two states: firing.
An Introduction to Simulated Annealing Kevin Cannons November 24, 2005.
1 Ch. 2: Getting Started. 2 About this lecture Study a few simple algorithms for sorting – Insertion Sort – Selection Sort (Exercise) – Merge Sort Show.
A study of simulated annealing variants Ana Pereira Polytechnic Institute of Braganca, Portugal Edite Fernandes University of Minho,
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
CS623: Introduction to Computing with Neural Nets (lecture-12) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Lecture 9 Model of Hopfield
Financial Data mining and Tools CSCI 4333 Presentation Group 6 Date10th November 2003.
IJCNN, July 27, 2004 Extending SpikeProp Benjamin Schrauwen Jan Van Campenhout Ghent University Belgium.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Balanced Billing Cycles and Vehicle Routing of Meter Readers by Chris Groër, Bruce Golden, Edward Wasil University of Maryland, College Park American University,
Nonlinear balanced model residualization via neural networks Juergen Hahn.
Meta-controlled Boltzmann Machine toward Accelerating the Computation Tran Duc Minh (*), Junzo Watada (**) (*) Institute Of Information Technology-Viet.
SemiBoost : Boosting for Semi-supervised Learning Pavan Kumar Mallapragada, Student Member, IEEE, Rong Jin, Member, IEEE, Anil K. Jain, Fellow, IEEE, and.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Stat 261 Two phase method.
The minimum cost flow problem
NPDAs Accept Context-Free Languages
Subject Name: Operation Research Subject Code: 10CS661 Prepared By:Mrs
Introduction to Simulated Annealing
Boltzmann Machine (BM) (§6.4)
Ch. 2: Getting Started.
… NPDAs continued.
More on HW 2 (due Jan 26) Again, it must be in Python 2.7.
Simulated Annealing & Boltzmann Machines
Presentation transcript:

A Modified Meta-controlled Boltzmann Machine Tran Duc Minh, Le Hai Khoi (*), Junzo Watada (**), Teruyuki Watanabe (***) (*) Institute Of Information Technology-Viet Nam Academy of Science & Technology (**) Graduate School of Information, Production and System, Waseda University, Japan (***) Osaka Institute of Technology 03/2004

2 CONTENT Introduction The portfolio selection problem Inner behaviors of the Meta-controlled Boltzmann machine A Modified Meta-controlled Boltzmann machine Conclusion

3 Introduction H. Markowitz proposed a method to allocate an amount of funds to plural stocks for investment Model of Meta-controlled Boltzmann Machine The ability of Meta-controlled Boltzmann Machine in solving the quadratic programming problem

4 Maximize Minimize Subject to and with m i  {0, 1}, i = 1,.., n where  ij denotes a covariance between stocks i and j,  i is an expected return rate of stock i, x i is investing rate to stock i, n denotes the total number of stocks and S denotes the number of stocks selected, and finally, m i denotes a selection variable of investing stocks. The portfolio selection problem

5 Convert the objective function into the energy functions of the two components that are Meta- controlling layer (Hopfield Network) and the Lower-layer (Boltzmann Machine) as described below: Meta-Controlling layer Lower Layer where K u, K l are weights of the expected return rate for each layer and s i is the output value of the ith unit of the Meta-Controlling layer.

6 Algorithm of the Meta-controlled Boltzmann machine Step 1. Set each parameter to its initial value. Step 2. Input the values of Ku and Kl. Step 3. Execute the Meta-controlling layer. Step 4. If the output value of a unit in the Meta-controlling layer is 1, add some amount of value to the corresponding unit in the lower layer. Execute the lower layer. Step 5. After executing the lower layer the constant number of times, decreases the temperature. Step 6. If the output value is sufficiently large, add a certain amount of value to the corresponding unit in the Meta-controlling layer. Step 7. Iterate from Step 3 to Step 6 until the temperature reaches the restructuring temperature. Step 8. Restructure the lower layer using the selected units of the Meta-controlling layer. Step 9. Execute the lower layer until reaching at the termination.

7 Inner behaviors of the Meta- controlled Boltzmann machine Some times, the Hopfield layer may converge to a local minimum but the disturb values make it to get over The changes of Meta layer’s energy function are very small, while the lower layer’s energy function’s is quite large The number of cycles to execute the Meta layer is much smaller than the cycles for the lower layer Similar to the simulated annealing that we will “try to go downhill most of the time instead of always going downhill” The time to converge is much shorter than a conventional Boltzmann machine All the neurons that are “encouraged” will be selected before the system goes to the final computation.

8 Chart of behaviors of Meta-controlled Boltzmann Machine Disturb back value = 80%

9 Chart of behaviors of Meta-controlled Boltzmann Machine Disturb back value = 1 %

10 Comparison of computing time between a Conventional Boltzmann machine and a Meta-controlled Boltzmann Machine (1286 units)

11 Some hints on accelerating the Meta- controlled Boltzmann machine Trying to use only a layer of Boltzmann Machine, modify the algorithm of original Boltzmann Machine by removing the discouraged units before goes into final computation. Trying to use only a layer of Boltzmann Machine, modify the algorithm of original Boltzmann Machine by removing the discouraged units before goes into final computation. Modify the original Boltzmann Machine by replacing deterministic neurons by stochastic neurons since the disturb from the lower layer to the upper layer may not be worth. Modify the original Boltzmann Machine by replacing deterministic neurons by stochastic neurons since the disturb from the lower layer to the upper layer may not be worth.

12 A Modified Meta-controlled Boltzmann machine Step 1. Set each parameter to its initial value. Step 2. Input the values of Ku, Kl. Step 3. Execute the Meta-controlling layer. Step 4. If the output value of a unit in the Meta-controlling layer is 1, add some amount of value to the corresponding unit in the lower layer. Execute the lower layer. Step 5. After executing the lower layer the constant number of times, decreases the temperature. Step 6. Iterate Step 4, 5 until the temperature reaches the restructuring temperature. Step 7. Restructure the lower layer using the selected units of the Meta- controlling layer. Step 8. Execute the lower layer until reaching at the termination. Algorithm of the Modified Meta-controlled Boltzmann machine

13

14

15 Comparing performance

16 CONCLUSION The trend of accelerating algorithms is focused mainly on heuristic modification and numeric optimization technique, i.e. toward the faster convergence of algorithms whereas keeping the correctness for them. The Meta-controlled Boltzmann Machine can be used to solve quadratic programming problems. Future works: Try the model with other quadratic programming problem. Evaluate the modified Meta-controlled Boltzmann Machine.

17 THANK YOU!