Vaida Bartkutė, Leonidas Sakalauskas

Slides:



Advertisements
Similar presentations
Probabilistic models Haixu Tang School of Informatics.
Advertisements

Neural and Evolutionary Computing - Lecture 4 1 Random Search Algorithms. Simulated Annealing Motivation Simple Random Search Algorithms Simulated Annealing.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
CHAPTER 8 A NNEALING- T YPE A LGORITHMS Organization of chapter in ISSO –Introduction to simulated annealing –Simulated annealing algorithm Basic algorithm.
CHAPTER 2 D IRECT M ETHODS FOR S TOCHASTIC S EARCH Organization of chapter in ISSO –Introductory material –Random search methods Attributes of random search.
All Hands Meeting, 2006 Title: Grid Workflow Scheduling in WOSE (Workflow Optimisation Services for e- Science Applications) Authors: Yash Patel, Andrew.
Gizem ALAGÖZ. Simulation optimization has received considerable attention from both simulation researchers and practitioners. Both continuous and discrete.
11 - Markov Chains Jim Vallandingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Suggested readings Historical notes Markov chains MCMC details
Optimization methods Morten Nielsen Department of Systems biology, DTU.
By : L. Pour Mohammad Bagher Author : Vladimir N. Vapnik
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
The Nature of Statistical Learning Theory by V. Vapnik
Stochastic Differentiation Lecture 3 Leonidas Sakalauskas Institute of Mathematics and Informatics Vilnius, Lithuania EURO Working Group on Continuous.
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
Job Release-Time Design in Stochastic Manufacturing Systems Using Perturbation Analysis By: Dongping Song Supervisors: Dr. C.Hicks & Dr. C.F.Earl Department.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Evaluating Hypotheses
Back-Propagation Algorithm
1 IOE/MFG 543 Chapter 14: General purpose procedures for scheduling in practice Section 14.4: Local search (Simulated annealing and tabu search)
Planning operation start times for the manufacture of capital products with uncertain processing times and resource constraints D.P. Song, Dr. C.Hicks.
Topologically Adaptive Stochastic Search I.E. Lagaris & C. Voglis Department of Computer Science University of Ioannina - GREECE IOANNINA ATHENS THESSALONIKI.
Nonlinear Stochastic Programming by the Monte-Carlo method Lecture 4 Leonidas Sakalauskas Institute of Mathematics and Informatics Vilnius, Lithuania EURO.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
CHAPTER 15 S IMULATION - B ASED O PTIMIZATION II : S TOCHASTIC G RADIENT AND S AMPLE P ATH M ETHODS Organization of chapter in ISSO –Introduction to gradient.
Photo-realistic Rendering and Global Illumination in Computer Graphics Spring 2012 Stochastic Radiosity K. H. Ko School of Mechatronics Gwangju Institute.
Stochastic Approximation and Simulated Annealing Lecture 8 Leonidas Sakalauskas Institute of Mathematics and Informatics Vilnius, Lithuania EURO Working.
Introduction to Adaptive Digital Filters Algorithms
Efficient Model Selection for Support Vector Machines
Based on: The Nature of Statistical Learning Theory by V. Vapnick 2009 Presentation by John DiMona and some slides based on lectures given by Professor.
Instructor: Prof.Dr.Sahand Daneshvar Presented by: Seyed Iman Taheri Student number: Non linear Optimization Spring EASTERN MEDITERRANEAN.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
Target Tracking with Binary Proximity Sensors: Fundamental Limits, Minimal Descriptions, and Algorithms N. Shrivastava, R. Mudumbai, U. Madhow, and S.
CHAPTER 4 S TOCHASTIC A PPROXIMATION FOR R OOT F INDING IN N ONLINEAR M ODELS Organization of chapter in ISSO –Introduction and potpourri of examples Sample.
Chapter 7 Optimization. Content Introduction One dimensional unconstrained Multidimensional unconstrained Example.
An Empirical Likelihood Ratio Based Goodness-of-Fit Test for Two-parameter Weibull Distributions Presented by: Ms. Ratchadaporn Meksena Student ID:
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
Markov Decision Processes1 Definitions; Stationary policies; Value improvement algorithm, Policy improvement algorithm, and linear programming for discounted.
Module 1: Statistical Issues in Micro simulation Paul Sousa.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Simulated Annealing.
Lecture 2 Basics of probability in statistical simulation and stochastic programming Leonidas Sakalauskas Institute of Mathematics and Informatics Vilnius,
Method of Hooke and Jeeves
Clustering and Testing in High- Dimensional Data M. Radavičius, G. Jakimauskas, J. Sušinskas (Institute of Mathematics and Informatics, Vilnius, Lithuania)
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Non-Bayes classifiers. Linear discriminants, neural networks.
Monte-Carlo method for Two-Stage SLP Lecture 5 Leonidas Sakalauskas Institute of Mathematics and Informatics Vilnius, Lithuania EURO Working Group on Continuous.
CHAPTER 6 STOCHASTIC APPROXIMATION AND THE FINITE-DIFFERENCE METHOD
Introduction to Simulated Annealing Study Guide for ES205 Xiaocang Lin & Yu-Chi Ho August 22, 2000.
Heuristic Methods for the Single- Machine Problem Chapter 4 Elements of Sequencing and Scheduling by Kenneth R. Baker Byung-Hyun Ha R2.
Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.
A study of simulated annealing variants Ana Pereira Polytechnic Institute of Braganca, Portugal Edite Fernandes University of Minho,
Optimization of Nonlinear Singularly Perturbed Systems with Hypersphere Control Restriction A.I. Kalinin and J.O. Grudo Belarusian State University, Minsk,
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
1 Design of experiment for computer simulations Let X = (X 1,…,X p )  R p denote the vector of input values chosen for the computer program Each X j is.
Intro. ANN & Fuzzy Systems Lecture 37 Genetic and Random Search Algorithms (2)
Metaheuristics for the New Millennium Bruce L. Golden RH Smith School of Business University of Maryland by Presented at the University of Iowa, March.
1 Contents 1. Basic Concepts 2. Algorithm 3. Practical considerations Simulated Annealing (SA)
Amir Yavariabdi Introduction to the Calculus of Variations and Optical Flow.
Parallel Simulated Annealing using Genetic Crossover Tomoyuki Hiroyasu Mitsunori Miki Maki Ogura November 09, 2000 Doshisha University, Kyoto, Japan.
Scientific Research Group in Egypt (SRGE)
Heuristic Optimization Methods
Slides for Introduction to Stochastic Search and Optimization (ISSO) by J. C. Spall CHAPTER 15 SIMULATION-BASED OPTIMIZATION II: STOCHASTIC GRADIENT AND.
Introduction to Simulated Annealing
Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014
Bioinformatics, Vol.17 Suppl.1 (ISMB 2001)
Nonlinear Conjugate Gradient Method for Supervised Training of MLP
Stochastic Methods.
Presentation transcript:

Vaida Bartkutė, Leonidas Sakalauskas APPLICATION OF ORDER STATISTICS TO TERMINATION OF STOCHASTIC ALGORITHMS Vaida Bartkutė, Leonidas Sakalauskas

Outline Introduction; Application of order statistics to optimality testing and termination of the algorithm: Stochastic Approximation algorithms; Simulated Annealing algorithm; Experimental results; Conclusions.

Outline Introduction; Application of order statistics to optimality testing and termination of the algorithm: Stochastic Approximation algorithms; Simulated Annealing algorithm; Experimental results; Conclusions.

Introduction Termination of the algorithm is a topical problem in stochastic and heuristic optimization. We consider the application of order statistics to establish the optimality in Markov type optimization algorithms. We build a method for the estimation of minima confidence intervals using order statistics, which is implemented for optimality testing and termination.

Statement of the problem The optimization problem is (minimization) as follows: where is a bounded from below locally Lipshitz function. Denote the generalized gradient of this function by Let be the sequence constructed by stochastic search algorithm, where ηt=f(xt), t = 0, 1, …. .

The Markovian algorithms for optimization The Markovian algorithm of random searching represents a Markov chain in which the distribution of probabilities of a point xt+1 depends on a location of the previous point xt and value of function ηt=f(xt) in it, that Examples: Stochastic Approximation; Simulated Annealing; Random Search (Rastrigin method) and etc.

Order statistics and target values for optimality testing and termination Beginning of the problem: Mockus (1968) Theoretical background: Zilinskas, Zhigljavsky (1991) Application to maximum location: Chen (1996) Time-to-target-solution value: Aiex, Resende, & Ribeiro, (2002), Pardalos (2005).

Method for optimality testing by order statistics We build a method for estimation of minimum M of the objective function using values of the function provided in optimization: Let only k+1 order statistics from the sample H to be chosen: , where .

Let apply linear estimators for estimation of the minimum: where . We examine a simple set (Hall (1982)): Let apply linear estimators for estimation of the minimum: where . We examine a simple set (Hall (1982)): Let apply linear estimators for estimation of the minimum: where . We examine a simple set (Hall (1982)):

the objective function is: The one side confidence interval for the minimum value of the objective function is: [ ] where , where  is a confidence level. - the parameter of extreme values distribution; n – dimension; - the parameter of homogeneity of the function f(x) (Zilinskas & Zhigliavsky (1991)).

Stochastic Approximation The smoothing is the standard way for the nondifferentiable optimization. We consider a function smoothed by Lipshitz perturbation operator: where is the value of the perturbation parameter, is a random vector distributed with density p(.). If density p(.) is locally Lipshitz then functions smoothed by this operator are twice continuously differentiable (Rubinstein & Shapiro (1993), Bartkute & Sakalauskas (2004)).

where stochastic gradient, is a scalar multiplier. This scheme is the same for different Stochastic Approximation algorithms whose distinguish only by approach to stochastic gradient estimation. The minimizing sequence converges a.s. to solution of the optimization problem under conditions typical for SA algorithms (Ermolyev (1976), Mikhalevitch et at (1987), Spall (1992), Bartkute & Sakalauskas (2004)).

ESTIMATE OF STOCHASTIC GRADIENT ALGORITHM ESTIMATE OF STOCHASTIC GRADIENT SPSAL, Lipshitz smoothing density (Bartkute & Sakalauskas (2007)) - uniformly distributed in the unit ball. SPSAU uniformly distributed density in the hypercube (Michalevitch et al (1976), (1987)) - uniformly distributed in the hypercube [-1;1]n. FDSA standard finite differences (Ermoliev (1988), Mikhalevitch et al (1987)) - uniformly distributed in the unit ball, - with zero components except ith , equal to 1. - smoothing parameter.

Rate of Convergence Let consider that the function f(x) has a sharp minimum in the point , in which the algorithm converges when Then where A>0, H>0, K>0 are certain constants, is minimum point of the smoothed function (Sakalauskas, Bartkute (2007)).

Experimental results Unimodal testing functions (SPSAL, SPSAU, FDSA) Generated funkcions with sharp minimum- CB3- Rozen Suzuki- Multiextremal testing functions (Simulated Annealing (SA)) Branin- Beale- Rastrigin-

The samples of T=500 test functions were generated, when The samples of T=500 test functions were generated, when and minimized by SPSA with Lipshitz perturbation. The coefficients of the optimizing sequence were chosen according to convergence conditions (Bartkute & Sakalauskas (2006)):

Testing hypothesis about Pareto distribution If order statistics follows from Weibull distribution, then distributed with respect to Pareto distribution (Žilinskas, Zhigljavsky (1991)): Thus, statistical hypothesis tested: . H0:

Testing hypothesis about Pareto distribution The hypothesis tested by criteria 2 ( ) for various stochastic algorithms (critical value 0,46)

One side confidence interval [ ], =0.95 V. Bartkute, L. Sakalauskas. Application of Order Statistics to termination of Stochastic lgorithms. CWU Workshop, December 10-12, 2007

Confidence bounds of the minimum

Confidence bounds of the hitting probability

Termination criterion of the algorithms To stop the algorithm when minima confidence interval becomes less admissible value :

Number of iterations after the termination of the algorithm

Simulated Annealing Algorithm I. Choose temperature updating function neighborhood size function solution generation density function and initial solution x0 (Yang (2000)). II. Construct the optimizing sequence:

F(x,y) = (1.5-x+xy)2 + (2.25-x+xy2)2 + (2.625-x+xy3)2, Experimental results Let consider results of optimality testing with Beale testing function: F(x,y) = (1.5-x+xy)2 + (2.25-x+xy2)2 + (2.625-x+xy3)2, where search domain: -4.5 ≤ x,y ≤ 4.5. It is known that this function has few local minima and global minimum is 0 at the point (3; 0.5).

Confidence bounds of the minimum

Confidence bounds of the hitting probability

Number of iterations after the termination of the algorithm

Conclusions Linear estimator for minimum has been proposed using theory of order statistics, which was studied by experimental way; Developed procedures are simple and depend only on the parameter of extreme values distribution ; Parameter  is easily estimated using a homogeneity of the objective function or by statistical way; Theoretical considerations and computer examples have shown that we can estimate the confidence interval of a function extremum with an admissible accuracy, when the number of iterations increased; Termination rule using the minimum confidence interval was proposed and implemented to Stochastic Approximation and Simulated Annealing.