Methods of Solving the Problem of Construction the Optimum Regression Model as a Discrete Optimization Task Ivan Melnyk The International Research and.

Slides:



Advertisements
Similar presentations
How to Schedule a Cascade in an Arbitrary Graph F. Chierchetti, J. Kleinberg, A. Panconesi February 2012 Presented by Emrah Cem 7301 – Advances in Social.
Advertisements

Using Parallel Genetic Algorithm in a Predictive Job Scheduling
Crew Scheduling Housos Efthymios, Professor Computer Systems Laboratory (CSL) Electrical & Computer Engineering University of Patras.
Ch11 Curve Fitting Dr. Deshi Ye
The General Linear Model. The Simple Linear Model Linear Regression.
Gizem ALAGÖZ. Simulation optimization has received considerable attention from both simulation researchers and practitioners. Both continuous and discrete.
Institute of Intelligent Power Electronics – IPE Page1 Introduction to Basics of Genetic Algorithms Docent Xiao-Zhi Gao Department of Electrical Engineering.
1 Minimum Ratio Contours For Meshes Andrew Clements Hao Zhang gruvi graphics + usability + visualization.
Visual Recognition Tutorial
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Data Mining Techniques Outline
Evolutionary Computational Intelligence Lecture 8: Memetic Algorithms Ferrante Neri University of Jyväskylä.
TECH Computer Science Graph Optimization Problems and Greedy Algorithms Greedy Algorithms  // Make the best choice now! Optimization Problems  Minimizing.
Ant Colony Optimization: an introduction
Learning Objectives for Section 2.1 Functions
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Evolutionary algorithms
Genetic Algorithm.
A Genetic Algorithms Approach to Feature Subset Selection Problem by Hasan Doğu TAŞKIRAN CS 550 – Machine Learning Workshop Department of Computer Engineering.
Evolutionary Intelligence
Efficient Model Selection for Support Vector Machines
by B. Zadrozny and C. Elkan
FDA- A scalable evolutionary algorithm for the optimization of ADFs By Hossein Momeni.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
Chapter 12 Multiple Linear Regression Doing it with more variables! More is better. Chapter 12A.
1 The Euclidean Non-uniform Steiner Tree Problem by Ian Frommer Bruce Golden Guruprasad Pundoor INFORMS Annual Meeting Denver, Colorado October 2004.
Lecture 8: 24/5/1435 Genetic Algorithms Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Investigation of the Effect of Neutrality on the Evolution of Digital Circuits. Eoin O’Grady Final year Electronic and Computer Engineering Project.
Zorica Stanimirović Faculty of Mathematics, University of Belgrade
Department of Electrical Engineering, Southern Taiwan University Robotic Interaction Learning Lab 1 The optimization of the application of fuzzy ant colony.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
ASC2003 (July 15,2003)1 Uniformly Distributed Sampling: An Exact Algorithm for GA’s Initial Population in A Tree Graph H. S.
Neural and Evolutionary Computing - Lecture 6
1 A method of successive elimination of spurious arguments for effective solution the search- based modelling tasks Oleksandr Samoilenko, Volodymyr Stepashko.
Genetic Algorithms Introduction Advanced. Simple Genetic Algorithms: Introduction What is it? In a Nutshell References The Pseudo Code Illustrations Applications.
Genetic Algorithms Siddhartha K. Shakya School of Computing. The Robert Gordon University Aberdeen, UK
© Negnevitsky, Pearson Education, Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Introduction,
Exact and heuristics algorithms
1 1 © 2003 Thomson  /South-Western Slide Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Introduction to Genetic Algorithms. Genetic Algorithms We’ve covered enough material that we can write programs that use genetic algorithms! –More advanced.
Genetic Algorithms CSCI-2300 Introduction to Algorithms
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Introduction Metaheuristics: increasingly popular in research and industry mimic natural metaphors to solve complex optimization problems efficient and.
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
A Brief Maximum Entropy Tutorial Presenter: Davidson Date: 2009/02/04 Original Author: Adam Berger, 1996/07/05
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
Application of the GA-PSO with the Fuzzy controller to the robot soccer Department of Electrical Engineering, Southern Taiwan University, Tainan, R.O.C.
Onlinedeeneislam.blogspot.com1 Design and Analysis of Algorithms Slide # 1 Download From
Machine Learning: A Brief Introduction Fu Chang Institute of Information Science Academia Sinica ext. 1819
Genetic Algorithm Dr. Md. Al-amin Bhuiyan Professor, Dept. of CSE Jahangirnagar University.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
Computacion Inteligente Least-Square Methods for System Identification.
1 Comparative Study of two Genetic Algorithms Based Task Allocation Models in Distributed Computing System Oğuzhan TAŞ 2005.
CORRELATION-REGULATION ANALYSIS Томский политехнический университет.
BDD-based Synthesis of Reversible Logic for Large Functions Robert Wille Rolf Drechsler DAC’09 Presenter: Meng-yen Li.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Chapter 7. Classification and Prediction
R. E. Wyllys Copyright 2003 by R. E. Wyllys Last revised 2003 Jan 15
Task: It is necessary to choose the most suitable variant from some set of objects by those or other criteria.
Structural testing, Path Testing
ICS 353: Design and Analysis of Algorithms
Hidden Markov Models Part 2: Algorithms
MATS Quantitative Methods Dr Huw Owens
Genetic Algorithms CSCI-2300 Introduction to Algorithms
Boltzmann Machine (BM) (§6.4)
Artificial Intelligence CIS 342
Presentation transcript:

Methods of Solving the Problem of Construction the Optimum Regression Model as a Discrete Optimization Task Ivan Melnyk The International Research and Training Centre of Information Technologies and Systems of the National Academy of Sciences of Ukraine and Ministry of Science and Education of Ukraine Akademik Glushkov prospect, 40, Kyiv 03680, Ukraine

1. Introduction Construction the optimum regression model is a special task to select from the set of independent variables (regressors) a subset for constructing the regression model using the proper function. This is a multiextremal task of the discrete optimization. There are no effective algorithms for decision such a task. Construction the optimum regression model is a special task to select from the set of independent variables (regressors) a subset for constructing the regression model using the proper function. This is a multiextremal task of the discrete optimization. There are no effective algorithms for decision such a task. In this article the problem is formulated as a task of discrete optimization on a special graph. The exact method of solving this task is proposed as the special task for searching the shortest path on this graph, and also as a heuristic method based on the ideas of classical genetic algorithm proposed in the Michigan University by J.H. Holland. In this article the problem is formulated as a task of discrete optimization on a special graph. The exact method of solving this task is proposed as the special task for searching the shortest path on this graph, and also as a heuristic method based on the ideas of classical genetic algorithm proposed in the Michigan University by J.H. Holland.

2. The Tasks of Research A complex system having input (independent) variables and one output (dependent) variable is considered to have a stochastic character and given by a sample of n observations. The structures of linear regression models with parameters estimated by the least-squares method are generated in the process of identication. Some functional of evaluation the model F (. ) is set as the criterion for selection the optimum model. This functional can depend from the model complexity (numbers of the estimated parameters) and/or the disparity of a regression equation on some subset of the statistical sample and other parameters of model. A complex system having input (independent) variables and one output (dependent) variable is considered to have a stochastic character and given by a sample of n observations. The structures of linear regression models with parameters estimated by the least-squares method are generated in the process of identication. Some functional of evaluation the model F (. ) is set as the criterion for selection the optimum model. This functional can depend from the model complexity (numbers of the estimated parameters) and/or the disparity of a regression equation on some subset of the statistical sample and other parameters of model.

For example, let us designate the coefficients of a regression model found by least-squares method with subset of numbers of input variables and the “charges” for receiving the statistical information about the independent variable value. The following functional of model estimation is considered (1): For example, let us designate the coefficients of a regression model found by least-squares method with subset of numbers of input variables and the “charges” for receiving the statistical information about the independent variable value. The following functional of model estimation is considered (1):

Second option, when the functional is defined as a maximal value of the residual sum of squares of the regression model on a part of the set. A next criterion to choose regressive model is considered further. Great number of points of statistical selection is broken-down into two parts: and, that is typical for GMDH. On the first part of selection a regressive model is built, and the value of functional of estimation of the built model is calculated on the second part (2): Second option, when the functional is defined as a maximal value of the residual sum of squares of the regression model on a part of the set. A next criterion to choose regressive model is considered further. Great number of points of statistical selection is broken-down into two parts: and, that is typical for GMDH. On the first part of selection a regressive model is built, and the value of functional of estimation of the built model is calculated on the second part (2): where – part from the great number of numbers of independent variables, and is - part of input variables which a regressive model is built on. where – part from the great number of numbers of independent variables, and is - part of input variables which a regressive model is built on.

The value to functional (2) is calculated as a maximal value to the square of differences between the statistical values of input variable and value of regression model on all observations of values of input variables of. The value to functional (2) is calculated as a maximal value to the square of differences between the statistical values of input variable and value of regression model on all observations of values of input variables of. It is necessary to select such a subset of variables from the whole set of input variables when the regression model built on the basis of this subset produce the optimum (minimum or maximum) of a given functional of model estimation F(. ). We determine the optimum model as the proper regression model which contains only the variables (regressors) from the chosen subset. It is necessary to select such a subset of variables from the whole set of input variables when the regression model built on the basis of this subset produce the optimum (minimum or maximum) of a given functional of model estimation F(. ). We determine the optimum model as the proper regression model which contains only the variables (regressors) from the chosen subset.

3. The Task Solving 3.1. The exact solving methods It is proposed to formulate and solve the task of selection the optimum regression model as a task of discrete optimization with Boolean variables on special graph (I, U). The set I of the graph nodes is formed in such a way. The graph node i is set in accordance with any independent variable Xi (i=1,...m). Two other nodes are added: the node (zero) and the node m+1. The set of arcs U of the graph is built as follows. The node 0 is linked to every node i (i=1,...m) and arc (0,i). Further, every node i is connected with every node j, j>i, arc (i, j). It is proposed to formulate and solve the task of selection the optimum regression model as a task of discrete optimization with Boolean variables on special graph (I, U). The set I of the graph nodes is formed in such a way. The graph node i is set in accordance with any independent variable Xi (i=1,...m). Two other nodes are added: the node (zero) and the node m+1. The set of arcs U of the graph is built as follows. The node 0 is linked to every node i (i=1,...m) and arc (0,i). Further, every node i is connected with every node j, j>i, arc (i, j).

It is considered that a node i Є I (i=1,...,m) corresponds to the situation when an independent variable Xi is included into regression model. So, any variant of construction of regression model corresponds to the path from the node 0 to the node m+1. The “length” of every such path μ corresponds to the value of the estimation functional of the proper model F(. ). The initial task of construction of optimum regression model (task of selection subset of regressors) reduces to finding the shortest path from the node 0 to the node m+1 on the graph (I, U). It is considered that a node i Є I (i=1,...,m) corresponds to the situation when an independent variable Xi is included into regression model. So, any variant of construction of regression model corresponds to the path from the node 0 to the node m+1. The “length” of every such path μ corresponds to the value of the estimation functional of the proper model F(. ). The initial task of construction of optimum regression model (task of selection subset of regressors) reduces to finding the shortest path from the node 0 to the node m+1 on the graph (I, U). Graph (I, U) allows to organize the searching variants for selection the optimum regression model consistently by an economical procedure. Finding the shortest path on the built graph (I, U) gives the solution of basic tasks of selection the optimum regression model. Graph (I, U) allows to organize the searching variants for selection the optimum regression model consistently by an economical procedure. Finding the shortest path on the built graph (I, U) gives the solution of basic tasks of selection the optimum regression model.

It is considered that a node corresponds to the situation when an independent variable is included into regression model. So, any variant of construction of regression model corresponds to the path from the node to the node. The “length” of every such path corresponds to the value of the estimation functional of the proper model F(. ). The initial task of construction of optimum regression model (task of selection subset of regressors) reduces to finding the shortest path from the node to the node on the graph. It is considered that a node corresponds to the situation when an independent variable is included into regression model. So, any variant of construction of regression model corresponds to the path from the node to the node. The “length” of every such path corresponds to the value of the estimation functional of the proper model F(. ). The initial task of construction of optimum regression model (task of selection subset of regressors) reduces to finding the shortest path from the node to the node on the graph. Graph allows to organize the searching variants for selection the optimum regression model consistently by an economical procedure. Finding the shortest path on the built graph gives the solution of basic tasks of selection the optimum regression model. Graph allows to organize the searching variants for selection the optimum regression model consistently by an economical procedure. Finding the shortest path on the built graph gives the solution of basic tasks of selection the optimum regression model.

3.2. The genetic algorithm of the task solving The genetic algorithm of decision short cut search task is offered on a graph as a method of decision of task of construction of optimum regression model, based on the ideas of classic genetic algorithm. According to the terminology and definitions of the genetic algorithms theory, every path from the node to the node on graph (that is variant of solving the regression model selection), which the Boolean vector of dimension corresponds to, is the chromosome of proper population and every element of this Boolean vector is the gene of this chromosome. The genetic algorithm of decision short cut search task is offered on a graph as a method of decision of task of construction of optimum regression model, based on the ideas of classic genetic algorithm. According to the terminology and definitions of the genetic algorithms theory, every path from the node to the node on graph (that is variant of solving the regression model selection), which the Boolean vector of dimension corresponds to, is the chromosome of proper population and every element of this Boolean vector is the gene of this chromosome.

We will define the function of adaptability P() for every chromosome on the basis of the special goal function F(. ) of all the chromosomes of population as the following:P() = F( )/ (SUM F(. ) – F()), (3) We will define the function of adaptability P() for every chromosome on the basis of the special goal function F(. ) of all the chromosomes of population as the following:P() = F( )/ (SUM F(. ) – F()), (3) where SUM F(. ) is an operator of summing for all the chromosomes of population. where SUM F(. ) is an operator of summing for all the chromosomes of population. It is possible to proof that if some genotype (chromosome) corresponds to a local (global) optimum for the basic task of selection the optimum regression model (for the special goal function F(. )) then this genotype (chromosomes) gives the local (global) optimum for the adjustment function P(. ) also and vice versa. And also if 1 and 2 are the chromosomes of proper population and F(1) < F(2) then F(1)/ F(2) < P (1) / P (2). It is possible to proof that if some genotype (chromosome) corresponds to a local (global) optimum for the basic task of selection the optimum regression model (for the special goal function F(. )) then this genotype (chromosomes) gives the local (global) optimum for the adjustment function P(. ) also and vice versa. And also if 1 and 2 are the chromosomes of proper population and F(1) < F(2) then F(1)/ F(2) < P (1) / P (2).

The offered genetic method is an iterative algorithmic process. Lets describe iteration (t=1,2,...) calculation process of this algorithm. After implementation of previous iteration t-1, for implementation of iteration we have the great numbers of population m of types:. The offered genetic method is an iterative algorithmic process. Lets describe iteration (t=1,2,...) calculation process of this algorithm. After implementation of previous iteration t-1, for implementation of iteration we have the great numbers of population m of types:. W1(t-1), W2(t-1),...,Wk(t-1),...Wm(t-1). W1(t-1), W2(t-1),...,Wk(t-1),...Wm(t-1). In the offered models of genetic algorithm the procedure of constructions of initial great number of population of type Wk(0), (k=1,...,m) is important. Great numbers of population W1(0), W2(0),...,Wk(0),...Wm(0) for realization of first iteration (t=1) are determined by the proper procedure of determination of these initial great numbers, which is carried out by the special variant of method marks [5] on partial graphs (Ik, Uk), with the use of probable sensors for „equalizing” of variants of decisions (chromosomes) for these great numbers [6]. In the offered models of genetic algorithm the procedure of constructions of initial great number of population of type Wk(0), (k=1,...,m) is important. Great numbers of population W1(0), W2(0),...,Wk(0),...Wm(0) for realization of first iteration (t=1) are determined by the proper procedure of determination of these initial great numbers, which is carried out by the special variant of method marks [5] on partial graphs (Ik, Uk), with the use of probable sensors for „equalizing” of variants of decisions (chromosomes) for these great numbers [6].

STEP І (Large cycle). Choice of the type of population. The population is the set of chromosomes (paths) which necessarily contain a node (variable) with the numbers and larger than. Choice of the type of population. The population is the set of chromosomes (paths) which necessarily contain a node (variable) with the numbers and larger than.

STEP ІІ (Small cycle). Step 1. Presentation for iteration t of calculation genetic process of great number Wk(t-1) of population of type k, which contains Nk persons (chromosomes). Calculation of values of the goal function F(. ) and the function of adaptability P(.) for populations (chromosomes) from the great number Wk(0). Step 1. Presentation for iteration t of calculation genetic process of great number Wk(t-1) of population of type k, which contains Nk persons (chromosomes). Calculation of values of the goal function F(. ) and the function of adaptability P(.) for populations (chromosomes) from the great number Wk(0). Step 2. Choice of „parents” from the great number Wk(t-1) of populations (chromosomes) of type k on the basis of combination of two principles: near and distant „cognation”. Under „cognation” of chromosomes „parents” means distance between the members of population in the value of measure Hamming (hamming distance) between the sets (by the vectors of Boolean variables) of chromosomes „parents”. Step 2. Choice of „parents” from the great number Wk(t-1) of populations (chromosomes) of type k on the basis of combination of two principles: near and distant „cognation”. Under „cognation” of chromosomes „parents” means distance between the members of population in the value of measure Hamming (hamming distance) between the sets (by the vectors of Boolean variables) of chromosomes „parents”.

STEP ІІ (Small cycle). Step 3. The crossing over pairs of chromosomes of “parents” for the generation of chromosomes of “posterity’s”. Step 3. The crossing over pairs of chromosomes of “parents” for the generation of chromosomes of “posterity’s”. Step 4. Calculation of values of the goal function F(. ) and the function of adaptability P(. ) for the chromosomes of “posterity’s”. Step 4. Calculation of values of the goal function F(. ) and the function of adaptability P(. ) for the chromosomes of “posterity’s”. Step 5. Excepting the procedure of mutation for the chromosomes of “posterity’s” for the generation of “mutants” chromosomes. Step 5. Excepting the procedure of mutation for the chromosomes of “posterity’s” for the generation of “mutants” chromosomes. Step 6. Calculation of values of goal function F(. ) and function of adaptability P(.) for the “mutants” chromosomes. Step 6. Calculation of values of goal function F(. ) and function of adaptability P(.) for the “mutants” chromosomes. Step 7. Checking end of process of calculating of the procedures of “crossing over pairs” and iteration pf the type. Step 7. Checking end of process of calculating of the procedures of “crossing over pairs” and iteration pf the type.

STEP ІІ (Small cycle). Step 8. Conducting of the procedure of migration between by the great numbers of population of type k and other types. Step 8. Conducting of the procedure of migration between by the great numbers of population of type k and other types. Step 9. Calculation for chromosomes „migrants” of function having a special purpose and function of adaptability P(. ). Step 9. Calculation for chromosomes „migrants” of function having a special purpose and function of adaptability P(. ). Step 10. Conducting the selection procedure for the generation of great number Wk(t). Step 10. Conducting the selection procedure for the generation of great number Wk(t). Step 11. Transition to begin small cycle for choosing next type of population (with verification of condition k<m). If k=m, we pass to the large cycle for realization of next iteration (t+1). Step 11. Transition to begin small cycle for choosing next type of population (with verification of condition k<m). If k=m, we pass to the large cycle for realization of next iteration (t+1).

The model of genetic algorithm offered above is a so called „island model” [7], in which the general great number of population is broken down into several different parts (islands) of types of population. Each of these parts of general great number of population will separately develop on the basis of a genetic algorithm. The proper procedure of migration will take place between them. The ”island„ model allows to start a genetic algorithm at once several times for the best global decision. The model of genetic algorithm offered above is a so called „island model” [7], in which the general great number of population is broken down into several different parts (islands) of types of population. Each of these parts of general great number of population will separately develop on the basis of a genetic algorithm. The proper procedure of migration will take place between them. The ”island„ model allows to start a genetic algorithm at once several times for the best global decision.

4. Conclusion Presentation of task of search of optimum regression model on special graph allows the task of discrete optimization effectively to use the tool of technology of search of short cut on this graph. Presentation of task of search of optimum regression model on special graph allows the task of discrete optimization effectively to use the tool of technology of search of short cut on this graph. For arbitrary numbers of primary variables (regressors), including the enough large number, a special heuristic method is offered. The proposed genetic algorithm of solving the task of selection the optimum regression model does not guarantee the exact solution but gives an enough satisfactory one. For arbitrary numbers of primary variables (regressors), including the enough large number, a special heuristic method is offered. The proposed genetic algorithm of solving the task of selection the optimum regression model does not guarantee the exact solution but gives an enough satisfactory one. The offered genetic algorithm of decision of task of construction of optimum regression model from the class of so called «island models» allows effectively to decide the general global task of optimization by decomposition of calculation process of its decision into process of decision of separate local tasks. The offered genetic algorithm of decision of task of construction of optimum regression model from the class of so called «island models» allows effectively to decide the general global task of optimization by decomposition of calculation process of its decision into process of decision of separate local tasks.

References 1. Holland J. H.: Adaptation in Natural and Artificial Systems. – Ann Arbor: Michigan University Press, Goldberg D. S.: Genetic algorithms in search, optimization, and machine learning. – Reading, MA: Addison-Wesley, Melnyk I.M., Stepashko V.S.: On an approach to the optimum regression model selection using the discrete optimization method. – Proceedings of the International Workshop on Inductive Modelling. – Kyiv: IRTC ITS, – P (In Ukrainian) 4. Ivakhnenko A. G., Stepashko V.S.: Pomekhoustoichivost modelirovania (Noise-immunity of modeling). Kiev: Naukova Dumka, 216 p, (In Russian) 5. Yermoliev Yu.M., Melnyk I.M. Extremalnye zadachi na grafakh (Extreme tasks on the graphs). – Kiev: Naukova Dumka, – 159 с. (In Russian) 6. Melnyk I.M. Genetic algorithm for solving the task of an optimum regression model construction as a discrete optimization problem­. – Problems of Control and Informatics. – – №3. – P (In Russian) 7. Whitley W.D., Rana S.B., and Heckendorn R.B. Island model genetic algorithms and linearly separable problems. – Selected Papers from AISB Workshop on Evolutionary Computing. – London: Springer-Verlag, – P

Thank you!