A Parallel BOA-PSO Hybrid Algorithm for History Matching

Slides:



Advertisements
Similar presentations
Heuristic Search techniques
Advertisements

Running a model's adjoint to obtain derivatives, while more efficient and accurate than other methods, such as the finite difference method, is a computationally.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
Linkage Problem, Distribution Estimation, and Bayesian Networks Evolutionary Computation 8(3) Martin Pelikan, David E. Goldberg, and Erick Cantu-Paz.
Gizem ALAGÖZ. Simulation optimization has received considerable attention from both simulation researchers and practitioners. Both continuous and discrete.
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Estimation of Distribution Algorithms Ata Kaban School of Computer Science The University of Birmingham.
Evaluating Hypotheses
Lecture 10 Comparison and Evaluation of Alternative System Designs.
Classification: Internal Status: Draft Using the EnKF for combined state and parameter estimation Geir Evensen.
Radial Basis Function Networks
Cristian Urs and Ben Riveira. Introduction The article we chose focuses on improving the performance of Genetic Algorithms by: Use of predictive models.
Estimation of Distribution Algorithms (EDA)
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Bayesian Reasoning: Tempering & Sampling A/Prof Geraint F. Lewis Rm 560:
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
Optimization by Model Fitting Chapter 9 Luke, Essentials of Metaheuristics, 2011 Byung-Hyun Ha R1.
Reservoir Uncertainty Assessment Using Machine Learning Techniques Authors: Jincong He Department of Energy Resources Engineering AbstractIntroduction.
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.
Evolving RBF Networks via GP for Estimating Fitness Values using Surrogate Models Ahmed Kattan Edgar Galvan.
Paging Area Optimization Based on Interval Estimation in Wireless Personal Communication Networks By Z. Lei, C. U. Saraydar and N. B. Mandayam.
Sporadic model building for efficiency enhancement of the hierarchical BOA Genetic Programming and Evolvable Machines (2008) 9: Martin Pelikan, Kumara.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
1 Comparative Study of two Genetic Algorithms Based Task Allocation Models in Distributed Computing System Oğuzhan TAŞ 2005.
Grid-Based Genetic Algorithm Approach to Colour Image Segmentation Marco Gallotta Keri Woods Supervised by Audrey Mbogho.
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Breeding Swarms: A GA/PSO Hybrid 簡明昌 Author and Source Author: Matthew Settles and Terence Soule Source: GECCO 2005, p How to get: (\\nclab.csie.nctu.edu.tw\Repository\Journals-
Tom.h.wilson Department of Geology and Geography West Virginia University Morgantown, WV.
Tom.h.wilson Department of Geology and Geography West Virginia University Morgantown, WV.
Stats Methods at IC Lecture 3: Regression.
Chapter 14 Genetic Algorithms.
Alan P. Reynolds1. , Asaad Abdollahzadeh2, David W
Discrete ABC Based on Similarity for GCP
Michigan Technological University
Sofus A. Macskassy Fetch Technologies
Unit 5: Hypothesis Testing
Chapter 9: Testing a Claim
Rule Induction for Classification Using
Amit Suman and Tapan Mukerji
A paper on Join Synopses for Approximate Query Answering
Presented by: Dr Beatriz de la Iglesia
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Data Mining (and machine learning)
Forecasting The Future of Movies
CONCEPTS OF HYPOTHESIS TESTING
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Assessing uncertainties on production forecasting based on production Profile reconstruction from a few Dynamic simulations Gaétan Bardy – PhD Student.
A RESERVOIR MANAGEMENT REVOLUTION
Bayesian Models in Machine Learning
Discrete Event Simulation - 4
Upscaling of 4D Seismic Data
Geology Geomath Chapter 7 - Statistics tom.h.wilson
I. Statistical Tests: Why do we use them? What do they involve?
Significance Tests: The Basics
Chapter 9: Testing a Claim
Principles of the Global Positioning System Lecture 11
Chapter 6 Network Flow Models.
Chapter 9: Testing a Claim
Chapter 9: Testing a Claim
Chapter 9: Testing a Claim
DESIGN OF EXPERIMENTS by R. C. Baker
Unit 5: Hypothesis Testing
Chapter 5: Sampling Distributions
Population Methods.
Design Issues Lecture Topic 6.
Presentation transcript:

A Parallel BOA-PSO Hybrid Algorithm for History Matching Alan Reynolds, Asaad Abdollahzadeh, David Corne, Mike Christie, Brian Davies and Glyn Williams We’ll start by describing the motivation for history matching, and what history matching is.

Development Optimization: An Overview We have access to a reservoir model and a reservoir simulator. We wish to optimize a plan for development of the reservoir. Can we apply a metaheuristic to the task of development optimization, using the simulator and the model to evaluate each solution? Well, the answer is “yes but…”. If the model accurately describes what is going on underground, and if the simulator is also accurate, then development optimization would simply be a fairly typical optimization problem – albeit one with an expensive evaluation function. However, the main problem is that there must be considerable uncertainty in the reservoir model – subsurface geology cannot be directly observed.

History Matching: An Overview We have access to: A reservoir model, with considerable uncertainties; A reservoir simulator; Pressure, production and saturation data collected during the lifetime of the reservoir. We wish to adjust the values of the reservoir model parameters, in order to produce alternative, improved reservoir models. Model quality is determined via simulation and comparison of the results with the collected data. So a model is evaluated according to a misfit measure and a metaheuristic can be used to optimize the model. In this paper we focus solely on minimizing misfit. Note though, that in reality we actually want to find a range of diverse solutions, with misfits providing a measure of the likelihood of each of the models. The set of models can then used not only to make predictions but also to give an idea of the uncertainty surrounding the predictions.

History Matching and Dev. Optimization Production Data (Rates) This is an optimization process History Dataset Other Obsvtns (PLTs, RFTs) Accuracy Criteria History Matching Static Data (Structure, Rocks) Uncertainties Multiple Models Run Model Output Filter Models Selected Models Formal & Informal Uncertainties ? Regenerate This is an optimization process too Economic Model Development Planning Dave Output (per model) w/ uncertainties Multiple Plans Chosen development plan w/uncertainties Development rules and constraints Run Rank & Risk Results Regenerate

PUNQ-S3 A synthetic problem based on a real field example. Pressure, volume, temperature and aquifer data are taken from the real-world model, while porosity and permeability fields for the ‘truth’ case were generated randomly, but in a manner so as to be as consistent as possible with the geological model. Simulation is used to produce the production data (bottom hole pressure, water cut, gas-oil ratio), after which Gaussian noise is added to simulate measurement error. Simulation model contains 2660 (19x28x5) grid blocks, of which 1761 are active, and 6 production wells. This make the model reasonable in size and runtime and suitable for experimentation with algorithm parameters. The field is structurally bounded to the east and south by a fault, while the link to a strong aquifer to the north and west means that no injection wells are required. The field initially has a small gas cap at the centre of the structure, and production wells are located around this gas cap. We optimize the value of 45 parameters – porosity values in 5 layers, each of which are divided into 9 regions. Permeability values are calculated using the known correlation between porosity and permeability.

Koma A real world problem, with real world data. 32,000 active blocks in 57 layers – so simulation takes a lot longer. (1000 simulations in 24 hours rather than 3000 in about 30 minutes, using 20 nodes on a Beowulf cluster). Colours show ratio of net reservoir rock to total rock. 3 production wells (in green) and 3 injectors. We have selected 54 parameters to be optimized - 16 pore volume multipliers, 8 permeability multipliers, 21 inter-region transmissibility multipliers, 8 fault transmissibility multipliers, and 1 aquifer volume multiplier.

Estimation of Distribution Algorithms Success of genetic algorithms is sometimes attributed to the building block hypothesis. However, in practice, genetic operators frequently destroy the building blocks. An alternative approach: explicitly identify high quality building blocks. Estimation of distribution algorithms (EDAs) create an explicit model of the location of good solutions. The model is used to generate new solutions. New good solutions are used to update the model. The population implicitly contains information about what building blocks seem to produce better quality results.

UMDA The Univariate Marginal Distribution Algorithm is one of the most basic EDAs. Consider a problem with solutions consisting of 5 bits. Rather than evolve solutions, UMDA evolves a model consisting of 5 probability values. Each value is the probability that, in a good solution, the corresponding bit is (or should be) set. Initially, each value is set to 0.5. Here we see the operation of one iteration of UMDA.

BOA Bayesian Optimization Algorithm. Note that each iteration requires the optimization of the Bayesian network structure, which can take some time. However, this is easily outweighed by the amount of time required to perform reservoir simulations. Indeed, the cost of reservoir simulations is what led us to use more sophisticated but computation heavy optimization algorithms. Results using BOA showed a improvement over a highly tuned genetic algorithm used in industry. Similar results could be obtained using only two thirds of the number of solution evaluations.

PSO 𝑣𝑖𝑗 ← 𝑤𝑣𝑖𝑗 + 𝛼𝑟1 (𝑝𝑖𝑗 − 𝑥𝑖𝑗) + 𝛽𝑟2 (𝑔𝑗 − 𝑥𝑖𝑗) 𝑣𝑖𝑗 ← min (𝑣𝑖𝑗, 𝑉𝑚𝑎𝑥,𝑗) 𝑣𝑖𝑗 ← max (𝑣𝑖𝑗 ,−𝑉𝑚𝑎𝑥,𝑗) 𝑥𝑖𝑗 ← 𝑥𝑖𝑗 + 𝑣𝑖𝑗 Particle Swarm Optimization. PSO has also been shown to outperform the industry used GA. Indeed, it also outperforms BOA.

BOA-PSO hybrid Initial BOA population generated by: initializing a swarm of, for example, 50 particles at random; running the PSO until the initial population of, for example, 500 solutions has been generated. Subsequent ‘child’ populations produced by: generating, for example, 5 seed solutions using the Bayesian network; injecting the 5 seed solutions into the swarm; continuing PSO until the child population (of say 500 solutions) is full. Why look at hybrid algorithms for history matching? The focus of our project was the application of EDAs to history matching, yet the relatively simple PSO produced better results. We wondered whether the strengths of the two algorithms could be combined. Also, the trade-off between getting a very good solution according to misfit and getting many, not quite so good solutions, suggested combining two algorithms where one provides the diversity while the other provides the drive toward the best solutions. We wanted to combine the two algorithms so that this trade-off could be easily controlled – we wanted to be able to have 100% PSO, 0% BOA, or 0% PSO, 100% BOA, or any point in between. The algorithm is parallelized synchronously for the BOA component, and asynchronously for the PSO.

Results: PUNQ-S3 Population size: 200. Number of parents and number of children: 100. Swarm size: 50 Fifty runs per setting.

Results: Koma Based on 5 runs each, due to the much greater time requirements of the reservoir simulation.

Conclusions and Further Research BOA-PSO hybrid clearly outperforms BOA on its own, on both problems. On PUNQ-S3, results using the hybrid are marginally better than those obtained by PSO. The hybrid algorithm continues to outperform PSO on the larger, real-world instance (Koma). There is plenty of room for experimentation with and improvement to the hybrid algorithm. And since BOA has already been shown to outperform the commercialized GA used in industry, this is a good result. Though not with any statistical significance – there is enough to warrant further exploration, but little more. In particular, a method that is less disruptive to the operation of the PSO would have BOA suggest new or additional ‘targets’ for the PSO, rather than replacing solutions. Alternatively, we could use DE rather than PSO. (DE has already been shown to be a promising approach.)