1 December 12, 2009 Robust Asynchronous Optimization for Volunteer Computing Grids Department of Computer Science Department of Physics, Applied Physics.

Slides:



Advertisements
Similar presentations
Designing Services for Grid-based Knowledge Discovery A. Congiusta, A. Pugliese, Domenico Talia, P. Trunfio DEIS University of Calabria ITALY
Advertisements

Rolls-Royce supported University Technology Centre in Control and Systems Engineering UK e-Science DAME Project Alex Shenfield
1 An Adaptive GA for Multi Objective Flexible Manufacturing Systems A. Younes, H. Ghenniwa, S. Areibi uoguelph.ca.
Particle Swarm Optimization (PSO)
1 Wendy Williams Metaheuristic Algorithms Genetic Algorithms: A Tutorial “Genetic Algorithms are good at taking large, potentially huge search spaces and.
Multiobjective VLSI Cell Placement Using Distributed Simulated Evolution Algorithm Sadiq M. Sait, Mustafa I. Ali, Ali Zaidi.
Parallelized Evolution System Onur Soysal, Erkin Bahçeci Erol Şahin Dept. of Computer Engineering Middle East Technical University.
A Heuristic Bidding Strategy for Multiple Heterogeneous Auctions Patricia Anthony & Nicholas R. Jennings Dept. of Electronics and Computer Science University.
Department of Engineering, Control & Instrumentation Research Group 22 – Mar – 2006 Optimisation Based Clearance of Nonlinear Flight Control Laws Prathyush.
Introduction to Genetic Algorithms Yonatan Shichel.
Improving Robustness in Distributed Systems Jeremy Russell Software Engineering Honours Project.
Genetic Programming. Agenda What is Genetic Programming? Background/History. Why Genetic Programming? How Genetic Principles are Applied. Examples of.
Genetic Algorithms: A Tutorial
RESEARCH DIRECTIONS IN GRID COMPUTING Dr G Sudha Sadasivam Professor CSE Department, PSG College of Technology.
Particle Swarm Optimization Algorithms
1 Reasons for parallelization Can we make GA faster? One of the most promising choices is to use parallel implementations. The reasons for parallelization.
ADLB Update Recent and Current Adventures with the Asynchronous Dynamic Load Balancing Library Rusty Lusk Mathematics and Computer Science Division Argonne.
COGNITIVE RADIO FOR NEXT-GENERATION WIRELESS NETWORKS: AN APPROACH TO OPPORTUNISTIC CHANNEL SELECTION IN IEEE BASED WIRELESS MESH Dusit Niyato,
Evolutionary algorithms
Genetic Algorithm.
CSM6120 Introduction to Intelligent Systems Other evolutionary algorithms.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
From Analyzing the Tuberculosis Genome to Modeling the Milky Way Galaxy Using Volunteer Computing for Computational Science Travis Desell Department of.
Robust Network Supercomputing with Malicious Processes (Reliably Executing Tasks Upon Estimating the Number of Malicious Processes) Kishori M. Konwar*
Network Aware Resource Allocation in Distributed Clouds.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Highly Distributed Parallel Computing Neil Skrypuch COSC 3P93 3/21/2007.
Hybrid Behavior Co-evolution and Structure Learning in Behavior-based Systems Amir massoud Farahmand (a,b,c) (
What is Genetic Programming? Genetic programming is a model of programming which uses the ideas (and some of the terminology) of biological evolution to.
PSO and its variants Swarm Intelligence Group Peking University.
(Particle Swarm Optimisation)
1 IE 607 Heuristic Optimization Particle Swarm Optimization.
Robin McDougall Scott Nokleby Mechatronic and Robotic Systems Laboratory 1.
1 “Genetic Algorithms are good at taking large, potentially huge search spaces and navigating them, looking for optimal combinations of things, solutions.
Adaptive Web Caching CS411 Dynamic Web-Based Systems Flying Pig Fei Teng/Long Zhao/Pallavi Shinde Computer Science Department.
1 Computing Challenges for the Square Kilometre Array Mathai Joseph & Harrick Vin Tata Research Development & Design Centre Pune, India CHEP Mumbai 16.
Algorithms and their Applications CS2004 ( ) 13.1 Further Evolutionary Computation.
and Volunteer Computing at RPI Travis Desell RCOS, April 23, 2010.
O PTIMAL SERVICE TASK PARTITION AND DISTRIBUTION IN GRID SYSTEM WITH STAR TOPOLOGY G REGORY L EVITIN, Y UAN -S HUN D AI Adviser: Frank, Yeong-Sung Lin.
 Genetic Algorithms  A class of evolutionary algorithms  Efficiently solves optimization tasks  Potential Applications in many fields  Challenges.
Regrouping Particle Swarm Optimization: A New Global Optimization Algorithm with Improved Performance Consistency Across Benchmarks George I. Evers Advisor:
Biologically inspired algorithms BY: Andy Garrett YE Ziyu.
A l a p a g o s : a generic distributed parallel genetic algorithm development platform Nicolas Kruchten 4 th year Engineering Science (Infrastructure.
Particle Swarm Optimization † Spencer Vogel † This presentation contains cheesy graphics and animations and they will be awesome.
SwinTop: Optimizing Memory Efficiency of Packet Classification in Network Author: Chen, Chang; Cai, Liangwei; Xiang, Yang; Li, Jun Conference: Communication.
Innovative and Unconventional Approach Toward Analytical Cadastre – based on Genetic Algorithms Anna Shnaidman Mapping and Geo-Information Engineering.
Particle Swarm Optimization (PSO)
Genetic algorithms: A Stochastic Approach for Improving the Current Cadastre Accuracies Anna Shnaidman Uri Shoshani Yerach Doytsher Mapping and Geo-Information.
An Improved Quantum-behaved Particle Swarm Optimization Algorithm Based on Culture V i   v i 1, v i 2,.. v iD  Gao X. Z 2, Wu Ying 1, Huang Xianlin.
Selection and Recombination Temi avanzati di Intelligenza Artificiale - Lecture 4 Prof. Vincenzo Cutello Department of Mathematics and Computer Science.
Use of Performance Prediction Techniques for Grid Management Junwei Cao University of Warwick April 2002.
An gentle introduction to StupidAlgo Library Yi
A distributed PSO – SVM hybrid system with feature selection and parameter optimization Cheng-Lung Huang & Jian-Fan Dun Soft Computing 2008.
Resource Optimization for Publisher/Subscriber-based Avionics Systems Institute for Software Integrated Systems Vanderbilt University Nashville, Tennessee.
Genetic Algorithms. Solution Search in Problem Space.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Breeding Swarms: A GA/PSO Hybrid 簡明昌 Author and Source Author: Matthew Settles and Terence Soule Source: GECCO 2005, p How to get: (\\nclab.csie.nctu.edu.tw\Repository\Journals-
Swarm Intelligence. Content Overview Swarm Particle Optimization (PSO) – Example Ant Colony Optimization (ACO)
Particle Swarm Optimization (2)
Scientific Research Group in Egypt (SRGE)
Adnan Quadri & Dr. Naima Kaabouch Optimization Efficiency
Particle Swarm Optimization
Robust Asynchronous Optimization Using Volunteer Computing Grids
PSO -Introduction Proposed by James Kennedy & Russell Eberhart in 1995
Meta-heuristics Introduction - Fabien Tricoire
Joseph JaJa, Mike Smorul, and Sangchul Song
Ch 4. The Evolution of Analytic Scalability
Alternative Processor Panel Results 2008
Coevolutionary Automated Software Correction
Presentation transcript:

1 December 12, 2009 Robust Asynchronous Optimization for Volunteer Computing Grids Department of Computer Science Department of Physics, Applied Physics and Astronomy Rensselaer Polytechnic Institute E-Science 2009 December 12, Oxford, UK Travis Desell, Malik Magdon-Ismail, Boleslaw Szymanski, Carlos Varela, Heidi Newberg, Nathan Cole

December 12, Overview Introduction Motivation Driving Scientific Application Asynchronous Genetic Search Why asynchronous? Methodology Recombination Particle Swarm Optimization ‏Generic Optimization Framework Approach Architecture Results Convergence Rates Re-computation rates Conclusions & Future Work Questions?

December 12, Motivation Scientists need easily accessible distributed optimization tools Distribution is essential for scientific computing Scientific models are becoming increasingly complex Rates of data acquisition are far exceeding increases in computing power Traditional optimization strategies not well suited to large scale computing Lack scalability and fault tolerance

December 12, Astro-Informatics Observing from inside the Milky Way provides 3D data: SLOAN digital sky survey has collected over 10 TB data. Can determine it's structure – not possible for other galaxies. Very expensive – evaluating a single model of the Milky Way with a single set of parameters can take hours or days on a typical high-end computer. Models determine where different star streams are in the Milky Way, which helps us understand better its structure and how it was formed. What is the structure and origin of the Milky Way galaxy?

Computed Paths of Sagittarius Stream

December 12, Separation of Concerns Distributed Computing Optimization Scientific Modeling “Plug-and-Play” Simple & generic interfaces Generic Optimization Framework

December 12, Two Distribution Strategies Asynchronous evaluations Results may not be reported or reported late No processor dependencies o Faults can be ignored Grids & Internet Single parallel evaluation Always uses most evolved population Can use traditional methods o Faults require recalculation o Grids require load balancing Supercomputers & Grids

December 12, Asynchronous Architecture‏ Scientific Models Distributed Evaluation Framework Search Routines … Evaluator Creation Data Initialisation Integral Function Integral Composition Likelihood Function Likelihood Composition Initial Parameters Optimised Parameters BOINC (Internet)‏SALSA/Java (RPI Grid)‏ Evaluator (N)‏ Work Request Results Work Request ResultsWork Evolutionary Methods Genetic Search Particle Swarm Optimisation … Evaluator (1)‏

December 12, GMLE Architecture (Parallel-Asynchronous)‏ Results … Distribute Parameters Combine Results Evaluator (1)‏ Evaluator (N)‏ Evaluator (2)‏ Work Request Worker (1)‏ Results … Distribute Parameters Combine Results Evaluator (1)‏ Evaluator (M)‏ Evaluator (2)‏ Work Request Worker (Z)‏ … Communication Layer BOINC - HTTPGrid - TCP/IPSupercomputer - MPI Search Routines MPI

December 12, Issues With Traditional Optimization Traditional global optimization techniques are evolutionary, but dependent on previous steps and are iterative Current population is used to generate the next population Dependencies and iterations limit scalability and impact performance With volatile hosts, what if an individual in the next generation is lost? Redundancy is expensive Scalability limited by population size

December 12, Asynchronous Optimization Strategy Use an asynchronous methodology No dependencies on unknown results No iterations Continuously updated population N individuals are generated randomly for the initial population Fulfil work requests by applying recombination operators to the population Update population with reported results

December 12, Asynchronous Search Strategy‏ Work QueuePopulation Workers Generate members from population Request work when queue is low Parameter Set (1)‏ Parameter Set (2)‏ Parameter Set (n)‏ Fitness (1)‏ Fitness (2)‏ Fitness (n)‏ Unevaluated Parameter Set (1)‏ Unevaluated Parameter Set (2)‏ Unevaluated Parameter Set (m)‏ Report results and update population Send work Request work

December 12, Asynchronous Genetic Search Operators (1)‏ Average Simple operator for continuous problems Generated parameters are the average of two randomly selected parents Mutation Takes a parent and generates a mutation by randomly selecting a parameter and mutating it

December 12, Asynchronous Genetic Search Operators (2)‏ Double Shot - two parents generate three children Average of the parents Outside the less fit parent, equidistant to parent and average Outside the more fit parent, equidistant to parent and average

December 12, Asynchronous Genetic Search Operators (3)‏ Probabilistic Simplex N parents generate one or more children Points placed randomly along the line created by the worst parent, and the centroid (average) of the remaining parents

16 December 12, 2009 Particle Swarm Optimization Particles ‘fly’ around the search space. They move according to their previous velocity and are pulled towards the global best found position and their locally best found position. Analogies: cognitive intelligence (local best knowledge) social intelligence (global best knowledge) 16

17 December 12, 2009 Particle Swarm Optimization PSO: v i (t+1) = w * v i (t) + c 1 * r 1 * (l i - p i (t)) + c 2 * r 2 * (g - p i (t)) p i (t+1) = p i (t) + v i (t+1) w, c 1, c 2 = constants r 1, r 2 = random float between 0 and 1 v i (t) = velocity of particle i at iteration t p i (t) = position of particle i at iteration t l i = best position found by particle i g = global best position found by all particles 17

18 December 12, 2009 Asynchronous PSO Generating new positions does not necessarily require the fitness of the previous position 1. Generate new particle or individual positions to fill work queue 2. Update local and global best on results PSO: If result improves particle’s local best, update local best, particle’s position and velocity of the result 18

19 December 12, 2009 Particle Swarm Optimization (Example) 19 previous: p i (t-1) current: p i (t) local best global best c 1 * (l i - p i (t)) c 2 * (g - p i (t)) w * v i (t) velocity: v i (t) possible new positions

20 December 12, 2009 Particle Swarm Optimization (Example) 20 previous: p i (t-1) current: p i (t) local best global best c 2 * (g - p i (t)) w * v i (t) velocity: v i (t) possible new positions Particle finds a new local best position and the global best position previous: p i (t-1) current: p i (t) local best global best velocity: v i (t) new position

21 December 12, 2009 Particle Swarm Optimization (Example) 21 c 2 * (g - p i (t)) w * v i (t) possible new positions c 1 * (l i - p i (t)) previous: p i (t-1) current: p i (t) local best global best velocity: v i (t) Another particle finds the global best position

22 December 12, Population Fitness (1) Fitness (2) Fitness (n) Individual (1) Individual (2) Individual (n) Unevaluated Individuals Unevaluated Individual (1) Unevaluated Individual (2) Unevaluated Individual (n) Workers (Fitness Evaluation) Report results and update population Request Work Send Work Generate individuals when queue is low Local and global best updated if new individual has better fitness Select individual to generate new individual from in round-robin manner Asynchronous PSO

Computing Environment: BOINC etc >50,000users; 80,000 CPUs; 600 teams; from 99 countries; Second largest BOINC computation (among 100’s) About 500 Teraflops Donate your idle computer time to help perform our calculations. December 12,

24 December 12, 2009 – Growth of Power 24

December 12, Computing Environments - BOINC Multiple Asynchronous Workers Approximately 10,000 – 30,000 volunteered computers engages at a time Asynchronous architecture used Asynchronous Evaluation Volunteered computers can queue up to 20 pending individuals Population updated when results reported Individuals may be reported slowly or not at all

December 12, Handling of Work Units by the BOINC‏ Server

User Participation Users do more than volunteer computing resources (Citizen’s Science): Open-source code gives users access to the application Users have submitted many bug reports, fixes, and performance enhancements A user even created an ATI GPU capable version of the application Forums provide opportunities for users to learn about astronomy and computer science December 12, 2009

Malicious/Incorrect Result Verification With open-source application code, users can compile their own compiler-optimized versions and many do. However, there is also the possibility of users returning malicious results BOINC traditionally uses redundancy on every result to verify their correctness. This requires at least 2 results for every work unit! Asynchronous search doesn't require all work units to be verified, only those which improve the population We reduce the redundancy by comparing a result against the current partial results. December 12, 2009

29 December 12, 2009 Limiting Redundancy (Genetic Search) 29 60% verification found best solutions Increased verification reduces reliability Reliability and convergence by number of parents seems dependent on verification rate

30 December 12, 2009 Limiting Redundancy (PSO) 30 30% verification found best solutions Increased verification reduces reliability Not as dramatically as AGS Lower inertia weights give better results

31 December 12, 2009 Optimization Method Comparison 31 APSO found better solutions than AGS. APSO needed lower verification rates and was less effected by different verification rates.

December 12, Conclusions Asynchronous search is effective on large scale computing environments Fault tolerant without expensive redundancy Asynchronous evaluation on heterogeneous environment increases diversity BOINC converges almost as fast as the BlueGene, while offering more availability and computational power Even computers with slow result report rates are useful Particle Swarm and Simplex-Genetic Hybrid methods provide significant improvement in convergence

December 12, Future Work Optimization Use report times to determine how to generate individuals Simulate asynchrony for benchmarks Automate selection of parameters Distributed Computing Parallel asynchronous workers Handle Malicious “Volunteers” Continued Collaboration

December 12, Questions?