2000-2001 NPACI Alpha Project Review: Cellular Microphysiology on the Data Grid Fran Berman, UCSD Tom Bartol, Salk Institute.

Slides:



Advertisements
Similar presentations
Nimrod/G GRID Resource Broker and Computational Economy
Advertisements

Building a CFD Grid Over ThaiGrid Infrastructure Putchong Uthayopas, Ph.D Department of Computer Engineering, Faculty of Engineering, Kasetsart University,
A Proposal of Capacity and Performance Assured Storage in The PRAGMA Grid Testbed Yusuke Tanimura 1) Hidetaka Koie 1,2) Tomohiro Kudoh 1) Isao Kojima 1)
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Scheduling in Distributed Systems Gurmeet Singh CS 599 Lecture.
From Grid to Global Computing: Deploying Parameter Sweep Applications Henri Casanova Grid Research And Innovation Laboratory (GRAIL)
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Achieving Application Performance on the Information Power Grid Francine Berman U. C. San Diego and NPACI This presentation will probably involve audience.
Performance Prediction Engineering Francine Berman U. C. San Diego Rich Wolski U. C. San Diego and University of Tennessee This presentation will probably.
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
Achieving Application Performance on the Computational Grid Francine Berman U. C. San Diego This presentation will probably involve audience discussion,
CSE 160/Berman Programming Paradigms and Algorithms W+A 3.1, 3.2, p. 178, 6.3.2, H. Casanova, A. Legrand, Z. Zaogordnov, and F. Berman, "Heuristics.
Adaptive Computing on the Grid – The AppLeS Project Francine Berman U.C. San Diego.
AppLeS, NWS and the IPG Fran Berman UCSD and NPACI Rich Wolski UCSD, U. Tenn. and NPACI This presentation will probably involve audience discussion, which.
MCell Usage Scenario Project #7 CSE 260 UCSD Nadya Williams
Achieving Application Performance on the Computational Grid Francine Berman This presentation will probably involve audience discussion, which will create.
NetSolve Henri Casanova and Jack Dongarra University of Tennessee and Oak Ridge National Laboratory
The AppLeS Project: Harvesting the Grid Francine Berman U. C. San Diego This presentation will probably involve audience discussion, which will create.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
AppLeS / Network Weather Service IPG Pilot Project FY’98 Francine Berman U. C. San Diego and NPACI Rich Wolski U.C. San Diego, NPACI and U. of Tennessee.
Workload Management Massimo Sgaravatto INFN Padova.
New Development in the AppLeS Project or User-Level Middleware for the Grid Francine Berman University of California, San Diego.
Achieving Application Performance on the Grid: Experience with AppLeS Francine Berman U. C., San Diego This presentation will probably involve audience.
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
Nimrod/G GRID Resource Broker and Computational Economy David Abramson, Rajkumar Buyya, Jon Giddy School of Computer Science and Software Engineering Monash.
Parallel Tomography Shava Smallen CSE Dept. U.C. San Diego.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
Active Monitoring in GRID environments using Mobile Agent technology Orazio Tomarchio Andrea Calvagna Dipartimento di Ingegneria Informatica e delle Telecomunicazioni.
NATIONAL PARTNERSHIP FOR ADVANCED COMPUTATIONAL INFRASTRUCTURE Molecular Science in NPACI Russ B. Altman NPACI Molecular Science Thrust Stanford Medical.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Nimrod & NetSolve Sathish Vadhiyar. Nimrod Sources/Credits: Nimrod web site & papers.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
A Survey of Distributed Task Schedulers Kei Takahashi (M1)
Development Timelines Ken Kennedy Andrew Chien Keith Cooper Ian Foster John Mellor-Curmmey Dan Reed.
EFFECTIVE LOAD-BALANCING VIA MIGRATION AND REPLICATION IN SPATIAL GRIDS ANIRBAN MONDAL KAZUO GODA MASARU KITSUREGAWA INSTITUTE OF INDUSTRIAL SCIENCE UNIVERSITY.
1 Logistical Computing and Internetworking: Middleware for the Use of Storage in Communication Micah Beck Jack Dongarra Terry Moore James Plank University.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
Logistical Networking Micah Beck, Research Assoc. Professor Director, Logistical Computing & Internetworking (LoCI) Lab Computer.
1 Andreea Chis under the guidance of Frédéric Desprez and Eddy Caron Scheduling for a Climate Forecast Application ANR-05-CIGC-11.
Problem Solving with NetSolve Michelle Miller, Keith Moore,
Service - Oriented Middleware for Distributed Data Mining on the Grid ,劉妘鑏 Antonio C., Domenico T., and Paolo T. Journal of Parallel and Distributed.
1 Mobile Management of Network Files Alex BassiMicah Beck Terry Moore Computer Science Department University of Tennessee.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
1 Alexandru V Staicu 1, Jacek R. Radzikowski 1 Kris Gaj 1, Nikitas Alexandridis 2, Tarek El-Ghazawi 2 1 George Mason University 2 George Washington University.
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
Automatic Statistical Evaluation of Resources for Condor Daniel Nurmi, John Brevik, Rich Wolski University of California, Santa Barbara.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
1 On-line Parallel Tomography Shava Smallen UCSD.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Adaptive Computing on the Grid Using AppLeS Francine Berman, Richard Wolski, Henri Casanova, Walfredo Cirne, Holly Dail, Marcio Faerman, Silvia Figueira,
A Technical Overview Bill Branan DuraCloud Technical Lead.
National Institute of Advanced Industrial Science and Technology Developing Scientific Applications Using Standard Grid Middleware Hiroshi Takemiya Grid.
Cyber-Research: Meeting the Challenge of a Terascale Computing Infrastructure Francine Berman Department of Computer Science and Engineering, U. C. San.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
10 May 2001WP6 Testbed Meeting1 WP5 - Mass Storage Management Jean-Philippe Baud PDP/IT/CERN.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
Parallel Tomography Shava Smallen SC99. Shava Smallen SC99AppLeS/NWS-UCSD/UTK What are the Computational Challenges? l Quick turnaround time u Resource.
Resource Optimization for Publisher/Subscriber-based Avionics Systems Institute for Software Integrated Systems Vanderbilt University Nashville, Tennessee.
- DAG Scheduling with Reliability - - GridSolve - - Fault Tolerance In Open MPI - Asim YarKhan, Zhiao Shi, Jack Dongarra VGrADS Workshop April 2007.
VGES Demonstrations Andrew A. Chien, Henri Casanova, Yang-suk Kee, Richard Huang, Dionysis Logothetis, and Jerry Chou CSE, SDSC, and CNS University of.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Achieving Application Performance on the Computational Grid Francine Berman U. C. San Diego and NPACI This presentation will probably involve audience.
Workload Management Workpackage
Q: What Does the Future Hold for “Parallel” Languages?
Scheduled Accomplishments
Wide Area Workload Management Work Package DATAGRID project
Presentation transcript:

NPACI Alpha Project Review: Cellular Microphysiology on the Data Grid Fran Berman, UCSD Tom Bartol, Salk Institute

“MCell” Alpha project Project leaders: Terry Sejnowski, Salk Institute, Fran Berman, UCSD Senior Participants: –Tom Bartol, Salk –Joel Stiles, CMU (leveraged) –Edwin Salpeter, Cornell (leveraged) –Jack Dongarra, Rich Wolski, U. of Tenn. –Mark Ellisman, UCSD NCMIR –Henri Casanova, UCSD CSE (leveraged)

MCell Alpha Goals General goal: Implementation and deployment of MCell, a general Monte Carlo simulator of cellular microphysiology using NPACI high performance and distributed resources Specific goals: 1.Develop a Grid-enabled version of MCell available to all MCell and NPACI users 2.Develop an MPI/Open MP version suitable for MPP platforms such as Blue Horizon 3.Perform large-scale runs necessary for new disciplinary results 4.Extend the prototype and tech transfer APST user-level middleware for deploying MCell and similar parameter sweep applications to NPACI partners

MCell Alpha Project Previous Accomplishments –Prototype of Grid-enabled MCell code (via APST Grid middleware) developed SW integrates NetSolve, AppLeS, NWS –Initial MCell runs performed on Blue Horizon Agenda for this presentation: –Tom: What is MCell and what are its computational requirements? –Fran: How do we develop SW for performance- efficient distributed MCell runs? –FY plans and feedback

Tom’s Presentation

Grid-enabled MCell Previous work –Have developed prototype of APST (AppLeS Parameter Sweep Template) which can be used to deploy MCell in wide-area Grid environments Includes mechanism for targeting available services at remote resources (NetSolve, Globus, GASS, IBP, NWS) –Have developed a Grid MCell performance model –Have developed performance-efficient Grid-oriented scheduling heuristics for MCell NPACI Alpha Project Goals 1.Develop a Grid-enabled version of MCell with enhanced scheduling algorithms, I/O, and data storage model available targeted to MCell users and NPACI resources. 2.Extend the prototype and tech transfer APST user-level middleware for deploying MCell and similar parameter sweep applications to NPACI partners 3.Develop an MPI/Open MP version of MCell suitable for MPP platforms such as Blue Horizon 4.Perform large-scale runs necessary for new disciplinary results

Grid-enabled MCell Have performed initial wide-area MCell runs using APST prototype APST –APST = AppLeS Parameter Sweep Template –MCell used as driving application –Developed as user-level Grid middleware for scheduling and deploying MCell and other parameter sweep applications –Joint work with Henri Casanova –Research supported by NASA and NSF

Scheduling Issues for MCell Large shared files may complicate the scheduling process Post-processing must minimize file transfer time Adaptive scheduling necessary to account for dynamic environment

Contingency Scheduling: Allocation developed by dynamically generating a Gantt chart for scheduling unassigned tasks between scheduling events Basic skeleton 1.Compute the next scheduling event 2.Create a Gantt Chart G 3.For each computation and file transfer currently underway, compute an estimate of its completion time and fill in the corresponding slots in G 4.Select a subset T of the tasks that have not started execution 5.Until each host has been assigned enough work, heuristically assign tasks to hosts, filling in slots in G 6.Implement schedule Scheduling Approach used for MCell Network links Hosts (Cluster 1) Hosts (Cluster 2) Time Resources Computation G Scheduling event Scheduling event Computation

NetSolve Globus Legion NWS Ninf IBP Condor APST/MCell user-level middleware transport APIexecution API metadata API scheduler API Grid Resources and Middleware APST/MCell Daemon GASSIBP NFS GRAMNetSolve Condor, Ninf, Legion,.. NWS Workqueue Gantt chart heuristic algorithms Workqueue++ MinMinMaxMinSufferageXSufferage APST/MCell Client Controller interacts Command-line client Metadata Bookkeeper Actuator Scheduler triggers transferexecutequery store actuate report retrieve

MCell Computational Challenges Support for large-scale distributed MCell runs Support for large-scale parallel MCell runs Execution of large-scale runs Tech Transfer of APST for NPACI parameter sweep application developers User-level middleware facilitates use of Grid for wider class of users MCell algorithm and SW development allows for new disciplinary results

FY Plans Develop Grid-enabled MCell –Optimize scheduling strategy –Increase sensitivity of model to environmental constraints (data storage, I/O, post-processing, resource location) –Target SW to NPACI resources Robustify and tech transfer more general APST user- level middleware to NPACI metasystem Develop MPP-enabled MPI/OpenMP version of MCell –Adapt algorithm for performance in MPP environment –Develop MPP-enabled APST to efficiently deploy MCell tasks to parallel environments –Implement/deploy SW for large-scale Blue Horizon runs Perform larger-scale runs necessary for new disciplinary results

Feedback It would be easier to perform this work for NPACI if … –Allocation of NS thrust area computer time was more generous –Blue Horizon had larger scratch space –A rendering farm were available –Globus platform was more stable –NWS, NetSolve, and other services were more consistently available at NPACI partner sites

Previous work on Scheduling Heuristics Self-scheduling Algorithms workqueue workqueue w/ work stealing workqueue w/ work duplication... Gantt chart heuristics: MinMin, MaxMin Sufferage, XSufferage... Scheduling Algorithms for MCell Easy to implement and quick No need for performance predictions  Insensitive to data placement  More difficult to implement  Needs performance predictions Sensitive to data placement Simulation results (HCW ’00 paper, SC’00 paper) show that: Gantt chart heuristics are worth it Xsufferage is good heuristic even when predictions are bad Complex environments require better planning (Gantt chart) Gantt Chart Algorithms Min-min Max-min Sufferage, XSufferage

1.How frequent should the scheduling events be? 2.Which set of tasks should we schedule between scheduling events? 3.How accurate do our estimates of computation and data transfer times need to be? 4.What scheduling heuristics should we use? 5.How do input and output location and visualization requirements impact scheduling? Research Scheduling Issues G Network links Hosts (Cluster 1) Hosts (Cluster 2) Time Resources Computation Scheduling event Scheduling event Computation

Features Scheduler can be used for structurally similar set of Parameter Sweep Applications in addition to MCell –INS2D, INS3D (NASA Fluid Dynamics applications) –Tphot (SDSC, Proton Transport application) –NeuralObjects (NSI, Neural Network simulations) –CS simulation applications for our own research (Model validation) Actuator’s APIs are interchangeable and mixable –(NetSolve+IBP) + (GRAM+GASS) + (GRAM+NFS) Scheduler allows for dynamic adaptation, multithreading No Grid software is required –However lack of it (NWS, GASS, IBP) may lead to poorer performance APST is being beta-tested by on NASA IPG and at other sites

Preliminary Results Scheduling Mcell on the Grid Experimental Setting: Mcell simulation with 1,200 tasks: composed of 6 Monte-Carlo simulations input files: 1, 1, 20, 20, 100, and 100 MB 4 scenarios: Initially (a) all input files are only in Japan (b) 100MB files replicated in California (c) in addition, one 100MB file replicated in Tennessee (d) all input files replicated everywhere workqueue Gantt-chart algs

Evaluation of APST MCell Scheduling Heuristics Wanted to evaluate Mcell scheduling heuristics Experiment: –We ran large-sized instances of MCell across a distributed platform and compared execution times with both self-scheduling and Gantt chart heuristics. University of Tennessee, Knoxville NetSolve + IBP University of California, San Diego GRAM + GASS Tokyo Institute of Technology NetSolve + NFS NetSolve + IBP APST Daemon APST Client