Achieving Application Performance on the Information Power Grid Francine Berman U. C. San Diego and NPACI This presentation will probably involve audience.

Slides:



Advertisements
Similar presentations
Introduction to Grid Application On-Boarding Nick Werstiuk
Advertisements

SSRS 2008 Architecture Improvements Scale-out SSRS 2008 Report Engine Scalability Improvements.
Introduction CSCI 444/544 Operating Systems Fall 2008.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
The Network Weather Service A Distributed Resource Performance Forecasting Service for Metacomputing Rich Wolski, Neil T. Spring and Jim Hayes Presented.
ADAPT An Approach to Digital Archiving and Preservation Technology Principal Investigator: Joseph JaJa Lead Programmers: Mike Smorul and Mike McGann Graduate.
Performance Prediction Engineering Francine Berman U. C. San Diego Rich Wolski U. C. San Diego and University of Tennessee This presentation will probably.
Achieving Application Performance on the Computational Grid Francine Berman U. C. San Diego This presentation will probably involve audience discussion,
Adaptive Computing on the Grid – The AppLeS Project Francine Berman U.C. San Diego.
AppLeS, NWS and the IPG Fran Berman UCSD and NPACI Rich Wolski UCSD, U. Tenn. and NPACI This presentation will probably involve audience discussion, which.
Achieving Application Performance on the Computational Grid Francine Berman This presentation will probably involve audience discussion, which will create.
The Network Weather Service: A Distributed Resource Performance Forecasting Service for Metacomputing, Rich Wolski, Neil Spring, and Jim Hayes, Journal.
The AppLeS Project: Harvesting the Grid Francine Berman U. C. San Diego This presentation will probably involve audience discussion, which will create.
NPACI Alpha Project Review: Cellular Microphysiology on the Data Grid Fran Berman, UCSD Tom Bartol, Salk Institute.
WSN Simulation Template for OMNeT++
AppLeS / Network Weather Service IPG Pilot Project FY’98 Francine Berman U. C. San Diego and NPACI Rich Wolski U.C. San Diego, NPACI and U. of Tennessee.
Mike Smorul Saurabh Channan Digital Preservation and Archiving at the Institute for Advanced Computer Studies University of Maryland, College Park.
GHS: A Performance Prediction and Task Scheduling System for Grid Computing Xian-He Sun Department of Computer Science Illinois Institute of Technology.
Business Intelligence Dr. Mahdi Esmaeili 1. Technical Infrastructure Evaluation Hardware Network Middleware Database Management Systems Tools and Standards.
New Development in the AppLeS Project or User-Level Middleware for the Grid Francine Berman University of California, San Diego.
Maintaining and Updating Windows Server 2008
©Ian Sommerville 2004Software Engineering, 7th edition. Chapter 17 Slide 1 Rapid software development.
Slide 1 of 9 Presenting 24x7 Scheduler The art of computer automation Press PageDown key or click to advance.
CSE 160/Berman Programming Paradigms and Algorithms W+A 3.1, 3.2, p. 178, 5.1, 5.3.3, Chapter 6, 9.2.8, , Kumar Berman, F., Wolski, R.,
Chapter 10 EJB Concepts of EJB Three Components in Creating an EJB Starting/Stopping J2EE Server and Deployment Tool Installation and Configuration of.
The Pursuit for Efficient S/C Design The Stanford Small Sat Challenge: –Learn system engineering processes –Design, build, test, and fly a CubeSat project.
STRATEGIES INVOLVED IN REMOTE COMPUTATION
Overview of SQL Server Alka Arora.
1 DAN FARRAR SQL ANYWHERE ENGINEERING JUNE 7, 2010 SCHEMA-DRIVEN EXPERIMENT MANAGEMENT DECLARATIVE TESTING WITH “DEXTERITY”
Achieving Application Performance on the Grid: Experience with AppLeS Francine Berman U. C., San Diego This presentation will probably involve audience.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
Nimrod/G GRID Resource Broker and Computational Economy David Abramson, Rajkumar Buyya, Jon Giddy School of Computer Science and Software Engineering Monash.
Computer Science Program Center for Entrepreneurship and Information Technology, Louisiana Tech University This presentation will probably involve audience.
Parallel Tomography Shava Smallen CSE Dept. U.C. San Diego.
DISTRIBUTED COMPUTING
Cluster Reliability Project ISIS Vanderbilt University.
Performance Model & Tools Summary Hung-Hsun Su UPC Group, HCS lab 2/5/2004.
Combining the strengths of UMIST and The Victoria University of Manchester Utility Driven Adaptive Workflow Execution Kevin Lee School of Computer Science,
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
Development Timelines Ken Kennedy Andrew Chien Keith Cooper Ian Foster John Mellor-Curmmey Dan Reed.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
1 Geospatial and Business Intelligence Jean-Sébastien Turcotte Executive VP San Francisco - April 2007 Streamlining web mapping applications.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
The System and Software Development Process Instructor: Dr. Hany H. Ammar Dept. of Computer Science and Electrical Engineering, WVU.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GVis: Grid-enabled Interactive Visualization State Key Laboratory. of CAD&CG Zhejiang University, Hangzhou
LEGS: A WSRF Service to Estimate Latency between Arbitrary Hosts on the Internet R.Vijayprasanth 1, R. Kavithaa 2,3 and Raj Kettimuthu 2,3 1 Coimbatore.
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
Automatic Statistical Evaluation of Resources for Condor Daniel Nurmi, John Brevik, Rich Wolski University of California, Santa Barbara.
Architecture View Models A model is a complete, simplified description of a system from a particular perspective or viewpoint. There is no single view.
Adaptive Computing on the Grid Using AppLeS Francine Berman, Richard Wolski, Henri Casanova, Walfredo Cirne, Holly Dail, Marcio Faerman, Silvia Figueira,
Timeshared Parallel Machines Need resource management Need resource management Shrink and expand individual jobs to available sets of processors Shrink.
CS223: Software Engineering Lecture 13: Software Architecture.
Application-level Scheduling Sathish S. Vadhiyar Credits / Sources: AppLeS web pages and papers.
Parallel Tomography Shava Smallen SC99. Shava Smallen SC99AppLeS/NWS-UCSD/UTK What are the Computational Challenges? l Quick turnaround time u Resource.
Origami: Scientific Distributed Workflow in McIDAS-V Maciek Smuga-Otto, Bruce Flynn (also Bob Knuteson, Ray Garcia) SSEC.
UCSD SAN DIEGO SUPERCOMPUTER CENTER Fran Berman Grids in Context Dr. Francine Berman Director, San Diego Supercomputer Center Professor and HPC Endowed.
INTRODUCTION TO GRID & CLOUD COMPUTING U. Jhashuva 1 Asst. Professor Dept. of CSE.
Resource Characterization Rich Wolski, Dan Nurmi, and John Brevik Computer Science Department University of California, Santa Barbara VGrADS Site Visit.
Achieving Application Performance on the Computational Grid Francine Berman U. C. San Diego and NPACI This presentation will probably involve audience.
Connected Infrastructure
Clouds , Grids and Clusters
Connected Infrastructure
Final Project Presentation
واشوقاه إلى رمضان مرحباً رمضان
Evaluation of Data Fusion Methods Using Kalman Filtering and TBM
UmbrellaDB v0.5 Project Report #3
Presentation transcript:

Achieving Application Performance on the Information Power Grid Francine Berman U. C. San Diego and NPACI This presentation will probably involve audience discussion, which will create action items. Use PowerPoint to keep track of these action items during your presentation In Slide Show, click on the right mouse button Select “Meeting Minder” Select the “Action Items” tab Type in action items as they come up Click OK to dismiss this box This will automatically create an Action Item slide at the end of your presentation with your points entered.

IPG = “Distributed Computer” comprising –clusters of workstations –MPPs –remote instruments –visualization sites –data archives for users, performance is key criteria in evaluating platform

Program Performance Current grid programs achieve performance by –dedicating resources –careful staging of computation and data –considerable coordination It must be possible to achieve program performance on the IPG by ordinary users on ordinary days...

Achieving Performance On ordinary days, many users share system resources –load and availability of resources vary –application behavior hard to predict –poor predictions make scheduling hard Challenge: Develop application schedules which can leverage deliverable performance of system at execution time.

Whose Job Is It? Application scheduling can be performed by many entities –Resource scheduler –Job Scheduler –Programmer or User –System Administrator –Application Scheduler

Scheduling and Performance Goal of scheduling application is to promote application performance Achieving application performance can conflict with achieving performance for other system components –Resource Scheduler -- perf measure is utilization –Job Scheduler -- perf measure is throughput –System Administrator -- focuses on system perf –Programmer or User -- may miss most current info –Application Scheduler -- can access most current info

Everything in the system is evaluated in terms of its impact on the application. –performance of each system component can be considered as a measurable quantity –forecasts of quantities relevant to the application can be manipulated to determine schedule This simple paradigm forms the basis for AppLeS. Self-Centered Scheduling

AppLeS Joint project with Rich Wolski AppLeS = Application-Level Scheduler Each application has its own self-centered AppLeS Schedule achieved through –selection of potentially efficient resource sets –performance estimation of dynamic system parameters and application performance for execution time frame –adaptation to perceived dynamic conditions

AppLeS Architecture AppLeS incorporates –application-specific information –dynamic information –prediction Schedule developed to optimize user’s performance measure –minimal execution time –turnaround time = staging/waiting time + execution time –other measures: precision, resolution, speedup, etc. NWS (Wolski) User Prefs App Perf Model Planner Resource Selector Application Act. IPG resources/ infrastructure

Network Weather Service (Wolski) The NWS provides dynamic resource information for AppLeS NWS –monitors current system state –provides best forecast of resource load from multiple models Sensor Interface Reporting Interface Forecaster Model

SARA: An AppLeS-in-Progress SARA = Synthetic Aperture Radar Atlas –application developed at JPL and SDSC Goal: Assemble/process files for user’s desired image –thumbnail image shown to user –user selects desired bounding box within image for more detailed viewing –SARA provides detailed image in variety of formats

Focusing in with SARA Thumbnail imageBounding box

Simple Sara Focuses on obtaining remote data quickly Code developed by Alan Su Compute Server Data Server Data Server Data Server Computation servers and data servers are logical entities, not necessarily different nodes Network shared by variable number of users Computation assumed to be done at compute servers

Simple SARA AppLeS Focus on resource selection problem: Which site can deliver data the fastest? –Data for image accessed over shared networks –Data sets megabytes, representative of SARA file sizes –Servers used for experiments lolland.cc.gatech.edu sitar.cs.uiuc perigee.chpc.utah.edu mead2.uwashington.edu spin.cacr.caltech.edu via vBNS via general Internet

Simple SARA Experiments Ran back-to-back experiments from remote sites to UCSD/PCL Wolski’s Network Weather Service provides forecasts of network load and availability Experiments run during normal business hours mid-week

Which is “Closer”? Sites on the east coast or sites on the west coast? Sites on the vBNS or sites on the general Internet? Consistently the same site or different sites at different times?

Which is “Closer”? Sites on the east coast or sites on the west coast? Sites on the vBNS or sites on the general Internet? Consistently the same site or different sites at different times? Depends a lot on traffic...

Preliminary Results Experiment with larger data set (3 Mbytes) During this time-frame, general Internet provides data mostly faster than vBNS

9/21/98 Experiments Clinton Grand Jury webcast commenced at iteration 62

Experiment with smaller data set (1.4 Mbytes) During this time frame, east coast sites provide data mostly faster than west coast sites More Preliminary Results

Distributed Data Applications SARA representative of larger class of distributed data applications Simple SARA template being extended to accommodate –replicated data sources –multiple files per image –parallel data acquisition –intermediate compute sites –web interface, etc.

SARA AppLeS -- Phase 2 Client, servers are “logical” nodes, which servers should the client use? Client Comp. Server Comp. Server Comp. Server Data Server Data Server Data Server Data Server... Move the computation or move the data? Computation, data servers may “live” at the same nodes Data servers may access the same storage media. How long will data access take when data is needed?

A Bushel of AppLeS … almost During the first “phase” of the project, we’ve focused on getting experience building AppLeS –Jacobi2D, DOT, SRB, Simple SARA, Genetic Algorithm, Tomography,... Using this experience, we are beginning to build AppLeS “templates”/tools for –master/slave applications –parameter sweep applications –distributed data applications –proudly parallel applications, etc. What have we learned...

Lessons Learned from AppLeS Dynamic information is critical

Lessons Learned from AppLeS Program execution and parameters may exhibit a range of performance

Lessons Learned from AppLeS Knowing something about performance predictions can improve scheduling

Lessons Learned from AppLeS Performance of scheduling policy sensitive to application, data, and system characteristics

A First IPG AppLeS Focus on class of parameter sweep applications Building AppLeS template for INS2D that can be used with other applications from class AppLeS INS2D scheduler –first phase focuses on interactive clusters –second phase will target clusters and batch-scheduled platforms –goal is to minimize turnaround time

Parameter Sweep AppLeS Architecture Being developed by Dmitrii Zagorodnov AppLeS schedules work on interactive resources AppLeS tuned to leverage underlying resource management system AppLe S API Resources App- specific case gen. Exp Act Sched. Act Exp

INS2D AppLeS Project Goals Complete design and deployment of INS2D AppLeS for interactive cluster –focus on socket design for first phase Conduct experiments to assess AppLeS performance on interactive cluster and to compare with batch system performance Expand INS2D AppLeS to target both batch and interactive systems –target to evolving IPG resource management system

Show Stoppers Queue prediction time –How long will the program wait in a batch queue? –How accurate is the prediction? Experimental Verification –How do we verify the performance of schedulers in production environments? –How do we achieve reproducible and relevant results? –What are the right measures of success? Uncertainty –How do we capture time-dependent information? –What do we do if the range of information is large?

AppLeS and the IPG Usability, Integration development of basic IPG infrastructure Performance “grid-aware” programming Short-termMedium-termLong-term Application scheduling Resource scheduling Throughput scheduling Multi-scheduling Resource economy Integration of schedulers and other tools, performance interfaces You are here Integration of multiple grid constituencies architectural models which support multiple constituencies automation of program execution

Getting There: Current Projects AppLeS and more AppLeS –AppLeS applications –AppLeS templates/tools –Globus AppLeS, Legion AppLeS, IPG AppLeS –Plans for integration of AppLeS and NWS with NetSolve, Condor, Ninf Performance Prediction Engineering –structural modeling with stochastic predictions –development of quality of information measures accuracy lifetime overhead

New Directions Contingency Scheduling scheduling during execution Scheduling with partial information, poor information, dynamically changing information Multischeduling resource economies scheduling “social structure” X

The Brave New World Grid-aware Programming –development of adaptive poly-applications –integration of schedulers, PSEs and other tools PSEPSE Config. object program whole program compiler Source appli- cation libraries Realtime perf monitor Dynamic optimizer Grid runtime system negotiation Software components Service negotiator Scheduler Performance feedback Perf problem

Project Information Thanks to NSF, NPACI, Darpa, DoD, NASA AppLeS Corps: –Francine Berman –Rich Wolski –Walfredo Cirne –Marcio Faerman –Jaime Frey –Jim Hayes –Graziano Obertelli AppLeS Home Page: cse.ucsd.edu/groups/ hpcl/apples.html –Jenny Schopf –Gary Shao –Neil Spring –Shava Smallen –Alan Su –Dmitrii Zagorodnov