Parallel Tomography Shava Smallen CSE Dept. U.C. San Diego.

Slides:



Advertisements
Similar presentations
Open Science Grid Living on the Edge: OSG Edge Services Framework Kate Keahey Abhishek Rana.
Advertisements

A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
High Performance Computing Course Notes Grid Computing.
Advanced Workgroup System. Printer Admin Utility Monitors printers over IP networks Views Sharp and non-Sharp SNMP Devices Provided Standard with Sharp.
Threads, SMP, and Microkernels Chapter 4. Process Resource ownership - process is allocated a virtual address space to hold the process image Scheduling/execution-
USING THE GLOBUS TOOLKIT This summary by: Asad Samar / CALTECH/CMS Ben Segal / CERN-IT FULL INFO AT:
The Network Weather Service A Distributed Resource Performance Forecasting Service for Metacomputing Rich Wolski, Neil T. Spring and Jim Hayes Presented.
GridFlow: Workflow Management for Grid Computing Kavita Shinde.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Achieving Application Performance on the Information Power Grid Francine Berman U. C. San Diego and NPACI This presentation will probably involve audience.
Performance Prediction Engineering Francine Berman U. C. San Diego Rich Wolski U. C. San Diego and University of Tennessee This presentation will probably.
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
Tao Yang, UCSB CS 240B’03 Unix Scheduling Multilevel feedback queues –128 priority queues (value: 0-127) –Round Robin per priority queue Every scheduling.
AppLeS, NWS and the IPG Fran Berman UCSD and NPACI Rich Wolski UCSD, U. Tenn. and NPACI This presentation will probably involve audience discussion, which.
MCell Usage Scenario Project #7 CSE 260 UCSD Nadya Williams
Achieving Application Performance on the Computational Grid Francine Berman This presentation will probably involve audience discussion, which will create.
The AppLeS Project: Harvesting the Grid Francine Berman U. C. San Diego This presentation will probably involve audience discussion, which will create.
CSE 160/Berman Mapping and Scheduling W+A: Chapter 4.
NPACI Alpha Project Review: Cellular Microphysiology on the Data Grid Fran Berman, UCSD Tom Bartol, Salk Institute.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
AppLeS / Network Weather Service IPG Pilot Project FY’98 Francine Berman U. C. San Diego and NPACI Rich Wolski U.C. San Diego, NPACI and U. of Tennessee.
1 GRID D. Royo, O. Ardaiz, L. Díaz de Cerio, R. Meseguer, A. Gallardo, K. Sanjeevan Computer Architecture Department Universitat Politècnica de Catalunya.
Grids and Globus at BNL Presented by John Scott Leita.
1 Distributed Systems: Distributed Process Management – Process Migration.
Grid Computing Net 535.
Adaptive Server Farms for the Data Center Contact: Ron Sheen Fujitsu Siemens Computers, Inc Sever Blade Summit, Getting the.
Chapter 2 Computer Clusters Lecture 2.1 Overview.
Resource Management Reading: “A Resource Management Architecture for Metacomputing Systems”
CSE 160/Berman Programming Paradigms and Algorithms W+A 3.1, 3.2, p. 178, 5.1, 5.3.3, Chapter 6, 9.2.8, , Kumar Berman, F., Wolski, R.,
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Technology Education Copyright © 2006 by The McGraw-Hill Companies,
Achieving Application Performance on the Grid: Experience with AppLeS Francine Berman U. C., San Diego This presentation will probably involve audience.
High Performance Louisiana State University - LONI HPC Enablement Workshop – LaTech University,
Dynamic Firewalls and Service Deployment Models for Grid Environments Gian Luca Volpato, Christian Grimm RRZN – Leibniz Universität Hannover Cracow Grid.
Marcos Dias de Assunção 1,2, Alexandre di Costanzo 1 and Rajkumar Buyya 1 1 Department of Computer Science and Software Engineering 2 National ICT Australia.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
Young Suk Moon Chair: Dr. Hans-Peter Bischof Reader: Dr. Gregor von Laszewski Observer: Dr. Minseok Kwon 1.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
The Globus Project: A Status Report Ian Foster Carl Kesselman
Threads, SMP, and Microkernels Chapter 4. Process Resource ownership - process is allocated a virtual address space to hold the process image Scheduling/execution-
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
Net-Centric Software and Systems I/UCRC A Framework for QoS and Power Management for Mobile Devices in Service Clouds Project Lead: I-Ling Yen, Farokh.
TeraGrid Advanced Scheduling Tools Warren Smith Texas Advanced Computing Center wsmith at tacc.utexas.edu.
Part Five: Globus Job Management A: GRAM B: Globus Job Commands C: Laboratory: globusrun.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Using Map-reduce to Support MPMD Peng
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
Memory Hierarchy Adaptivity An Architectural Perspective Alex Veidenbaum AMRM Project sponsored by DARPA/ITO.
Faucets Queuing System Presented by, Sameer Kumar.
SAN DIEGO SUPERCOMPUTER CENTER Inca Control Infrastructure Shava Smallen Inca Workshop September 4, 2008.
Portal Update Plan Ashok Adiga (512)
1 On-line Parallel Tomography Shava Smallen UCSD.
Adaptive Computing on the Grid Using AppLeS Francine Berman, Richard Wolski, Henri Casanova, Walfredo Cirne, Holly Dail, Marcio Faerman, Silvia Figueira,
NASRULLAH KHAN.  Lecturer : Nasrullah   Website :
Static Process Scheduling
Globus Grid Tutorial Part 2: Running Programs Across Multiple Resources.
Application-level Scheduling Sathish S. Vadhiyar Credits / Sources: AppLeS web pages and papers.
Parallel Tomography Shava Smallen SC99. Shava Smallen SC99AppLeS/NWS-UCSD/UTK What are the Computational Challenges? l Quick turnaround time u Resource.
RobuSTore: Performance Isolation for Distributed Storage and Parallel Disk Arrays Justin Burke, Huaxia Xia, and Andrew A. Chien Department of Computer.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
The Globus Toolkit The Globus project was started by Ian Foster and Carl Kesselman from Argonne National Labs and USC respectively. The Globus toolkit.
CS 283Computer Networks Spring 2013 Instructor: Yuan Xue.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
Achieving Application Performance on the Computational Grid Francine Berman U. C. San Diego and NPACI This presentation will probably involve audience.
OpenMosix, Open SSI, and LinuxPMI
A. Rama Bharathi Regd. No: 08931F0040 III M.C.A
Department of Computer Science University of California, Santa Barbara
A Characterization of Approaches to Parrallel Job Scheduling
Presentation transcript:

Parallel Tomography Shava Smallen CSE Dept. U.C. San Diego

Parallel Tomography n Tomography: reconstruction of 3D image from 2D projections n Used for electron microscopy at National Center for Microscopy and Imaging Research (NCMIR) n Coarse-grain, embarrassingly parallel application n Goal of project: Achieve performance using AppLeS scheduling and Globus services

Parallel Tomography Structure driver reader ptomo writer preprocessor destsourc e Off-line Solid lines = data flow dashed lines = control

Parallel Tomography Structure driver reader ptomo writer preprocessor destsourc e On-line Solid lines = data flow dashed lines = control

Parallel Tomography Structure n driver: directs communication among processes, controls work queue n preprocessor: serially formats raw data from microscope for parallel processing n reader: reads preprocessed data and passes it to a ptomo process n ptomo: performs tomography processing n writer: writes processed data to destination (i.e. visualization device, tape)

NCMIR Environment n Platform: collection of heterogeneous resources u workstations (e.g. sgi, sun) u NPACI supercomputer time F (e.g. SP-2, T3E) n How can users achieve execution performance in heterogeneous, multi-user environments?

Performance and Scheduling n Users want to achieve minimum turnaround time in the following scenarios: u off-line: already collected data set u on-line: data streamed from electron microscope n Goal of our project is to develop adaptive scheduling strategies which promote both performance and flexibility for tomography in multi- user Globus environments

Adaptive Scheduling strategy n Develop schedule which adapts to deliverable resource performance at execution time n Application scheduler will dynamically u select sets of resources based on user-defined performance measure u plan possible schedules for each set of feasible resources u predict the performance for each schedule u implement best predicted schedule on selected infrastructure

AppLeS = Application-Level Scheduling n Each AppLeS is an adaptive application scheduler u AppLeS+application = self-scheduling application n scheduling decisions based on u dynamic information (i.e. resource load forecasts) u static application and system information

Resource Selection n available resources u workstations : run immediately  execution may be slow due to load u supercomputers : may have to wait in a queue  execution fast on dedicated nodes n We want to schedule using both types of resources together for an improved execution performance

Allocation Strategy n We have developed a strategy which simultaneously schedules on both u workstations u immediately available supercomputer nodes F avoid wait time in the batch queue F information is exported by batch scheduler n Overall, this strategy performs better than running on either type of resource alone

Preliminary Globus/AppLeS Tomography Experiments n Resources u 6 workstations available at Parallel Computation Laboratory (PCL) at UCSD u immediately available nodes on SDSC SP-2 (128 nodes) F Maui scheduler exports the number of immediately available nodes e.g. 5 nodes available for the next 30 mins 10 nodes available for the next 10 mins u Globus installed everywhere W W W W W W NCMIR SDSC SP-2 W W W W W W PCL

Allocation Strategies compared n 4 strategies compared: u SP2Immed/WS: workstations and immediately available SP-2 nodes u WS: workstations only u SP2Immed: immediately available SP-2 nodes only u SP2Queue(n): traditional batch queue submit using n nodes n experiments performed in production environment u ran experiments in sets, each set contains all strategies F e.g. SP2Immed, SP2Immed/WS, WS, SP2Queue(8) u within a set, experiments ran back-to-back

Results (8 nodes on SP-2)

Results (16 nodes on SP-2)

Results (32 nodes on SP-2)

Targeting Globus n AppLeS uses Globus services u GRAM, GSI, & RSL F process startup on workstations and batch queue systems F remote process control u Nexus F interprocess communication F multi-threaded F callback functions for fault tolerance

Experiences with Globus n What we would like to see improved: u free node information in MDS (e.g. time availability, frequency) u steep learning curve to initial installation F knowledge of underlying Globus infrastructure u startup scripts F more documentation F more flexible configuration

Experiences with Globus n What worked: u once installed, software works well u responsiveness and willingness to help from Globus team F mailing list F web pages

Future work n Develop contention model to address network overloading which includes u network capacity information u model of application communication requirements n Expansion of allocation policy to include additional resources u other Globus supercomputers, reserved resources (GARA) u availability of resources n NPACI Alpha project with NCMIR and Globus

AppLeS n AppLeS/NCMIR/Globus Tomography Project u Shava Smallen, Jim Hayes, Fran Berman, Rich Wolski, Walfredo Cirne (AppLeS) u Mark Ellisman, Marty Hadida-Hassan, Jaime Frey (NCMIR), u Carl Kesselman, Mei-Hui Su (Globus) n AppLeS Home Page u www-cse.ucsd.edu/groups/hpcl/apples n