Trace Generation to Simulate Large Scale Distributed Application Olivier Dalle, Emiio P. ManciniMar. 8th, 2012.

Slides:



Advertisements
Similar presentations
Network II.5 simulator ..
Advertisements

Barcelona Supercomputing Center. The BSC-CNS objectives: R&D in Computer Sciences, Life Sciences and Earth Sciences. Supercomputing support to external.
INTRODUCTION TO SIMULATION WITH OMNET++ José Daniel García Sánchez ARCOS Group – University Carlos III of Madrid.
Scalable Multi-Cache Simulation Using GPUs Michael Moeng Sangyeun Cho Rami Melhem University of Pittsburgh.
SALSA HPC Group School of Informatics and Computing Indiana University.
1 Advancing Supercomputer Performance Through Interconnection Topology Synthesis Yi Zhu, Michael Taylor, Scott B. Baden and Chung-Kuan Cheng Department.
SKELETON BASED PERFORMANCE PREDICTION ON SHARED NETWORKS Sukhdeep Sodhi Microsoft Corp Jaspal Subhlok University of Houston.
WS-VLAM: Towards a Scalable Workflow System on the Grid V. Korkhov, D. Vasyunin, A. Wibisono, V. Guevara-Masis, A. Belloum Institute.
Task Scheduling and Distribution System Saeed Mahameed, Hani Ayoub Electrical Engineering Department, Technion – Israel Institute of Technology
Scalable Application Layer Multicast Suman Banerjee Bobby Bhattacharjee Christopher Kommareddy ACM SIGCOMM Computer Communication Review, Proceedings of.
Exploiting Content Localities for Efficient Search in P2P Systems Lei Guo 1 Song Jiang 2 Li Xiao 3 and Xiaodong Zhang 1 1 College of William and Mary,
1 Introduction to Load Balancing: l Definition of Distributed systems. Collection of independent loosely coupled computing resources. l Load Balancing.
WSN Simulation Template for OMNeT++
Inferring the Topology and Traffic Load of Parallel Programs in a VM environment Ashish Gupta Peter Dinda Department of Computer Science Northwestern University.
Router Architectures An overview of router architectures.
Massive Ray Tracing in Fusion Plasmas on EGEE J.L. Vázquez-Poletti, E. Huedo, R.S. Montero and I.M. Llorente Distributed Systems Architecture Group Universidad.
Router Architectures An overview of router architectures.
© 2005, it - instituto de telecomunicações. Todos os direitos reservados. System Level Resource Discovery and Management for Multi Core Environment Javad.
Edge Based Cloud Computing as a Feasible Network Paradigm(1/27) Edge-Based Cloud Computing as a Feasible Network Paradigm Joe Elizondo and Sam Palmer.
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED SYSTEMS.
UAB Dynamic Monitoring and Tuning in Multicluster Environment Genaro Costa, Anna Morajko, Paola Caymes Scutari, Tomàs Margalef and Emilio Luque Universitat.
DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Computer and Automation Research Institute Hungarian Academy of Sciences Presentation and Analysis of Grid Performance Data Norbert Podhorszki and Peter.
MediaGrid Processing Framework 2009 February 19 Jason Danielson.
A Proposal of Application Failure Detection and Recovery in the Grid Marian Bubak 1,2, Tomasz Szepieniec 2, Marcin Radecki 2 1 Institute of Computer Science,
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
SALSA HPC Group School of Informatics and Computing Indiana University.
IPDPS 2005, slide 1 Automatic Construction and Evaluation of “Performance Skeletons” ( Predicting Performance in an Unpredictable World ) Sukhdeep Sodhi.
OMIS Approach to Grid Application Monitoring Bartosz Baliś Marian Bubak Włodzimierz Funika Roland Wismueller.
Parallel Programming on the SGI Origin2000 With thanks to Igor Zacharov / Benoit Marchand, SGI Taub Computer Center Technion Moshe Goldberg,
July 31 - August 4, 1999 SCI'99 / ISAS'99 Performance Prediction for Data Intensive Applications on Large Scale Parallel Systems Yuhong Wen and Geoffrey.
Nguyen Tuan Anh. VN-Grid: Goals  Grid middleware (focus of this presentation)  Tuan Anh  Grid applications  Hoai.
How to for compiling and running MPI Programs. Prepared by Kiriti Venkat.
HIERARCHICAL TREES OF UNSTEADY SIMULATION DATASETS Marek Gayer and Pavel Slavík C omputer G raphics G roup Department of Computer Science and Engineering.
6/29/1999PDPTA'991 Performance Prediction for Large Scale Parallel Systems Yuhong Wen and Geoffrey C. Fox Northeast Parallel Architecture Center (NPAC)
Scalable and Coordinated Scheduling for Cloud-Scale computing
Pre-calculated Fluid Simulator States Tree Marek Gayer and Pavel Slavík C omputer G raphics G roup Department of Computer Science and Engineering Faculty.
Scalable and Topology-Aware Load Balancers in Charm++ Amit Sharma Parallel Programming Lab, UIUC.
Distributed Topology Control In Wireless Sensor Networks with Assymetric Links Presenter: Himali Saxena.
On the Placement of Web Server Replicas Yu Cai. Paper On the Placement of Web Server Replicas Lili Qiu, Venkata N. Padmanabhan, Geoffrey M. Voelker Infocom.
Hierarchical Load Balancing for Large Scale Supercomputers Gengbin Zheng Charm++ Workshop 2010 Parallel Programming Lab, UIUC 1Charm++ Workshop 2010.
LACSI 2002, slide 1 Performance Prediction for Simple CPU and Network Sharing Shreenivasa Venkataramaiah Jaspal Subhlok University of Houston LACSI Symposium.
Background Computer System Architectures Computer System Software.
DataTAG is a project funded by the European Union International School on Grid Computing, 23 Jul 2003 – n o 1 GridICE The eyes of the grid PART I. Introduction.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED SYSTEMS.
1
1 Scalability and Accuracy in a Large-Scale Network Emulator Nov. 12, 2003 Byung-Gon Chun.
Towards a High Performance Extensible Grid Architecture Klaus Krauter Muthucumaru Maheswaran {krauter,
Interaction and Animation on Geolocalization Based Network Topology by Engin Arslan.
Introduction to Load Balancing:
California Institute of Technology
BDII Performance Tests
Grid Computing.
CRESCO Project: Salvatore Raia
Introduction to Parallelism.
Auburn University COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques (2) Dr. Xiao Qin Auburn University.
Chapter 17: Database System Architectures
ExaO: Software Defined Data Distribution for Exascale Sciences
The Globus Toolkit™: Information Services
Shreeni Venkataramaiah
Introduction to locality sensitive approach to distributed systems
Distributed Databases
CLUSTER COMPUTING.
Characteristics of Reconfigurable Hardware
CARLA Buenos Aires, Argentina - Sept , 2017
Outline Ganesan, D., Greenstein, B., Estrin, D., Heidemann, J., and Govindan, R. Multiresolution storage and search in sensor networks. Trans. Storage.
Gengbin Zheng, Esteban Meneses, Abhinav Bhatele and Laxmikant V. Kale
Presentation transcript:

Trace Generation to Simulate Large Scale Distributed Application Olivier Dalle, Emiio P. ManciniMar. 8th, 2012

O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 2 Outline Introduction The trace collection The hierarchical architecture The components An example Conclusion

Mar. 8th, 2012 O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 3 Introduction Most distributed systems, as the Grids, offer massively parallel but loosely coupled resources: an accurate application’s model can help the scheduling decisions Simulators of parallel and distributed applications need accurate model of application behavior: but the size of the traces for long running parallel applications tends to explode

Mar. 8th, 2012 O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 4 Introduction One solution is to buffer data locally, gathering them after the end of the program (post-mortem): there is some scalability issue We need to minimize the perturbation: the instrumentation compete with the application for the system’s resources.

Introduction A distributed application is composed by a set of cooperating tasks The connection between them are in general not homogenous Networks may present some hierarchy, e.g. fat trees, multi switch hops... Can we exploit that hierarchy on the trace generation/instrumentation purposes? O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 5

The Trace Collection: a Simplified Schema O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 6 The classical computational cluster execution model: Several task on several nodes (e.g., MPI) CPU GPU Core

The Trace Collection: a Simplified Schema O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 7 We need to measure some parameters on each task, collect local data, and gather them.

The Trace Collection: a Simplified Schema O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 8 We gather the data hierarchically, using local collectors, eventually making local decimations or pre- elaborations. We use the locality principle. In a Grid it is common to have a low quality connecting link between the V.O. sites In HPC the bandwidth of upper levels is shared between more hosts than lower levels

The Trace Collection O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications Infrastructure initialization 2. Execution with instrumentation a.Environment update (e.g., LD_PRELOAD) b.Middleware launcher (e.g., mpiexec, qsub …) 3. Data collection a.Overhead estimation b.Events’ measurement 4. Processing and Propagation a.Decimation b.Compression c.Buffering d.… 5. Trace generation a.Data collection b.Post-processing c.Simulator’s trace generation

The architecture O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 10

The sensors O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 11 The sensors: Instrument the application’s tasks Compute the instrumentation’s overhead Collect the raw data Send them to the first level collectors

The sensors O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 12 We assume the system to be heterogeneous Every sensor makes an overhead analysis Then it propagates the information to the management unit

The collectors Dec. 14th, 2011 O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 13 The collectors gather data from sensors and from other collectors Buffer incoming data Process collected data before sending them to upper levels Decimation Compression …

The Management Unit O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 14 Launches the collector daemons Launches the application Gather the data from the top collector Convert and store the data in the required format Managed with scripts or graphical interface

An Example of Data Collection O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 15 We are interested to analyze the I/O of a parallel synthetic benchmark We want to check the overhead The benchmark is a MPI application of n tasks Every task runs on a different node and writes random data on the local file system

An Example of Data Collection: O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 16 We use the management unit to: 1.Create a hierarchical schema 2.Create the MPI launch scripts 3.Launch the collectors and the instrumented application 1.Collect the results $MPIDIR/mpiexec.hydra \ -env LD_PRELOAD \ $DTDIR/libdt_sensor.so \ $HOME/bench/bench mpiexec qsub …

Conclusion O.Dalle, E.P. Mancini - Trace generation for large scale distributed applications- 17 Collecting large traces in distributed systems may perturb the application’s execution. We presented a system that efficiently collects traces at run-time or post-mortem. We use a hierarchical schema matching the network links’ capacity, with distributed buffering and processing Future improvement will include the automatic discovery of the network topology

Thank you