Sponsors and Acknowledgments This work is supported in part by the National Science Foundation under Grants No. OCI-0910812, IIP-0758596 and CNS-0821622.

Slides:



Advertisements
Similar presentations
First Steps in the Clouds
Advertisements

11 Application of CSF4 in Avian Flu Grid: Meta-scheduler CSF4. Lab of Grid Computing and Network Security Jilin University, Changchun, China Hongliang.
A Scalable Approach to Deploying and Managing Appliances Kate Keahey Rick Bradshaw, Narayan Desai, Tim Freeman Argonne National Lab, University of Chicago.
Virtualization, Cloud Computing, and TeraGrid Kate Keahey (University of Chicago, ANL) Marlon Pierce (Indiana University)
Wei Lu 1, Kate Keahey 2, Tim Freeman 2, Frank Siebenlist 2 1 Indiana University, 2 Argonne National Lab
Virtual Workspaces in the Grid Kate Keahey Argonne National Laboratory Ian Foster, Tim Freeman, Xuehai Zhang, Daniel Galron.
Science Clouds: Early Experiences in Cloud Computing for Scientific Applications Chicago, October 2008 Kate Keahey, Renato Figueiredo, Jose Fortes, Tim.
Nimbus or an Open Source Cloud Platform or the Best Open Source EC2 No Money Can Buy ;-) Kate Keahey Tim Freeman University of Chicago.
Adding the Easy Button to the Cloud with SnowFlock and MPI Philip Patchin, H. Andrés Lagar-Cavilla, Eyal de Lara, Michael Brudno University of Toronto.
Windows® Deployment Services
Education and training on FutureGrig Salt Lake City, Utah July 18 th 2011 Presented by Renato Figueiredo
FutureGrid related presentations at TG and OGF Sun. 17th: Introduction to FutireGrid (OGF) Mon. 18th: Introducing to FutureGrid (TG) Tue. 19th –Educational.
Advanced Computing and Information Systems laboratory Virtual Private Clusters: Virtual Appliances and Networks in the Cloud Renato Figueiredo ACIS Lab.
Future Grid Early Projects 2010 User Advisory Board Meeting Pittsburgh, PA.
Clouds C. Vuerli Contributed by Zsolt Nemeth. As it started.
Clouds from FutureGrid’s Perspective April Geoffrey Fox Director, Digital Science Center, Pervasive.
Advanced Computing and Information Systems laboratory Educational Virtual Clusters for On- demand MPI/Hadoop/Condor in FutureGrid Renato Figueiredo Panoat.
High memory instances Monthly SLA : Virtual Machines Validated & supported Microsoft workloads Price reduction: standard Windows (22%) & Linux (29%)
COMMA: Coordinating the Migration of Multi-tier applications 1 Jie Zheng* T.S Eugene Ng* Kunwadee Sripanidkulchai† Zhaolei Liu* *Rice University, USA †NECTEC,
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Virtualization for Cloud Computing
Virtualization Infrastructure Administration Cluster Jakub Yaghob.
Cloud Computing Why is it called the cloud?.
Yury Kissin Infrastructure Consultant Storage improvements Dynamic Memory Hyper-V Replica VM Mobility New and Improved Networking Capabilities.
Virtual Infrastructure in the Grid Kate Keahey Argonne National Laboratory.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Simulation of Cloud Environments
Purdue RP Highlights TeraGrid Round Table September 23, 2010 Carol Song Purdue TeraGrid RP PI Rosen Center for Advanced Computing Purdue University.
Virtual Machine Hosting for Networked Clusters: Building the Foundations for “Autonomic” Orchestration Based on paper by Laura Grit, David Irwin, Aydan.
SAIGONTECH COPPERATIVE EDUCATION NETWORKING Spring 2009 Seminar #1 VIRTUALIZATION EVERYWHERE.
Cloud Computing. What is Cloud Computing? Cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable.
A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.
Experimenting with FutureGrid CloudCom 2010 Conference Indianapolis December Geoffrey Fox
Advanced Computing and Information Systems laboratory The case for UF in PRAGMA Jose Fortes (also on behalf of Renato Figueiredo and Reed Beaman)
DISTRIBUTED COMPUTING
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
Improving Network I/O Virtualization for Cloud Computing.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Grids, Clouds and the Community. Cloud Technology and the NGS Steve Thorn Edinburgh University Matteo Turilli, Oxford University Presented by David Fergusson.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
SALSA HPC Group School of Informatics and Computing Indiana University.
FutureGrid Cyberinfrastructure for Computational Research.
Sponsored by the National Science Foundation Systematic Experimentation, Automation, and Scaling Up Sarah Edwards, GENI Project Office.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Virtual Workspaces Kate Keahey Argonne National Laboratory.
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid PI: Geoffrey Fox*, CoPIs: Kate Keahey +, Warren Smith -, Jose Fortes #, Andrew.
Windows Azure. Azure Application platform for the public cloud. Windows Azure is an operating system You can: – build a web application that runs.
Efficient Live Checkpointing Mechanisms for computation and memory-intensive VMs in a data center Kasidit Chanchio Vasabilab Dept of Computer Science,
Computing Research Testbeds as a Service: Supporting large scale Experiments and Testing SC12 Birds of a Feather November.
Vignesh Ravindran Sankarbala Manoharan. Infrastructure As A Service (IAAS) is a model that is used to deliver a platform virtualization environment with.
Purdue RP Highlights TeraGrid Round Table May 20, 2010 Preston Smith Manager - HPC Grid Systems Rosen Center for Advanced Computing Purdue University.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
Instituto de Biocomputación y Física de Sistemas Complejos Cloud resources and BIFI activities in JRA2 Reunión JRU Española.
Nimbus Update March 2010 OSG All Hands Meeting Kate Keahey Nimbus Project University of Chicago Argonne National Laboratory.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Daniele Lezzi Execution of scientific workflows on federated multi-cloud infrastructures IBERGrid Madrid, 20 September 2013.
Virtualization for Cloud Computing
Private Public FG Network NID: Network Impairment Device
StratusLab Final Periodic Review
StratusLab Final Periodic Review
Cloud Computing with Nimbus
Dr. John P. Abraham Professor, Computer Engineering UTPA
Sky Computing on FutureGrid and Grid’5000
Virtualization, Cloud Computing, and TeraGrid
Clouds from FutureGrid’s Perspective
Using and Building Infrastructure Clouds for Science
This material is based upon work supported by the National Science Foundation under Grant #XXXXXX. Any opinions, findings, and conclusions or recommendations.
Sky Computing on FutureGrid and Grid’5000
Presentation transcript:

Sponsors and Acknowledgments This work is supported in part by the National Science Foundation under Grants No. OCI , IIP and CNS and in part by the MCS Division subprogram of the Office of Advanced Scientific Computing Research, SciDAC Program, Office of Science, U.S. Department of Energy, under Contract DE-AC02-06CH The authors also acknowledge the support of the BellSouth Foundation. Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation or BellSouth Foundation. Experiments were carried out using the Grid'5000 experimental testbed, being developed under the INRIA ALADDIN development action with support from CNRS, RENATER and several Universities as well as other funding bodies (see Sky computing is an emerging computing model where resources from multiple cloud providers are leveraged to create large scale distributed infrastructures. Experimental Testbeds IntroductionIntroduction ConclusionConclusion ArchitectureArchitecture VM Image Propagation Mechanisms To deploy virtual clusters, each VM requires an independent replica of a common VM image. Nimbus transfers a copy of the required VM image to each VM host (a step called propagation), using SCP from a single repository. This propagation scheme doesnt scale with the number of VMs as it is limited by the repository disk or network bandwidth. To overcome this problem, we developed two new propagation mechanisms. The first one leverages the TakTuk and Kastafior tools developed at INRIA to create a broadcast chain used to transfer image data. The second one relies on Copy-on-Write capabilities of the Xen hypervisor. Sky Computing on FutureGrid and Grid5000 Pierre Riteau 1, Mauricio Tsugawa 2, Andrea Matsunaga 2, José Fortes 2, Tim Freeman 3, David LaBissoniere 4, Kate Keahey 3,4 1 Université de Rennes 1, IRISA/INRIA Rennes – Bretagne Atlantique 2 University of Florida 3 Argonne National Labs 4 University of Chicago Computation Institute The above graph compares instantiation times of virtual clusters using different propagation mechanisms. In the SCP and TakTuk cases, the image is compressed and is 2.2GB in size (12 GB uncompressed). In the QCOW case, the 12GB image is pre-propagated on all hypervisors. Propagation consists in creating a new Copy-On-Write volume and contextualizing the virtual cluster. FutureGrid is an experimental testbed for grid and cloud research. It is distributed over 6 sites in the US and offers more than 5,000 cores. Grid5000 is an experimental testbed for research in large-scale parallel and distributed systems. It is distributed over 9 sites in France and offers more than 5,500 cores. This work uses resources across two experimental projects: FutureGrid and Grid5000. This showcases not only the capabilities of the experimental platforms, but also their emerging collaboration. The two platforms are used to create a Sky Computing environment. To validate our approach in a real-world scenario, we run a MapReduce version of a popular bioinformatics application (BLAST). However, any kind of distributed application can be run on these infrastructures. The Sky Computing model allows the creation of large scale infrastructures using resources from multiple cloud providers. These infrastructures are able to run embarrassingly parallel computation with high performance. Our work shows how it is possible to federate multiple infrastructures and improve the speed of virtual cluster creation, using experimental testbeds in the US and in France as an example. Our Sky Computing deployment makes use of: Xen to minimize platform (hardware and operating system stack) differences Nimbus to offer VM provisioning and contextualization services (contextualization automatically assigns roles and configures VMs) ViNe, a virtual network based on an IP-overlay, to enable all-to-all communication between virtual machines spread across multiple clouds Hadoop for parallel fault-tolerant execution and dynamic cluster extension ScalabilityScalability We deployed a Sky Computing infrastructure consisting of 1114 CPU cores (457 VMs) distributed over 3 sites in FutureGrid and 3 sites in Grid5000 (OGF- 29 demo, Chicago, IL, June 2010). FutureGrid Grid5000 San Diego University of Florida University of Chicago Lille Rennes Sophia Queue ViNe Router Grid5000 firewall Cloud A Cloud B Nimbus ViNe Hadoop MapReduce App (e.g. BLAST) Cloud C Nimbus Distributed Application (e.g. MPI BLAST) Distributed Application (e.g. MPI BLAST) ViNe router VMs FutureGrid Grid5000