RAIN: A system to Dynamically Generate & Provision Images on Bare Metal by Application Users Presented by Gregor von Laszewski Authors: Javier Diaz, Gregor.

Slides:



Advertisements
Similar presentations
Overview of the FutureGrid Software
Advertisements

FutureGrid related presentations at TG and OGF Sun. 17th: Introduction to FutireGrid (OGF) Mon. 18th: Introducing to FutureGrid (TG) Tue. 19th –Educational.
FutureGrid UAB Meeting XSEDE13 San Diego July
2010 FutureGrid User Advisory Meeting Architecture Roadmap Long term vision 10:00-10:45, Monday, August 2, 2010 Pittsburgh, PA Gregor von Laszewski Representing.
FutureGrid Overview NSF PI Science of Cloud Workshop Washington DC March Geoffrey Fox
Future Grid Introduction March MAGIC Meeting Gregor von Laszewski Community Grids Laboratory, Digital Science.
Advanced Computing and Information Systems laboratory Educational Virtual Clusters for On- demand MPI/Hadoop/Condor in FutureGrid Renato Figueiredo Panoat.
Implementing a menu based application in FutureGrid
FutureGrid Image Repository: A Generic Catalog and Storage System for Heterogeneous Virtual Machine Images Javier Diaz, Gregor von Laszewski, Fugang Wang,
Jefferson Ridgeway 2, Ifeanyi Rowland Onyenweaku 3, Gregor von Laszewski 1*, Fugang Wang 1 1* Indiana University, Bloomington, IN 47408, U.S.A.,
Reproducible Environment for Scientific Applications (Lab session) Tak-Lon (Stephen) Wu.
INTRODUCTION TO CLOUD COMPUTING Cs 595 Lecture 5 2/11/2015.
Cloud Computing Systems Lin Gu Hong Kong University of Science and Technology Sept. 21, 2011 Windows Azure—Overview.
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid Geoffrey Fox, Andrew J. Younge, Gregor von Laszewski, Archit Kulshrestha, Fugang.
SC2010 Gregor von Laszewski (*) (*) Assistant Director of Cloud Computing, CGL, Pervasive Technology Institute.
Big Data and Clouds: Challenges and Opportunities NIST January Geoffrey Fox
Eucalyptus on FutureGrid: A case for Eucalyptus 3 Sharif Islam, Javier Diaz, Geoffrey Fox Gregor von Laszewski Indiana University.
Cyberaide Virtual Appliance: On-demand Deploying Middleware for Cyberinfrastructure Tobias Kurze, Lizhe Wang, Gregor von Laszewski, Jie Tao, Marcel Kunze,
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Raining Compute Environments on Resources by Application Users Gregor von Laszewski Indiana University Open Cirrus Summit 2011, Oct.
FutureGrid SOIC Lightning Talk February Geoffrey Fox
Science of Cloud Computing Panel Cloud2011 Washington DC July Geoffrey Fox
Cloud Computing. What is Cloud Computing? Cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable.
+ CS 325: CS Hardware and Software Organization and Architecture Cloud Architectures.
Experimenting with FutureGrid CloudCom 2010 Conference Indianapolis December Geoffrey Fox
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
Gregor von Laszewski*, Geoffrey C. Fox, Fugang Wang, Andrew Younge, Archit Kulshrestha, Greg Pike (IU), Warren Smith, (TACC) Jens Vöckler (ISI), Renato.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Future Grid FutureGrid Overview Dr. Speaker. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research on the future.
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid Dynamic Provisioning Experiments including Hadoop Fugang Wang, Archit Kulshrestha, Gregory G. Pike, Gregor von Laszewski, Geoffrey C. Fox.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
What’s Hot in Clouds? Analyze (superficially) the ~140 Papers/Short papers/Workshops/Posters/Demos in CloudCom Each paper may fall in more than one category.
Future Grid FutureGrid Overview Geoffrey Fox SC09 November
Eucalyptus 3 (&3.1). Eucalyptus 3 Product Overview – Govind Rangasamy.
Image Management and Rain on FutureGrid: A practical Example Presented by Javier Diaz, Fugang Wang, Gregor von Laszewski.
FutureGrid SC10 New Orleans LA IU Booth November Geoffrey Fox
FutureGrid Connection to Comet Testbed and On Ramp as a Service Geoffrey Fox Indiana University Infra structure.
Image Generation and Management on FutureGrid CTS Conference 2011 Philadelphia May Geoffrey Fox
Image Management and Rain on FutureGrid Javier Diaz - Fugang Wang – Gregor von.
Magellan: Experiences from a Science Cloud Lavanya Ramakrishnan.
FutureGrid SOIC Lightning Talk February Geoffrey Fox
FutureGrid Cyberinfrastructure for Computational Research.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
SAN DIEGO SUPERCOMPUTER CENTER Inca TeraGrid Status Kate Ericson November 2, 2006.
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid PI: Geoffrey Fox*, CoPIs: Kate Keahey +, Warren Smith -, Jose Fortes #, Andrew.
Computing Research Testbeds as a Service: Supporting large scale Experiments and Testing SC12 Birds of a Feather November.
Recipes for Success with Big Data using FutureGrid Cloudmesh SDSC Exhibit Booth New Orleans Convention Center November Geoffrey Fox, Gregor von.
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Future Grid Future Grid Overview. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research that will invent the future.
Vignesh Ravindran Sankarbala Manoharan. Infrastructure As A Service (IAAS) is a model that is used to deliver a platform virtualization environment with.
Directions in eScience Interoperability and Science Clouds June Interoperability in Action – Standards Implementation.
Grappling Cloud Infrastructure Services with a Generic Image Repository Javier Diaz Andrew J. Younge, Gregor von Laszewski, Fugang.
PLATFORM TO EASE THE DEPLOYMENT AND IMPROVE THE AVAILABILITY OF TRENCADIS INFRASTRUCTURE IberGrid 2013 Miguel Caballer GRyCAP – I3M - UPV.
WP5 – Infrastructure Operations Test and Production Infrastructures StratusLab kick-off meeting June 2010, Orsay, France GRNET.
Introductory Tutorial: OpenStack, Chef, Hadoop, Hbase, Pig I590 Data Science Curriculum Big Data Open Source Software and Projects September Geoffrey.
Private Public FG Network NID: Network Impairment Device
Digital Science Center II
StratusLab First Periodic Review
StratusLab Final Periodic Review
StratusLab Final Periodic Review
FutureGrid Computing Testbed as a Service
Versatile HPC: Comet Virtual Clusters for the Long Tail of Science SC17 Denver Colorado Comet Virtualization Team: Trevor Cooper, Dmitry Mishin, Christopher.
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Sky Computing on FutureGrid and Grid’5000
1Indiana University, 2now Rutgers University
Gregor von Laszewski Indiana University
Sky Computing on FutureGrid and Grid’5000
Presentation transcript:

RAIN: A system to Dynamically Generate & Provision Images on Bare Metal by Application Users Presented by Gregor von Laszewski Authors: Javier Diaz, Gregor von Laszewski, Fugang Wang, Geoffrey Fox, Andrew Younge

Acknowledgment: People Many people have worked on FuturGrid and we are not be able to list all them here. We will attempt to keep a list available on the portal Web site. Many others have contributed to this tutorial!! Thanks!!

Acknowledgement The FutureGrid project is funded by the National Science Foundation (NSF) and is led by Indiana University with University of Chicago, University of Florida, San Diego Supercomputing Center, Texas Advanced Computing Center, University of Virginia, University of Tennessee, University of Southern California, Dresden, Purdue University, and Grid 5000 as partner sites.

Reuse of slides If you reuse the slides please indicate that they are copied from this tutorial. Include a link to its biblographical entry that you can find at – We discourage the printing the slide material on paper due to two reasons: – We like to minimize the impact on the environment for paper and ink usage – We intend to keep the tutorials up to date on the Web site

FutureGrid Overview

Selected FutureGrid key Features International testbed for Grids, Clouds, HPC Can be used by Industry and Academia Education & Training A flexible development and testing platform for middleware and application – users looking at interoperability, functionality, performance or evaluation In future each use of FutureGrid is an experiment that is reproducible Distributed infrastructure – Indiana University, University of Chicago, University of Florida, San Diego Supercomputing Center, Texas Advanced Computing Center, University of Virginia, University of Tennessee, University of Southern California, Dresden, Purdue University, and Grid 5000 as partner sites. Indiana UniversityUniversity of ChicagoUniversity of FloridaSan Diego Supercomputing CenterTexas Advanced Computing Center University of VirginiaUniversity of TennesseeUniversity of Southern CaliforniaDresdenPurdue UniversityGrid 5000

Selected List of Services Offered PaaS Hadoop (Twister) (Sphere/Sector) IaaS Nimbus Eucalyptus ViNE (OpenStack) (OpenNebula) Grid Genesis II Unicore SAGA (Globus) HPCC MPI OpenMP ScaleMP (XD Stack) Others Portal Inca Ganglia (Exper. Manag./(Pegasus (Rain) (will be added in future)

Services Offered 1.ViNe can be installed on the other resources via Nimbus  2.Access to the resource is requested through the portal  3.Pegasus available via Nimbus and Eucalyptus images

User demand influences service deployment Based on User input we focused on – Nimbus (53%) – Eucalyptus (51%) – Hadoop (37%) – HPC (36%) Eucalyptus: 64(50.8%) High Performance Computing Environment: 45(35.7%) Nimbus: 67(53.2%) Hadoop: 47(37.3%) MapReduce: 42(33.3%) Twister: 20(15.9%) OpenNebula: 14(11.1%) Genesis II: 21(16.7%) Common TeraGrid Software Stack: 34(27%) Unicore 6: 13(10.3%) gLite: 12(9.5%) OpenStack: 16(12.7%) * Note: We will improve the way we gather statistics in order to avoid inaccuracy during the information gathering at project and user registration time.

Dynamic Service Allocation

Rain in FutureGrid

Software Architecture

FutureGrid Software Architecture Note on Authentication and Authorization We have different environments and requirements from XSEDE Non trivial to integrate/align security model with XSEDE

Detailed Software Architecture

Next we present selected Services      

Image Management and Dynamic Provisioning

Terminology Image Management provides the low level software (create, customize, store, share and deploy images) needed to achieve Dynamic Provisioning and Rain Dynamic Provisioning is in charge of providing machines with the requested OS. The requested OS must have been previously deployed in the infrastructure RAIN is our highest level component that uses Dynamic Provisioning and Image Management to provide custom environments that may or may not exits. Therefore, a Rain request may involve the creation, deployment and provision of one or more images in a set of machines

Motivation The goal is to create and maintain platforms in custom FG images that can be retrieved, deployed, and provisioned on demand Imagine the following scenario for FG:  fg-image-generate –o ubuntu –v maverick -s openmpi-bin,gcc,fftw2,emacs –n ubuntu-mpi-dev (store img in repo with id 1234)  fg-image-deploy –x india.futuregrid.org –r 1234  fg-rain –provision -n 32 ubuntu-mpi-dev

Architecture Image management is supported by a number of tightly-coupled services essential for FG The major services are –Image Repository –Image Generator –Image Deployment –RAIN – Dynamic provisioning –External Services

Image Management

Image Generation Creates and customizes images according to user requirements : o OS type o OS version o Architecture o Software Packages Image is stored in the Image Repository or returned to the users Images are not aimed to any specific infrastructure

Image Generation (Implementation View)

Image Verification (I) Images will be verified to guarantee some minimum security requirements Only if the image passes predefined tests, it is marked as deployable Verification takes place several times on an image –Time of generation –Before and after the deployment –Once a time threshold is reached –Periodically

Image Deployment Customizes (network IP, DNS, file system table, kernel modules, etc) and deploys images for specific infrastructures Two main infrastructures types –HPC deployment: it means that we are going to create network bootable images that can run in bare metal machines –Cloud deployment: it means that we are going to convert the images in VMs

Image Deployment (Implementation View)

Image Repository (I) Integrated service that enables storing and organizing images from multiple cloud efforts in the same repository Images are augmented with metadata to describe their properties like the software stack installed or the OS Access to the images can be restricted to single users, groups of users or system administrators

Image Repository (II) Maintains data related with the usage to assist performance monitoring and accounting Quota management to avoid space restrictions Pedigree to recreate image on demand Repository’s interfaces: API's, a command line, an interactive shell, and a REST service Other cloud frameworks could integrate with this image repository by accessing it through an standard API

Image Repository II

Rain – Dynamic Provisioning

Classical Dynamic Provisioning Dynamically partition a set of resources Dynamically allocate resources to users Dynamically define the environment that a resource is going to use Dynamically assign them based on user request Deallocate the resources so they can be dynamically allocated again

Use Cases of Dynamic Provisioning Static provisioning: o Resources in a cluster may be statically reassigned based on the anticipated user requirements, part of an HPC or cloud service. It is still dynamic, but control is with the administrator. (Note some call this also dynamic provisioning.) Automatic Dynamic provisioning: o Replace the administrator with intelligent scheduler. Queue-based dynamic provisioning: o provisioning of images is time consuming, group jobs using a similar environment and reuse the image. User just sees queue. Deployment: o dynamic provisioning features are provided by a combination of using XCAT and Moab

Generic Reprovisioning

Dynamic Provisioning Examples Give me a virtual cluster with 30 nodes based on Xen Give me 15 KVM nodes each in Chicago and Texas linked to Azure and Grid5000 Give me a Eucalyptus environment with 10 nodes Give 32 MPI nodes running on first Linux and then Windows Give me a Hadoop environment with 160 nodes Give me a 1000 BLAST instances linked to Grid5000 Run my application on Hadoop, Dryad, Amazon and Azure … and compare the performance

From Dynamic Provisioning to “RAIN” In FG, dynamic provisioning goes beyond the services offered by common scheduling tools that provide such features RAIN (Runtime Adaptable INsertion Configurator) We want to provide custom HPC environment, Cloud environment, or virtual networks on-demand with little effort Example: “rain” a Hadoop environment into a set of machines o fg-rain -n 8 -app Hadoop … o Users and administrators do not have to set up the Hadoop environment as it is being done for them

Future FG RAIN Commands fg-rain –h hostfile –iaas nimbus –image img fg-rain –h hostfile –paas hadoop … fg-rain –h hostfile –paas dryad … fg-rain –h hostfile –gaas gLite … fg-rain –h hostfile –image img Additional Authorization is required to use fg-rain without virtualization.

Rain in FutureGrid

Dynamic Provisioning Results Time elapsed between requesting a job and the jobs reported start time on the provisioned node. The numbers here are an average of 2 sets of experiments. Number of nodes

Status and Plan

Image Generation FeatureY1Y2Y3 QualityPrototype (proof of concept scripts) Production development and deployment for selected users General Deployment OS supportedUbuntuCentOS/UbuntuCentOS/Ubuntu/Fed ora/Suse Multi-tenancyNoYes SecurityNoYes AuthenticationNoYes – LDAPYes - LDAP Client InterfaceCLI CLI, Rest API, Portal Interface ScalabilityNoHigh (uses OpenNebula to deploy a VM per request) InteroperabilityPoor (based on base-images) High (VM with different OS)

Image Deployment FeatureY1Y2Y3 QualityPrototype (proof of concept scripts) Production development and deployment for selected users General Deployment Deployment type - Eucalyptus - Proof of Concept for HPC Eucalyptus, HPC (Moab/torque – xCAT) Eucalyptus, OpenStack, Nimbus, OpenNebula, HPC (Moab/torque) OS supportedUbuntu for EucalyptusCentOS for HPCCentOS/Ubuntu/Fed ora/Suse Multi-tenancyNoYes SecurityNoYes AuthenticationNoYes – LDAPYes - LDAP Client InterfaceCLI CLI, Rest API, Portal Interface

Image Repository FeatureY1Y2Y3 QualityEarly developmentProduction development General Deployment Client-Server Communication SshTLS/SSL Sockets Multi-tenancyYes – limitedYes SecurityYes - sshYes AuthenticationYes – sshYes – LDAPYes - LDAP Client InterfaceCLICLI, Rest APICLI, Rest API, Portal Interface Manage ImagesStore, retrieve, modify metadata, share Manage UsersNousers, roles and quotas StatisticsNoYes (#Images, usage,logs) Storage BackendFilesystemFilesystem, MongoDB, Cumulus, Swift, Mysql

Lessons Learned Users can customize bare metal images We provide base images that can be extended We have developed an environment allowing multiple users to do this at the same time Changing version of XCAT Moab supports a different kind of dynamic provisioning. E.g. Administrator needs to provide the image (not scalable)