Advanced Computing and Information Systems laboratory A Case for Grid Computing on Virtual Machines Renato Figueiredo Assistant Professor ACIS Laboratory,

Slides:



Advertisements
Similar presentations
Wei Lu 1, Kate Keahey 2, Tim Freeman 2, Frank Siebenlist 2 1 Indiana University, 2 Argonne National Lab
Advertisements

Virtualization Dr. Michael L. Collard
Education and training on FutureGrig Salt Lake City, Utah July 18 th 2011 Presented by Renato Figueiredo
1 Scoped and Approximate Queries in a Relational Grid Information Service Dong Lu, Peter A. Dinda, Jason A. Skicewicz Prescience Lab, Dept. of Computer.
Nondeterministic Queries in a Relational Grid Information Service Peter A. Dinda Dong Lu Prescience Lab Department of Computer Science Northwestern University.
Virtualization and Cloud Computing. Definition Virtualization is the ability to run multiple operating systems on a single physical system and share the.
The Case for Enterprise Ready Virtual Private Clouds Timothy Wood, Alexandre Gerber *, K.K. Ramakrishnan *, Jacobus van der Merwe *, and Prashant Shenoy.
Advanced Computing and Information Systems laboratory Educational Virtual Clusters for On- demand MPI/Hadoop/Condor in FutureGrid Renato Figueiredo Panoat.
A Case for Virtualizing Nodes on Network Experimentation Testbeds Konrad Lorincz Harvard University June 1, 2015June 1, 2015June 1, 2015.
Towards High-Availability for IP Telephony using Virtual Machines Devdutt Patnaik, Ashish Bijlani and Vishal K Singh.
1 Virtual Machine Resource Monitoring and Networking of Virtual Machines Ananth I. Sundararaj Department of Computer Science Northwestern University July.
1 In VINI Veritas: Realistic and Controlled Network Experimentation Jennifer Rexford with Andy Bavier, Nick Feamster, Mark Huang, and Larry Peterson
Towards Virtual Networks for Virtual Machine Grid Computing Ananth I. Sundararaj Peter A. Dinda Prescience Lab Department of Computer Science Northwestern.
Automatic Run-time Adaptation in Virtual Execution Environments Ananth I. Sundararaj Advisor: Peter A. Dinda Prescience Lab Department of Computer Science.
Increasing Application Performance In Virtual Environments Through Run-time Inference and Adaptation Ananth I. Sundararaj Ashish Gupta Peter A. Dinda Prescience.
Increasing Application Performance In Virtual Environments Through Run-time Inference and Adaptation Ananth I. Sundararaj Ashish Gupta Peter A. Dinda Prescience.
Virtuoso: Distributed Computing Using Virtual Machines Peter A. Dinda Prescience Lab Department of Computer Science Northwestern University
Virtuoso: Distributed Computing Using Virtual Machines Peter A. Dinda Prescience Lab Department of Computer Science Northwestern University
Dynamic Topology Adaptation of Virtual Networks of Virtual Machines Ananth I. Sundararaj Ashish Gupta Peter A. Dinda Prescience Lab Department of Computer.
Towards Virtual Networks for Virtual Machine Grid Computing Ananth I. Sundararaj Peter A. Dinda Prescience Lab Department of Computer Science Northwestern.
The Whats and Whys of Whole System Virtualization Peter A. Dinda Prescience Lab Department of Computer Science Northwestern University
Dynamic Topology Adaptation of Virtual Networks of Virtual Machines Ananth I. Sundararaj Ashish Gupta Peter A. Dinda Prescience Lab Department of Computer.
Jennifer Rexford Princeton University MW 11:00am-12:20pm SDN Software Stack COS 597E: Software Defined Networking.
The Origin of the VM/370 Time-sharing system Presented by Niranjan Soundararajan.
+ Virtualization in Clusters and Grids Dr. Lizhe Wang.
Distributed Systems CS Virtualization- Overview Lecture 22, Dec 4, 2013 Mohammad Hammoud 1.
Jaeyoung Yoon Computer Sciences Department University of Wisconsin-Madison Virtual Machines in Condor.
CSE598C Virtual Machines and Their Applications Operating System Support for Virtual Machines Coauthored by Samuel T. King, George W. Dunlap and Peter.
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
Cloud Computing Why is it called the cloud?.
Virtual Infrastructure in the Grid Kate Keahey Argonne National Laboratory.
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Microkernels, virtualization, exokernels Tutorial 1 – CSC469.
A Survey on Virtualization Technologies. Virtualization is “HOT” Microsoft acquires Connectix Corp. EMC acquires VMware Veritas acquires Ejascent IBM,
Disco : Running commodity operating system on scalable multiprocessor Edouard et al. Presented by Jonathan Walpole (based on a slide set from Vidhya Sivasankaran)
CS533 Concepts of Operating Systems Jonathan Walpole.
Operating System Support for Virtual Machines Samuel T. King, George W. Dunlap,Peter M.Chen Presented By, Rajesh 1 References [1] Virtual Machines: Supporting.
Virtualization Concepts Presented by: Mariano Diaz.
การติดตั้งและทดสอบการทำคลัสเต อร์เสมือนบน Xen, ROCKS, และไท ยกริด Roll Implementation of Virtualization Clusters based on Xen, ROCKS, and ThaiGrid Roll.
Grid Appliance – On the Design of Self-Organizing, Decentralized Grids David Wolinsky, Arjun Prakash, and Renato Figueiredo ACIS Lab at the University.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
Simplifying Resource Sharing in Voluntary Grid Computing with the Grid Appliance David Wolinsky Renato Figueiredo ACIS Lab University of Florida.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
Center for Autonomic Computing Intel Portland, April 30, 2010 Autonomic Virtual Networks and Applications in Cloud and Collaborative Computing Environments.
Advanced Computing and Information Systems laboratory IP over P2P: Enabling Self- configuring Virtual IP Networks for Grid Computing Arijit Ganguly, Abhishek.
Copyright © cs-tutorial.com. Overview Introduction Architecture Implementation Evaluation.
Virtual Workspaces Kate Keahey Argonne National Laboratory.
 Virtual machine systems: simulators for multiple copies of a machine on itself.  Virtual machine (VM): the simulated machine.  Virtual machine monitor.
VMware vSphere Configuration and Management v6
Types of Operating Systems 1 Computer Engineering Department Distributed Systems Course Assoc. Prof. Dr. Ahmet Sayar Kocaeli University - Fall 2015.
A. Frank - P. Weisberg Operating Systems Structure of Operating Systems.
Full and Para Virtualization
Introduction Why are virtual machines interesting?
1 Agility in Virtualized Utility Computing Hangwei Qian, Elliot Miller, Wei Zhang Michael Rabinovich, Craig E. Wills {EECS Department, Case Western Reserve.
Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing,
Distributed Systems Unit – 1 Concepts of DS By :- Maulik V. Dhamecha Maulik V. Dhamecha (M.Tech.)
Hands-On Virtual Computing
Advanced Computing and Information Systems laboratory The Advanced Computing and Information Systems Laboratory José Fortes Dep. of Electrical and Computer.
Protection of Processes Security and privacy of data is challenging currently. Protecting information – Not limited to hardware. – Depends on innovation.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Virtualization Review and Discussion
Clouds , Grids and Clusters
GGF15 – Grids and Network Virtualization
Middleware for Grid Computing On Virtual Machines
A Survey on Virtualization Technologies
CSE 451: Operating Systems Autumn Module 24 Virtual Machine Monitors
Presentation transcript:

Advanced Computing and Information Systems laboratory A Case for Grid Computing on Virtual Machines Renato Figueiredo Assistant Professor ACIS Laboratory, Dept. of ECE University of Florida José Fortes ACIS Laboratory, Dept. of ECE University of Florida Peter Dinda Prescience Lab, Dept. of Computer Science Northwestern University

Advanced Computing and Information Systems laboratory 2 The “Grid problem” “Flexible, secure, coordinated resource sharing among dynamic collections of individuals, institutions, and resources” 1 1 “ The Anatomy of the Grid: Enabling Scalable Virtual Organizations”, I. Foster, C. Kesselman, S. Tuecke. International J. Supercomputer Applications, 15(3), 2001

Advanced Computing and Information Systems laboratory 3 Example – PUNCH Since 1995 >1,000 users >100,000 jobs Kapadia, Fortes, Lundstrom, Adabala, Figueiredo et al

Advanced Computing and Information Systems laboratory 4 Resource sharing Traditional solutions: Multi-task operating systems User accounts File systems Evolved from centrally-admin. domains Functionality available for reuse However, Grids span administrative domains

Advanced Computing and Information Systems laboratory 5 Sharing – owner’s perspective I own a resource (e.g. cluster) and wish to sell/donate cycles to a Grid √User “A” is trusted and uses an environment common to my cluster ×If user “B” is not to be trusted? May compromise resource, other users ×If user “C” has different O/S, application needs? Administrative overhead May not be possible to support “C” without dedicating resource or interfering with other users A B C

Advanced Computing and Information Systems laboratory 6 Sharing – user’s perspective I wish to use cycles from a Grid I develop my apps using standard Grid interfaces, and trust users who share resource A ×If I have a grid-unaware application? Provider B may not support the environment my application expects: O/S, libraries, packages, … ×If I do not trust who is sharing a resource C? If another user compromises C’s O/S, they also compromise my work A BC

Advanced Computing and Information Systems laboratory 7 Alternatives? “Classic” Virtual Machines (VMs) Virtualization of instruction sets (ISAs) Language-independent, binary-compatible (not JVM) 70’s (IBM 360/370..) – 00’s (VMware, Connectix, zVM)

Advanced Computing and Information Systems laboratory 8 “Classic” Virtual Machines “A virtual machine is taken to be an efficient, isolated, duplicate copy of the real machine” 2 “A statistically dominant subset of the virtual processor’s instructions is executed directly by the real processor” 2 “…transforms the single machine interface into the illusion of many” 3 “Any program run under the VM has an effect identical with that demonstrated if the program had been run in the original machine directly” 2 2 “Formal Requirements for Virtualizable Third-Generation Architectures”, G. Popek and R. Goldberg, Communications of the ACM, 17(7), July “Survey of Virtual Machine Research”, R. Goldberg, IEEE Computer, June 1974

Advanced Computing and Information Systems laboratory 9 VMs for Grid computing Security VMs isolated from physical resource, other VMs Flexibility/customization Entire environments (O/S + applications) Site independence VM configuration independent of physical resource Binary compatibility Resource control VM1 (Linux RH7.3) VM2 (Win98) Physical (Win2000)

Advanced Computing and Information Systems laboratory 10 Outline Motivations VMs for Grid Computing Architecture Challenges Performance analyses Related work Outlook and conclusions

Advanced Computing and Information Systems laboratory 11 How can VMs be deployed? Statically Like any other node on the network, except it is virtual Not controlled by middleware Dynamically May be created, terminated by middleware User-customized Per-user state, persistent A personal, virtual workspace One-for-many, “clonable” State shared across users; non-persistent Sandboxes; application-tailored nodes

Advanced Computing and Information Systems laboratory 12 Architecture – dynamic VMs Indirection layer: Physical resources: where virtual machines are instantiated Virtual machines: where application execution takes place Coordination: Grid middleware

Advanced Computing and Information Systems laboratory 13 Middleware Abstraction: VM consists of a process (VMM) and data (system image) Core middleware support is available VM-raised challenges Resource and information management How to represent VMs as resources? How to instantiate, configure, terminate VMMs? Data management How to provide (large) system images to VMs? How to access user data from within VM instances?

Advanced Computing and Information Systems laboratory 14 Image management Proxy-based Grid virtual file systems On-demand transfers (NFS virtualization) RedHat 7.3: 1.3GB, <5% reboot+exec SpecSEIS User-level extensions for client caching/sharing Shareable (read) portions NFS client NFS server proxy ssh tunnel disk cache NFS protocol inter-proxy extensions [HPDC’2001] VM image

Advanced Computing and Information Systems laboratory 15 Resource management Extensions to Grid information services (GIS) VMs can be active/inactive VMs can be assigned to different physical resources URGIS project GIS based on the relational data model Virtual indirection Virtualization table associates unique id of virtual resources with unique ids of their constituent physical resources Futures An URGIS object that does not yet exist Futures table of unique ids

Advanced Computing and Information Systems laboratory 16 GIS extensions Compositional queries (joins) “Find physical machines which can instantiate a virtual machine with 1 GB of memory” “Find sets of four different virtual machines on the same network with a total memory between 512 MB and 1 GB” Virtual/future nature of resource hidden unless query explicitly requests it

Advanced Computing and Information Systems laboratory 17 Example: In-VIGO virtual workspace Front end ‘F’ Physical server pool P User ‘X’ Image Server I Data Server D1 User ‘Y’ Data Server D2 1: user request Information service 2: query (data, image, compute server) 3: setup VM image V1 X 5: copy/access user data 6: return handler to user (URL) 7: VNC X-window, HTTP file manager 4: start VM User request V2 Y isolation How fast to instantiate? Run-time overhead?

Advanced Computing and Information Systems laboratory 18 Performance – VM instantiation Instantiate VM “clone” via Globus GRAM Persistent (full copy) vs. non-persistent (link to base disk, writes to separate file) Full state copying is expensive VM can be rebooted, or resumed from checkpoint Restoring from post-boot state has lower latency Experimental setup: physical: dual Pentium III 933MHz, 512MB memory, RedHat 7.1, 30GB disk; virtual: Vmware Workstation 3.0a, 128MB memory, 2GB virtual disk, RedHat 2.0

Advanced Computing and Information Systems laboratory 19 Performance – VM instantiation Local and mounted via virtual file system Disk caching – low latency StartupDiskGrid Virtual FSLANWAN Reboot48sCache: cold121s434s Cache: warm52s56s Resume4sCache: cold80s1386s Cache: warm7s16s Experimental setup: Physical client is a dual Pentium-4, 1.8GHz, 1GB memory, 18GB Disk, RedHat 7.3. Virtual client: 128MB memory, 1.3GB disk, RedHat 7.3. LAN server is an IBM zSeries virtual machine, RedHat 7.1, 32GB disk, 256MB memory. WAN server is a VMware virtual machine, identical configuration to virtual client. WAN GridVFS is tunneled through ssh between UFL and NWU.

Advanced Computing and Information Systems laboratory 20 Performance – VM run-time ApplicationResourceExecTime (10^3 s) Overhead SpecHPC Seismic (serial, medium) Physical16.4N/A VM, local % VM, Grid virtual FS % SpecHPC Climate (serial, medium) Physical9.31N/A VM, local % VM, Grid virtual FS % Experimental setup: physical: dual Pentium III 933MHz, 512MB memory, RedHat 7.1, 30GB disk; virtual: Vmware Workstation 3.0a, 128MB memory, 2GB virtual disk, RedHat 2.0 NFS-based grid virtual file system between UFL (client) and NWU (server) Small relative virtualization overhead; compute-intensive

Advanced Computing and Information Systems laboratory 21 Related work Entropia virtual machines Application-level sandbox via Win32 binary modifications; no full O/S virtualization Denali at U. Washington Light-weight virtual machines; ISA modifications CoVirt at U. Michigan; User Mode Linux O/S VMMs, host extensions for efficiency “Collective” at Stanford Migration and caching of personal VM workspaces Internet Suspend/Resume at CMU/Intel Migration of VM environment for mobile users; explicit copy-in/copy-out of entire state files

Advanced Computing and Information Systems laboratory 22 Outlook Interconnecting VMs via virtual networks Virtual nodes: VMs Virtual switches, routers, bridges: host processes Virtual links: tunneling through physical resources Layer-3 virtual networks (e.g. VPNs) Layer-2 virtual networks (virtual bridges) “In-VIGO” On-demand virtual systems for Grid computing

Advanced Computing and Information Systems laboratory 23 Conclusions VMs enable fundamentally different approach to Grid computing: Physical resources – Grid-managed distributed providers of virtual resources Virtual resources – engines where computation occurs; logically connected as virtual network domains Towards secure, flexible sharing of resources Demonstrated feasibility of the architecture For current VM technology, compute-intensive tasks On-demand transfer; difference-copy, resumable clones; application-transparent image caches

Advanced Computing and Information Systems laboratory 24 Acknowledgments NSF Middleware Initiative NSF Research Resources IBM Shared University Research VMware Ivan Krsul, In-VIGO and Virtuoso teams at UFL/NWU