Advanced Computing and Information Systems laboratory Self-configuring Condor Virtual Machine Appliances for Ad-Hoc Grids Renato Figueiredo Arijit Ganguly,

Slides:



Advertisements
Similar presentations
Virtual Machine Technology Dr. Gregor von Laszewski Dr. Lizhe Wang.
Advertisements

Advanced Computing and Information Systems laboratory Virtual Private Clusters: Virtual Appliances and Networks in the Cloud Renato Figueiredo ACIS Lab.
CPSC Network Layer4-1 IP addresses: how to get one? Q: How does a host get IP address? r hard-coded by system admin in a file m Windows: control-panel->network->configuration-
Transitioning to IPv6 April 15,2005 Presented By: Richard Moore PBS Enterprise Technology.
Center for Autonomic Computing Intel Portland, April 30, 2010 Autonomic Virtual Networks and Applications in Cloud and Collaborative Computing Environments.
1.1 © 2004 Pearson Education, Inc. Exam Managing and Maintaining a Microsoft® Windows® Server 2003 Environment Lesson 1: Introducing Windows Server.
Module 5: Configuring Access for Remote Clients and Networks.
Nada Abdulla Ahmed.  SmoothWall Express is an open source firewall distribution based on the GNU/Linux operating system. Designed for ease of use, SmoothWall.
Advanced Computing and Information Systems laboratory Educational Virtual Clusters for On- demand MPI/Hadoop/Condor in FutureGrid Renato Figueiredo Panoat.
Building Your Own Firewall Chapter 10. Learning Objectives List and define the two categories of firewalls Explain why desktop firewalls are used Explain.
Towards Virtual Networks for Virtual Machine Grid Computing Ananth I. Sundararaj Peter A. Dinda Prescience Lab Department of Computer Science Northwestern.
Automatic Run-time Adaptation in Virtual Execution Environments Ananth I. Sundararaj Advisor: Peter A. Dinda Prescience Lab Department of Computer Science.
Subnetting.
1 Version 3.0 Module 9 TCP/IP Protocol and IP Addressing.
The Whats and Whys of Whole System Virtualization Peter A. Dinda Prescience Lab Department of Computer Science Northwestern University
Center for Autonomic Computing Intel Portland, April 30, 2010 Autonomic Virtual Networks and Applications in Cloud and Collaborative Computing Environments.
TCP/IP Tools Lesson 5. Objectives Skills/ConceptsObjective Domain Description Objective Domain Number Using basic TCP/IP commands Understanding TCP/IP3.6.
Network+ Guide to Networks 6 th Edition Chapter 10 Virtual Networks and Remote Access.
© 2005,2006 NeoAccel Inc. Training Access Modes. © 2005,2006 NeoAccel Inc. Agenda 2. Access Terminals 6. Quick Access Terminal Client 3. SSL VPN-Plus.
© 2007 Cisco Systems, Inc. All rights reserved.ISCW-Mod3_L7 1 Network Security 2 Module 6 – Configure Remote Access VPN.
Week #10 Objectives: Remote Access and Mobile Computing Configure Mobile Computer and Device Settings Configure Remote Desktop and Remote Assistance for.
Network Services Lesson 6. Objectives Skills/ConceptsObjective Domain Description Objective Domain Number Setting up common networking services Understanding.
LB VIP:Input Endpoint Internal Endpoint foo.cloudapp.net  VIP.
Support Protocols and Technologies. Topics Filling in the gaps we need to make for IP forwarding work in practice – Getting IP addresses (DHCP) – Mapping.
On the Design of Autonomic, Decentralized VPNs David Wolinsky, Kyungyong Lee, Oscar Boykin, and Renato Figueiredo ACIS P2P Group University of Florida.
Network+ Guide to Networks 6 th Edition Chapter 10 Virtual Networks and Remote Access.
Module 3: Planning and Troubleshooting Routing and Switching.
About the Presentations The presentations cover the objectives found in the opening of each chapter. All chapter objectives are listed in the beginning.
DHCP Dynamic Host Configuration Protocol. Introduction Client administration:  IP address management: They need to ease the process of joining the network.
Advanced Computing and Information Systems laboratory Plug-and-play Virtual Appliance Clusters Running Hadoop Dr. Renato Figueiredo ACIS Lab - University.
+ CS 325: CS Hardware and Software Organization and Architecture Cloud Architectures.
Remote Access Chapter 4. Learning Objectives Understand implications of IEEE 802.1x and how it is used Understand VPN technology and its uses for securing.
Remote Access Chapter 4. Learning Objectives Understand implications of IEEE 802.1x and how it is used Understand VPN technology and its uses for securing.
1.  PRAGMA Grid test-bed : Shares clusters which managed by multiple sites Realizes a large-scale computational environment. › Expects as a platform.
Grid Appliance – On the Design of Self-Organizing, Decentralized Grids David Wolinsky, Arjun Prakash, and Renato Figueiredo ACIS Lab at the University.
Windows 7 Firewall.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Module 12: Routing Fundamentals. Routing Overview Configuring Routing and Remote Access as a Router Quality of Service.
CIS 3360: Internet: Network Layer Introduction Cliff Zou Spring 2012.
The Open Source Virtual Lab: a Case Study Authors: E. Damiani, F. Frati, D. Rebeccani, M. Anisetti, V. Bellandi and U. Raimondi University of Milan Department.
Windows Small Business Server 2003 Setting up and Connecting David Overton Partner Technical Specialist.
Module 2: Installing and Maintaining ISA Server. Overview Installing ISA Server 2004 Choosing ISA Server Clients Installing and Configuring Firewall Clients.
Advanced Computing and Information Systems laboratory Virtual Appliances for Training and Education in FutureGrid Renato Figueiredo Arjun Prakash, David.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
Module 11: Implementing ISA Server 2004 Enterprise Edition.
Simplifying Resource Sharing in Voluntary Grid Computing with the Grid Appliance David Wolinsky Renato Figueiredo ACIS Lab University of Florida.
Advanced Topics StratusLab Tutorial (Orsay, France) 28 November 2012.
Virtual Private Ad Hoc Networking Jeroen Hoebeke, Gerry Holderbeke, Ingrid Moerman, Bard Dhoedt and Piet Demeester 2006 July 15, 2009.
Center for Autonomic Computing Intel Portland, April 30, 2010 Autonomic Virtual Networks and Applications in Cloud and Collaborative Computing Environments.
ETICS All Hands meeting Bologna, October 23-25, 2006 NMI and Condor: Status + Future Plans Andy PAVLO Peter COUVARES Becky GIETZEL.
A Survey on Parallel Computing in Heterogeneous Grid Environments Takeshi Sekiya Chikayama-Taura Laboratory M1 Nov 24, 2006.
OS Services And Networking Support Juan Wang Qi Pan Department of Computer Science Southeastern University August 1999.
Advanced Computing and Information Systems laboratory IP over P2P: Enabling Self- configuring Virtual IP Networks for Grid Computing Arijit Ganguly, Abhishek.
EVGM081 Multi-Site Virtual Cluster: A User-Oriented, Distributed Deployment and Management Mechanism for Grid Computing Environments Takahiro Hirofuchi,
Plethora: Infrastructure and System Design. Introduction Peer-to-Peer (P2P) networks: –Self-organizing distributed systems –Nodes receive and provide.
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Tools and techniques for managing virtual machine images Andreas.
Module 10: Windows Firewall and Caching Fundamentals.
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Hands-On Virtual Computing
You there? Yes Network Health Monitoring Heartbeats are sent to monitor health status of network interfaces Are sent over all cluster.
CEG 2400 FALL 2012 Chapter 10 Virtual Networks and Remote Access 1.
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
Group # 14 Dhairya Gala Priyank Shah. Introduction to Grid Appliance The Grid appliance is a plug-and-play virtual machine appliance intended for Grid.
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
Securing the Network Perimeter with ISA 2004
Network+ Guide to Networks 6th Edition
Network Virtualization
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Goals Introduce the Windows Server 2003 family of operating systems
Cloud computing mechanisms
Presentation transcript:

Advanced Computing and Information Systems laboratory Self-configuring Condor Virtual Machine Appliances for Ad-Hoc Grids Renato Figueiredo Arijit Ganguly, David Wolinsky, J. Rhett Aultman, P. Oscar Boykin, ACIS Lab, University of Florida

Advanced Computing and Information Systems laboratory 2 Outline Motivations Background Condor Virtual Appliance: features On-going and future work

Advanced Computing and Information Systems laboratory 3 Motivations Goal: plug-and-play deployment of Condor grids High-throughput computing; LAN and WAN Collaboration: file systems, messaging,.. Synergistic approach: VM + virtual network + Condor “WOWs” are wide-area NOWs, where: Nodes are virtual machines Network is virtual: IP-over-P2P (IPOP) overlay VMs provide: Sandboxing; software packaging; decoupling Virtual network provides: Virtual private LAN over WAN; self-configuring and capable of firewall/NAT traversal Condor provides: Match-making, reliable scheduling, … unmodified

Advanced Computing and Information Systems laboratory 4 1. Prime base VM image with O/S, Condor, Virtual network; publish (Web/Torrent) Condor WOWs - outlook 2. Download image; boot using free VM monitor (e.g. VMware Player or Server) 4. Download base and custom VM images; boot up 3. Create virtual IP namespace for pool: MyGrid: / Prime custom image with virtual namespace, desired tools Bootstrap manager(s) VMs obtain IP addresses from MyGrid Virtual DHCP server, join virtual IP network, discover available manager(s), and join pool b. VMs obtain IP addresses from OtherGrid Virtual DHCP server, join virtual IP network, discover available manager(s), and join pool

Advanced Computing and Information Systems laboratory 5 Condor WOW snapshot Zurich Gainesville Long Beach

Advanced Computing and Information Systems laboratory 6 Roadmap The basics: 1.1 VMs and appliances 1.2 IPOP: IP-over-P2P virtual network 1.3 Grid Appliance and Condor The details: 2.1 Customization, updates 2.2 User interface 2.3 Security 2.4 Performance Usage experience

Advanced Computing and Information Systems laboratory 7 1.1: VMs and appliances System VMs: VMware, KVM, Xen Homogenous system Sandboxing Co-exist with unmodified hosts Virtual appliances: Hardware/software configuration packaged in easy to deploy VM images Only dependences: ISA (x86), VMM

Advanced Computing and Information Systems laboratory 8 1.2: IPOP virtual networking Key technique: IP-over-P2P tunneling Interconnect VM appliances WAN VMs perceive a virtual LAN environment IPOP is self-configuring Avoid administrative overhead of VPNs NAT and firewall traversal IPOP is scalable and robust P2P routing deals with node joins and leaves IPOP networks are isolated One or more private IP address spaces Decentralized DHCP serves addresses for each space

Advanced Computing and Information Systems laboratory 9 1.2: IPOP virtual networking App IPOP Node B eth0 ( ) IPOP Node A eth0 ( ) A B tap0 ( ) tap0 ( ) P2P Overlay Structured overlay network topology Bootstrap 1-hop IP tunnels on demand Discover NAT mappings; decentralized hole punching VM keeps IPOP address even if it migrates on WAN [Ganguly et al, IPDPS 2006, HPDC 2006]

Advanced Computing and Information Systems laboratory Grid appliance and Condor Base: Debian Linux; Condor; IPOP Works on x86 Linux/Windows/MacOS; VMware, KVM/QEMU 157MB zipped Uses NAT and host-only NICs No need to get IP address on host network Managed negotiator/collector VMs Easy to deploy schedd/startd VMs Flocking is easy – virtual network is a LAN

Advanced Computing and Information Systems laboratory : Customization and updates VM image: Virtual Disks Portable medium for data Growable after distribution Disks are logically stacked Leverage UnionFS file system Three stacks: Base – O/S, Condor, IPOP Module – site specific configuration (e.g. nanoHUB) Home – user persistent data Major updates: replace base/module Minor updates: automatic, apt-based

Advanced Computing and Information Systems laboratory : User interface (Windows host) VM console: X11 GUI Host-mounted loop-back Samba folder Loopback SSH

Advanced Computing and Information Systems laboratory : User interface (Mac host) VM console: X11 GUI Host-mounted loop-back Samba folder Loopback SSH

Advanced Computing and Information Systems laboratory : User interface (Linux host) VM console: X11 GUI Host-mounted loop-back Samba folder Loopback SSH

Advanced Computing and Information Systems laboratory Security Appliance firewall eth0: block all outgoing Internet packets Except DHCP, DNS, IPOP’s UDP port Only traffic within WOW allowed eth1 (host-only): allow ssh, Samba IPsec X.509 host certificates Authentication and end-to-end encryption VM joins WOW only with signed certificate bound to its virtual IP Private net/netmask: ~10 lines of IPsec configuration for an entire class A network!

Advanced Computing and Information Systems laboratory : Performance User-level C# IPOP implementation (UDP): Link bandwidth: 25-30Mbit/s Latency overhead: ~4ms Connection times: ~5-10s to join P2P ring and obtain DHCP address ~10s to create shortcuts, UDP hole-punching SimpleScalar 3.0 (cycle-accurate CPU simulator)

Advanced Computing and Information Systems laboratory 17 Experiences Bootstrap WOW with VMs at UF and partners Currently ~300 VMs, IPOP overlay routers (Planetlab) Exercised with 10,000s of Condor jobs from real users nanoHUB: 3-week long, 9,000-job batch (BioMoca) submitted via a Condor-G gateway P2Psim, CH3D, SimpleScalar Pursuing interactions with users and the Condor community for broader dissemination

Advanced Computing and Information Systems laboratory 18 Time scales and expertise Development of baseline VM image: VM/Condor/IPOP expertise; weeks/months Development of custom module: Domain-specific expertise; hours/days/weeks Deployment of VM appliance: No previous experience with VMs or Condor minutes to download and install VMM minutes to download and unzip appliance minutes to boot appliance, automatically connect to a Condor pool, run condor_status and a demo condor_submit job

Advanced Computing and Information Systems laboratory 19 On-going and future work Enhancing self-organization at the Condor level: Structured P2P for manager publish/discovery Distributed hash table (DHT); primary and flocking Condor integration via configuration files, DHT scripts Unstructured P2P for matchmaking Publish/replicate/cache classads on P2P overlay Support for arbitrary queries Condor integration: proxies for collector/negotiator Decentralized storage, cooperative caching Virtual file systems (NFS proxies) Distribution of updates, read-only code repositories Caching and COW for diskless, net-boot appliances

Advanced Computing and Information Systems laboratory 20 Acknowledgments National Science Foundation NMI, CI-TEAM SURA SCOOP (Coastal Ocean Observing and Prediction) Publications, Brunet/IPOP code (GPL’ed C#), Condor Grid appliance

Advanced Computing and Information Systems laboratory 21 Questions?

Advanced Computing and Information Systems laboratory 22 Self-organizing NAT traversal, shortcuts Node A Node B CTM request: connect to me at my NAT IP:port Sends CTM request - A starts exchanging IP packets with B - Traffic inspection triggers request to create shortcut - Connect-to-me (CTM) - “A” tells “B” its known address(es): - “A” had learned NATed public IP/port when it joined overlay

Advanced Computing and Information Systems laboratory 23 - “B” sends CTM reply – routed through overlay - “B” tells “A” its address(es) - “B” initiates linking protocol by attempting to connect to “A” directly Node A Node B CTM reply through overlay: send NAT (IP:port) B Self-organizing NAT traversal, shortcuts Link request: NAT endpoint (IP:port) A

Advanced Computing and Information Systems laboratory 24 - B’s linking protocol message to A pokes hole on B’s NAT - A’s linking protocol message to B pokes hole on A’s NAT CTM protocol establishes direct shortcut A Gets CTM reply; initiates linking Node A Node B Self-organizing NAT traversal, shortcuts

Advanced Computing and Information Systems laboratory 25 Performance considerations CPU-intensive application, Condor SimpleScalar 3.0d execution-driven computer architecture simulator

Advanced Computing and Information Systems laboratory 26 Performance considerations I/O: PostMark Version 1.51 Parameters: Minimum file size: 500 bytes Maximum file size: 4.77 MB Transactions: 5,000

Advanced Computing and Information Systems laboratory 27 Performance considerations User-level C# IPOP implementation (UDP): Link bandwidth: 25-30Mbit/s (LAN) Latency overhead: ~4ms Connection times: (Fine-tuning has reduced mean acquire time to ~ 6-10s, with degree of redundancy n=8)

Advanced Computing and Information Systems laboratory 28 Condor Appliance on a desktop Linux, Condor, IPOP Domain- specific tools User files Swap VM Hardware configuration

Advanced Computing and Information Systems laboratory 29 Related Work Virtual Networking VIOLIN VNET; topology adaptation ViNe Internet Indirection Infrastructure (i3) Support for mobility, multicast, anycast Decouples packet sending from receiving Based on Chord p2p protocol IPv6 tunneling IPv6 over UDP (Teredo protocol) IPv6 over P2P (P6P)