Predrag Buncic (CERN/PH-SFT) Virtualizing LHC Applications.

Slides:



Advertisements
Similar presentations
Virtual Machine Technology Dr. Gregor von Laszewski Dr. Lizhe Wang.
Advertisements

Copyright 2009 FUJITSU TECHNOLOGY SOLUTIONS PRIMERGY Servers and Windows Server® 2008 R2 Benefit from an efficient, high performance and flexible platform.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Network+ Guide to Networks 6 th Edition Chapter 10 Virtual Networks and Remote Access.
Network+ Guide to Networks 6 th Edition Chapter 10 Virtual Networks and Remote Access.
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
Cloud computing is the use of computing resources (hardware and software) that are delivered as a service over the Internet. Cloud is the metaphor for.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 7 2/23/2015.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
1 port BOSS on Wenjing Wu (IHEP-CC)
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Network Plus Virtualization Concepts. Virtualization Overview Virtualization is the emulation of a computer environment called a Virtual Machine. A Hypervisor.
Grids, Clouds and the Community. Cloud Technology and the NGS Steve Thorn Edinburgh University Matteo Turilli, Oxford University Presented by David Fergusson.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
Predrag Buncic (CERN/PH-SFT) CernVM - a virtual software appliance for LHC applications C. Aguado-Sanchez 1), P. Buncic 1), L. Franco 1), A. Harutyunyan.
Predrag Buncic (CERN/PH-SFT) CernVM - a virtual software appliance for LHC applications C. Aguado-Sanchez 1), P. Buncic 1), L. Franco 1), S. Klemer 1),
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
Cloud Status Laurence Field IT/SDC 09/09/2014. Cloud Date Title 2 SaaS PaaS IaaS VMs on demand.
Copyright © cs-tutorial.com. Overview Introduction Architecture Implementation Evaluation.
Predrag Buncic (CERN/PH-SFT) WP9 - Workshop Summary
ALICE Offline Week | CERN | November 7, 2013 | Predrag Buncic AliEn, Clouds and Supercomputers Predrag Buncic With minor adjustments by Maarten Litmaath.
Desktop Virtualization
Predrag Buncic (CERN/PH-SFT) Introduction to WP9 Portable Analysis Environment Using Virtualization Technology IBM-VM 360, CERNVM,
Windows Azure Virtual Machines Anton Boyko. A Continuous Offering From Private to Public Cloud.
DPHEP Workshop CERN, December Predrag Buncic (CERN/PH-SFT) CernVM R&D Project Portable Analysis Environments using Virtualization.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
WLCG Overview Board, September 3 rd 2010 P. Mato, P.Buncic Use of multi-core and virtualization technologies.
Virtual Machines Created within the Virtualization layer, such as a hypervisor Shares the physical computer's CPU, hard disk, memory, and network interfaces.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
Ian Gable HEPiX Spring 2009, Umeå 1 VM CPU Benchmarking the HEPiX Way Manfred Alef, Ian Gable FZK Karlsruhe University of Victoria May 28, 2009.
1 Volunteer Computing at CERN past, present and future Ben Segal / CERN (describing the work of many people at CERN and elsewhere ) White Area lecture.
1 BOINC Case Study: Ben Segal / CERN with: Predrag Buncic / CERN Daniel Lombrana Gonzalez / Univ. Extremadura David Weir / Imperial.
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-1.
Workload management, virtualisation, clouds & multicore Andrew Lahiff.
SPI NIGHTLIES Alex Hodgkins. SPI nightlies  Build and test various software projects each night  Provide a nightlies summary page that displays all.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen Budapest
The CernVM Infrastructure Insights of a paradigmatic project Carlos Aguado Sanchez Jakob Blomer Predrag Buncic.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI VM Management Chair: Alexander Papaspyrou 2/25/
Predrag Buncic (CERN/PH-SFT) Software Packaging: Can Virtualization help?
1 Cloud Services Requirements and Challenges of Large International User Groups Laurence Field IT/SDC 2/12/2014.
2nd ASPERA Workshop May 2011, Barcelona, Spain P. Mato /CERN.
Predrag Buncic (CERN/PH-SFT) Virtualization – the road ahead.
The CernVM Project A new approach to software distribution Carlos Aguado Jakob Predrag
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
NA61 Collaboration Meeting CERN, December Predrag Buncic, Mihajlo Mudrinic CERN/PH-SFT Enabling long term data preservation.
Predrag Buncic (CERN/PH-SFT) CernVM Status. CERN, 24/10/ Virtualization R&D (WP9)  The aim of WP9 is to provide a complete, portable and easy.
EGI-InSPIRE RI EGI Webinar EGI-InSPIRE RI Porting your application to the EGI Federated Cloud 17 Feb
Journées Informatiques de l'IN2P May 2010, Aussois, France P. Mato /CERN.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Trusted Virtual Machine Images the HEPiX Point of View Tony Cass October 21 st 2011.
Amazon Web Services. Amazon Web Services (AWS) - robust, scalable and affordable infrastructure for cloud computing. This session is about:
Predrag Buncic (CERN/PH-SFT) Virtualization R&D (WP9) Status Report.
Predrag Buncic, CERN/PH-SFT The Future of CernVM.
CHEP 2010 Taipei, 19 October Predrag Buncic Jakob Blomer, Carlos Aguado Sanchez, Pere Mato, Artem Harutyunyan CERN/PH-SFT.
Guide to Operating Systems, 5th Edition
Use of HLT farm and Clouds in ALICE
Virtualisation for NA49/NA61
NA61/NA49 virtualisation:
Blueprint of Persistent Infrastructure as a Service
Dag Toppe Larsen UiB/CERN CERN,
What is Virtualization Last Update
Dag Toppe Larsen UiB/CERN CERN,
Virtualisation for NA49/NA61
Virtualization in the gLite Grid Middleware software process
CernVM Status Report Predrag Buncic (CERN/PH-SFT).
WLCG Collaboration Workshop;
Network+ Guide to Networks 6th Edition
Guide to Operating Systems, 5th Edition
PerformanceBridge Application Suite and Practice 2.0 IT Specifications
Presentation transcript:

Predrag Buncic (CERN/PH-SFT) Virtualizing LHC Applications

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis locally and on the Grid independent of physical software and hardware platform (Linux, Windows, MacOS)  Code check-out, edition, compilation, local small test, debugging, …  Grid submission, data access…  Event displays, interactive data analysis, …  Suspend, resume… Decouple application lifecycle from evolution of system infrastructure Reduce effort to install, maintain and keep up to date the experiment software Web site: 2 Virtualization R&D

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June From Application to Appliance Build types  Installable CD/DVD  Stub Image  Raw Filesystem Image  Netboot Image  Compressed Tar File  Demo CD/DVD (Live CD/DVD)  Raw Hard Disk Image  Vmware ® Virtual Appliance  Vmware ® ESX Server Virtual Appliance  Microsoft ® VHD Virtual Apliance  Xen Enterprise Virtual Appliance  Virtual Iron Virtual Appliance  Parallels Virtual Appliance  Amazon Machine Image  Update CD/DVD  Appliance Installable ISO Starting from experiment software… …ending with custom Linux specialised for a given task

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Login to Web interface 2. Create user account 3. Select experiment, appliance flavor and preferences As easy as 1,2,3

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June “Thin” Virtual Machine The experiment are packaging a lot of code  but really use only fraction of it at runtime CernVM downloads what is needed and puts it in the cache  Does not require persistent network connection (offline mode)

Publishing Releases 1. Each experiment is given a VM to install and test their software using own installation tools 2. Publishing is an atomic operation

CernVM Infrastructure

Service Level Status

Where are our users? ~800 different IP adresses

Download statistics

Download image types

CVMFS Performance

AFS vs CVMFS In this setup measure the time penalty  t=t AFS,CVMFS - t local  resulting from having the application binaries, search paths, and include paths reside on a network file system while running ROOT stressHEPIX benchmark.

CVMFS shows consistently better performance than AFS in case of ‘cold cache’ irrespectively if latency or bandwidth constraints Results We compare the performance of AFS (1.4.8) and current version of CVMS as well as development version of CVMFS that includes extra optimization

Proxy Server Proxy Server CernVM HTTP server HTTP server Proxy Server HTTP server HTTP server Proxy Server HTTP server HTTP server Proxy Server Scaling up… Proxy and slave servers could be deployed on strategic locations to reduce latency and provide redundancy

HTTP server HTTP server Proxy Server CernVM Content Distributio n Network WAN Use existing Content Delivery Networks to remove single point of failure  Amazon CloudFront (  Coral CDN ( LAN Use P2P like mechanism for discovery of nearby CernVMs and cache sharing between them. No need to manually setup proxy servers (but they could still be used where exist) LAN & WAN

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Release status Stable release  Available for download from Development release  Available now for download from devel/releaseshttp://rbuilder.cern.ch/project/cernvm- devel/releases Can be run on  Linux (VMware Player, VirtualBox, KVM)  Windows(VMware Player, VirtualBox)  Mac (VMware Fusion, VirtualBox, Parallels) Appliance can be configured and used with ALICE, LHCb, ATLAS, CMS and LCD software frameworks Future releases will probably come in two editions  Basic (text development environment, suitable for ssh login, ~200MB)  Desktop (full desktop environment, works on VMware & VirtualBox, ~500MB)

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June What’s next? Support for existing CernVM users Helping new ones get on board  LCD - Linear Collider Detector studies  NA61 (software and data access preservation) Continuing develoment of CVMFS  New version feature complete  Should go into production release by the end of summer Moving on to CernVM 2.0  Will be based on upstream SL5 (unlike current version based on RH4)  Release by the end of the year Continuing to develop tools (CoPilot) to ease deployment of CernVM in cloud like environments  Nimbus, EC2  BOINC - platform for computing using volunteered resources Possibly investigating possibilities for deployment on the Grid

GRID

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June CernVM as job hosting environemnt Ideally, users would like run their applications on the grid (or cloud) infrastructure in exactly the same conditions in which they were developed CernVM already provides development environment and can be deployed on cloud (EC2)  One image supports all four LHC experiments  Easily extensible to other communities

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Possible deployment scenario Hypervisor deployed on physical nodes running instances of CernVM  Number of instances under control of experiment software admin as well as instance parameters (# of cores, memory) Virtual machine images thin provisioned from a shared storage  Required to be able to move VMs between physical node VMs run on a private network  No incoming connectivity to VMs  Only limited outgoing connectivity via gateway node  Outgoing HTTP connectivity via caching proxy Access to storage via tactical storage proxy  Equivalent to HTTP proxy for data files HTTP Proxy Tactical Storage Proxy Shared Storage NAT Gateway CernVM

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Advantages Simple and should be able to fulfill the needs of experiments  Exactly the same environment for development and job execution  For as long as Pilot Jobs can be run in VM there should be no difference between this model and what is currently going on on the grid  Software can be efficiently installed using CVMFS HTTP proxy assures very fast access to software even if VM cache is cleared  Can accommodate multi-core jobs

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Addressing site concerns Should answer to site concerns  VMs are deployed on a private network, no incoming connectivity No shortage of IP numbers  Tactical storage proxy should provide mechanism to VM to efficiently access bulk of data files Possible implementation using xrootd  Depending on hypervisor choice monitoring of VM may or may not be different from current practices  CernVM has integrated rAA agent to assure appliance software updates However, these should be less frequent as less components are installed on the system and far less critical as they run on private network  All remaining network activity (beyond data access and HTTP) can be monitored and policy enforced on gateway node

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Variations… In the simplest case VMs are simply started according to predefined share per experiment An alternative is to deploy VM provisioning infrastructure that will instantiate VMs according to the request and specification made by authorized person in the experiment  Nimbus, Open Nebula or vendor tools (vSphere…)  Gives more freedom to experiments to express their optional requirements in terms of memory, number of jobs If we trust what vendors are telling us, we can over commit resources and let hypervisor and/or management tools do their job in optimizing resource usage

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Batch or not to batch now? For this to work we need a mechanism to force people to shutdown their VMs once they are not in use  On EC2 this is simple - Amazon charges your credit card for every CPU (wall time) hour and this is usually sufficient incentive  Do we need economy model for scheduling VMs?  Should we start thinking about accounting in terms of VM slot hours instead of real CPU hours?  Wouldn’t it make sense to start thinking in terms of advanced slot reservations instead of sending batch jobs to the queue? Hypervisors can again help to some extent  Unused VMs can be parked and resources used by others if we over commit them  They can spread and balance the load over available physical resources

Workshop on adapting applications and computing services to multi-core and virtualization, CERN, June Conclusions Lots of interest from LHC experiments and huge momentum in industry  ATLAS, LHCb, CMS, ALICE, LCD Hypervisors are nowadays available for free (Linux, Mac and Windows)  But managing tools and support are not CernVM approach solves the problem of efficient software distribution  Using its own dedicated file system  One (thin) image fits all Initially developed as user interface for laptop/desktop  Already deployable on the cloud (EC2, Nimbus)  Can be deployed on managed (and unmanaged infrastructure) without necessarily compromising the site security Deployment on the grid or in the computer centre environment requires changes to some of the current practices and thinking  Utilizing private networks to avoid shortage of IP numbers and to hide VMs form public internet  Use proxy/caches wherever possible  Move away from traditional batch job scheduling towards advanced slot reservations for VMs carry out the accounting in the same terms