Virtualisation for NA49/NA61

Slides:



Advertisements
Similar presentations
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Advertisements

Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
Apache : Installation, Configuration, Basic Security Presented by, Sandeep K Thopucherela, ECE Department.
Cambodia-India Entrepreneurship Development Centre - : :.... :-:-
Installing software on personal computer
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
1 port BOSS on Wenjing Wu (IHEP-CC)
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Data production using CernVM and lxCloud Dag Toppe Larsen Belgrade
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
Data production using CernVM and LxCloud Dag Toppe Larsen Warsaw,
Copyright © cs-tutorial.com. Overview Introduction Architecture Implementation Evaluation.
Installing, running, and maintaining large Linux Clusters at CERN Thorsten Kleinwort CERN-IT/FIO CHEP
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
WLCG Overview Board, September 3 rd 2010 P. Mato, P.Buncic Use of multi-core and virtualization technologies.
Windows Azure. Azure Application platform for the public cloud. Windows Azure is an operating system You can: – build a web application that runs.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen Budapest
Predrag Buncic (CERN/PH-SFT) Software Packaging: Can Virtualization help?
T3g software services Outline of the T3g Components R. Yoshida (ANL)
MIS Week 5 Site:
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Virtualisation: status and plans Dag Toppe Larsen
Amazon Web Services. Amazon Web Services (AWS) - robust, scalable and affordable infrastructure for cloud computing. This session is about:
System Architecture CS 560. Project Design The requirements describe the function of a system as seen by the client. The software team must design a system.
CVMFS Alessandro De Salvo Outline  CVMFS architecture  CVMFS usage in the.
Predrag Buncic, CERN/PH-SFT The Future of CernVM.
HEPiX Virtualisation working group Andrea Chierici INFN-CNAF Workshop CCR 2010.
CernVM and Volunteer Computing Ivan D Reid Brunel University London Laurence Field CERN.
Progress Apama Fundamentals
CernVM-FS vs Dataset Sharing
WLCG IPv6 deployment strategy
Chapter 6: Securing the Cloud
Use of HLT farm and Clouds in ALICE
Chapter 1: Introduction
Update on revised HEPiX Contextualization
WWW and HTTP King Fahd University of Petroleum & Minerals
Virtualisation for NA49/NA61
NA61/NA49 virtualisation:
Blueprint of Persistent Infrastructure as a Service
Dag Toppe Larsen UiB/CERN CERN,
Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław
Dag Toppe Larsen UiB/CERN CERN,
ATLAS Cloud Operations
GWE Core Grid Wizard Enterprise (
StratusLab Final Periodic Review
StratusLab Final Periodic Review
WLCG experiments FedCloud through VAC/VCycle in the EGI
ATLAS Software Distribution
Generator Services planning meeting
PES Lessons learned from large scale LSF scalability tests
Virtualization overview
CernVM Status Report Predrag Buncic (CERN/PH-SFT).
Discussions on group meeting
WLCG Collaboration Workshop;
Building and Testing using Condor
Haiyan Meng and Douglas Thain
CERN Certificates platform Emmanuel Ormancey / Anatoly Gladkov
Module 01 ETICS Overview ETICS Online Tutorials
Ivan Reid (Brunel University London/CMS)
Prof. Leonardo Mostarda University of Camerino
Azure Container Service
Web Servers (IIS and Apache)
Grid Computing Software Interface
Presentation transcript:

Virtualisation for NA49/NA61 Dag Toppe Larsen UiB/CERN Zagreb, 11.10.2011

Outline Recapitulation Reference cloud NA49/NA61 data processing Why virtualisation? CERNVM CVMFS Reference cloud NA49/NA61 data processing Status Next steps Outlook

Why virtualisation? Data preservation Very flexible Avoids complexity of grid Easy software distribution Processing not constrained to CERN Take advantage of new LXCLOUD Take advantage of commercial clouds, e.g. Amazon EC2 Can develop in same VM as data will be processed on – should reduce failing jobs

CERNVM: introduction Dedicated Linux distribution for virtual machines Currently based on SLC5 Newer and updated versions will be made available for new software Old versions will still be available for legacy analysis software Supports all common hypervisors Supports Amazon EC2 clouds

CERNVM: layout

CVMFS: introduction Distributed file system based on HTTP Read-only Distribution of binary files – no need for local compile & install All libraries & software that can not be expected to be found on “standard” Linux should be distributed Each experiment has one or more persons responsible for providing updates and resolve dependencies

CVMFS: software repositories Several repositories mounted under /cvmfs/ Each repository typically corresponds to one “experiment” (or other “entity”) Experiments have “localised” names, e.g. /cvmfs/na61.cern.ch/ Common software in separate repositories, e.g. ROOT in /cvmfs/sft.cern.ch/ Several versions of software may be distributed in parallel – user can choose version to run

Reference cloud: introduction Small CERNVM reference private cloud Condor batch system OpenNebula management Amazon EC2 interface Reference installation for other clouds Detailed/simple step-by-step instructions for replication at other sites will be provided Attempt to make “uniform” installations Site customisation possible for monitoring, etc.

Reference cloud: OpenNebula framework Popular framework for management of virtual machines Supports most common hypervisors Choice: KVM/QEMU – fast, does not require modifications to OS Amazon EC2 interface Possible to include VMs from other clouds, and provide hosts to other clouds Web management interface

Reference cloud: Amazon EC2 interface EC2 is the commercial cloud offered by Amazon EC2 also describes an interface for managing VMs Has become de-facto interface for all clouds, including private Hence, using the EC2 interface allows for great flexibility in launching VMs on both private and commercial clouds

Reference cloud: public vs. private clouds

Reference cloud: Elasticfox web user interface VM management through browser Can configure/start/stop VM instances, add/remove VM images Through Amazon EC2 interface Similar interface needed for data processing

NA49/NA61 processing: status CVMFS software installation Software for NA61 installed Issues with some set-up file options? Can also be used for processing on LXPLUS/ LXBATCH No need to adapt scripts (except for environment) NA49 software in progress Processing on CERNVM: Currently, reconstruction can be run “by hand” Batch system exists, scripts being adapted

NA49/NA61 processing: NA61 CVMFS installation Available under /cvmfs/na61.cern.ch/ On CERNVM virtual machines On “ordinary” computers having CVMFS installed, including LXPLUS/LXBATCH Script to set up environment: . /cvmfs/na61.cern.ch/library/etc/na61_env.sh

NA49/NA61 processing: next steps Two main tasks: Software validation CERNVM processing set-up Can be largely done in parallel Suggested steps on following slides Use LXBATCH for initial software validation Then validate on CERNVM Then set up production system

NA49/NA61 processing: next steps Step 1a: software validation on LXBATCH Select reference data set already processed by LXBATCH using software on AFS Reprocess the data on LXBATCH, but using software on CVMFS (instead of AFS) Compare output from CVMFS and AFS software Correct eventual problems Decouple issues related to CVMFS software installation from CERNVM set-up Ready to start this step now

NA49/NA61 processing: next steps Step 1b: CERNVM set-up, convert all processing scripts LXPLUS/LXBATCH uses PBS, CERNVM Condor Remove AFS references Castor Kerberos authentication (distribute kinit keytab file?) In progress, soon ready for reconstruction

NA49/NA61 processing: next steps Step 2a: software validation on LXBATCH Use CVMFS software for “normal” processing on LXBATCH (instead of AFS) Step 2b, CERNVM set-up: Select reference data set already processed by LXBATCH using software on AFS Reprocess the same data using CERNVM on test/ reference cloud, using software on CVMFS Compare output from CERNVM and LXBATCH Correct eventual problems CVMFS issues should already be found in step 1a

NA49/NA61 processing: next steps Step 3: production processing on LXCLOUD LXCLOUD is new cloud service offered by CERN IT Experience from reference cloud directly transferable Possible to set up processing facilities at other sites than CERN, based on reference cloud, if needed

NA49/NA61 processing: what is needed To successfully adapt processing to CERNVM, some input is needed Overview of (all) processing types performed on LXBATCH Scripts in use How to set-up/configure Also analysis (not only reconstruction)? Reference data sets to compare output Who is responsible for the various (software) components – who can/wants to participate? Carrying out the steps on the previous slides

NA49/NA61 processing: web user interface Web user interface for managing VM instances/ images exists Needed: processing-centred web user interface What data to process and what type of analysis List of available data and status, request processing Specify versions of software and VM Specify requirements for processing nodes Both VM and processing management in the same interface, or two separate interfaces? Generic experiment interface?

NA49/NA61 processing: web user interface “Step 4” CERNVM processing does not depend on it But will make the user experience much improved Considering to extend an existing tool for managing VMs to also manage data/processing

Status/outlook Reference cloud up and running NA61 software available on CVMFS NA49 software soon available on CVMFS Software validation ready to begin Data processing on CERNVM: Currently by hand, using batch soon Needed: better understanding of different processing task, who is responsible for what Needed: processing-centred web interface

Backup

Data preservation: motivation Preserve historic record Even after experiment end-of-life, data reprocessing might be desirable if future experiments reach incompatible results Many past experiments have already lost this possibility

Data preservation: challenges Two parts: data & software Data: preserve by migration to newer storage technologies Software: more complicated Just preserving source code/binaries not enough Strongly coupled to OS/library/compiler version (software environment) Software environment strongly coupled to hardware, platform will eventually become unavailable Porting to new platform requires big effort

Data preservation: solution Possible solution: virtualisation “Freeze” hardware in software Can run legacy analysis software on legacy versions of Linux they were originally developed for in VMs Software environment preserved, no need to modify code Comes for free if processing is already done on VMs

CERNVM: use cases Two main use cases: Computing centre Images for head and batch nodes Includes Condor batch system Personal computers Desktop (GUI) and basic (CL) images “Personal” use Code can be developed (desktop image) in similar environment/platform it will be processed (batch node image)

CERNVM: contextualisation All CERNVM instances initially identical Experiment specific software configuration/set- up introduced via contextualisation Two types CD-ROM image – mainly site specific configuration EC2 user data – mainly experiment specific Executed during start-up of VM

CVMFS: design Compressed files on HTTP server Downloaded, decompressed and cached locally on first use Possible to run software without Internet connection A hierarchy of standard HTTP proxy servers distribute the load Can also be used by non-VMs, e.g. LXPLUS/LXBATCH, other clusters, personal laptops

Reference cloud: virtual distributed Condor cluster Based on VMs in cloud Can be distributed over several sites Even if nodes are on different sites, they will appear to be in the same cluster A tier 1 can include VMs provided by tier 2s in its virtual Condor cluster Can be very work-saving, as the tier 2s do not need to set up job management themselves Other possibility: local CERNVM batch system run local jobs (like normal cluster)