Cracow Grid Workshop 2004 - Grid Software Installation Tools

Slides:



Advertisements
Similar presentations
LCFG overview and future Michael George University of Liverpool
Advertisements

Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Torsten Antoni – LCG Operations Workshop, CERN 02-04/11/04 Global Grid User Support - GGUS -
Andrew McNab - Manchester HEP - 2 May 2002 Testbed and Authorisation EU DataGrid Testbed 1 Job Lifecycle Software releases Authorisation at your site Grid/Web.
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
INFN - Ferrara BaBarGrid Meeting SPGrid Efforts in Italy BaBar Collaboration Meeting - SLAC December 11, 2002 Enrica Antonioli - Paolo Veronesi.
27-29 September 2002CrossGrid Workshop LINZ1 USE CASES (Task 3.5 Test and Integration) Santiago González de la Hoz CrossGrid Workshop at Linz,
Torrent-based Software Distribution in ALICE.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
Andrew McNab - Manchester HEP - 22 April 2002 UK Rollout and Support Plan Aim of this talk is to the answer question “As a site admin, what are the steps.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) Overview of software tools for gLite installation & configuration.
Glite I/O Storm Testing in EDG-LCG Framework Elena Slabospitskaya, Vadim Petukhov, (IHEP, Russia) Gilbert Grosdidier, (CNRC, France) NEC'2005, Sept 16.
C. Loomis – Testbed Status – 28/01/2002 – n° 1 Future WP6 Tasks Charles Loomis January 28, 2002
INFSO-RI Enabling Grids for E-sciencE Status of LCG-2 porting Stephen Childs, Brian Coghlan and Eamonn Kenny Grid-Ireland/EGEE October.
1 Dynamic Application Installation (Case of CMS on OSG) Introduction CMS Software Installation Overview Software Installation Issues Validation Considerations.
J.T Moscicki CERN LCG - Software Process & Infrastructure1 SPI Software Process & Infrastructure for LCG Software Packaging and Distribution LCG Application.
Distribution After Release Tool Natalia Ratnikova.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
LCG Middleware Testing in 2005 and Future Plans E.Slabospitskaya, IHEP, Russia CERN-Russia Joint Working Group on LHC Computing March, 6, 2006.
WP8 Status – Stephen Burke – 30th January 2003 WP8 Status Stephen Burke (RAL) (with thanks to Frank Harris)
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
05/29/2002Flavia Donno, INFN-Pisa1 Packaging and distribution issues Flavia Donno, INFN-Pisa EDG/WP8 EDT/WP4 joint meeting, 29 May 2002.
EGEE is a project funded by the European Union under contract IST Experiment Software Installation in the LHC Computing Grid
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
Open Science Grid OSG CE Quick Install Guide Siddhartha E.S University of Florida.
EGEE is a project funded by the European Union under contract IST JRA1-SA1 requirement gathering Maite Barroso JRA1 Integration and Testing.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
Deployment work at CERN: installation and configuration tasks WP4 workshop Barcelona project conference 5/03 German Cancio CERN IT/FIO.
Ariel Garcia LCG cluster installation, EGEE training, Ariel Garcia - IWR LCG Cluster Installation Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft.
Feedback from the POOL Project User Feedback from the POOL Project Dirk Düllmann, LCG-POOL LCG Application Area Internal Review October 2003.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
DataGRID Testbed Enlargement EDG Retreat Chavannes, august 2002 Fabio HERNANDEZ
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
CERN IT Department t LHCb Software Distribution Roberto Santinelli CERN IT/GS.
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Tools and techniques for managing virtual machine images Andreas.
G.Govi CERN/IT-DB 1 September 26, 2003 POOL Integration, Testing and Release Procedure Integration  Packages structure  External dependencies  Configuration.
SAM Sensors & Tests Judit Novak CERN IT/GD SAM Review I. 21. May 2007, CERN.
VO Box Issues Summary of concerns expressed following publication of Jeff’s slides Ian Bird GDB, Bologna, 12 Oct 2005 (not necessarily the opinion of)
Ariel Garcia DataGrid WP6, Heidelberg, 26 th September 2003 Ariel García CrossGrid testbed status Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft.
EGEE-II INFSO-RI Enabling Grids for E-sciencE YAIM Overview MiMOS Grid tutorial HungChe, ASGC OPS Team.
INFSO-RI Enabling Grids for E-sciencE /10/20054th EGEE Conference - Pisa1 gLite Configuration and Deployment Models JRA1 Integration.
EGEE is a project funded by the European Union under contract IST Package Manager Predrag Buncic JRA1 ARDA 21/10/04
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
EGEE is a project funded by the European Union under contract IST Experiment Software Installation in LCG-2 Biomed Application.
Open Science Grid Build a Grid Session Siddhartha E.S University of Florida.
CernVM-FS Infrastructure for EGI VOs Catalin Condurache - STFC RAL Tier1 EGI Webinar, 5 September 2013.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
EGEE is a project funded by the European Union under contract IST Experiment Software Installation toolkit on LCG-2
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Mario Reale – GARR NetJobs: Network Monitoring Using Grid Jobs.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
EGEE is a project funded by the European Union under contract IST Experiment Software Installation in LCG-2 INFN-GRID/EGEE.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Maite Barroso – WP4 Workshop – 10/12/ n° 1 -WP4 Workshop- Developers’ Guide Maite Barroso 10/12/2002
CERN LCG1 to LCG2 Transition Markus Schulz LCG Workshop March 2004.
INFSO-RI Enabling Grids for E-sciencE Padova site report Massimo Sgaravatto On behalf of the JRA1 IT-CZ Padova group.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
Installation of the ALICE Software
Regional Operations Centres Core infrastructure Centres
lcg-infosites documentation (v2.1, LCG2.3.1) 10/03/05
First Latin-american Grid Workshop
The ATLAS software in the Grid Alessandro De Salvo <Alessandro
Torrent-based software distribution
Transactional Grid Deployment
The EU DataGrid Fabric Management Services
gLite The EGEE Middleware Distribution
Presentation transcript:

Cracow Grid Workshop Grid Software Installation Tools Forschungszentrum Karlsruhe GmbH, Germany Forschungszentrum Karlsruhe In der Helmholtz Gemeinschaft

Cracow Grid Workshop Introduction ● The issue of managing software packages is not new ● Many approaches exist: ● RPM (Redhat, SuSE, Mandrake,...), Portage (Gentoo), Ports collection (FreeBSD, NetBSD, OpenBSD,...), Pacman (NMI), dpkg/APT (Debian) ● These approaches (and usually grid middleware) often require root privileges for installation ● Applications can usually be installed by any User

Cracow Grid Workshop Use cases in Grid Environments ● Middleware Development: ● Install as root ● Homogenous Testbed (all sites at the same versions) ● Quick rollout procedure ● Central control ● HEP Experiments: ● Non root installation ● Large amounts of software (up to 6GB per Experiment) ● Installation per site ● Central control ● Generic Grid User: ● Independence of central instances ● Independence of organizational boundaries ● Portability (as few dependencies as possible)

Cracow Grid Workshop X# Deployment ● Design goals ● Focused on middleware installation ● Homogenous development environment ● Central control over software that is installed ● Defined deployment procedures ● Release Management ● definition ● rollback ● production TB – development TB ● Build on top of existing technologies

Cracow Grid Workshop X# Deployment ● How it works ● Packages are in RPM format ● Packages provided via autobuild ● Distributed via webserver ● Local ConFiGuration tool (LCFG) ● All sites use LCFG to install/configure ● Common part of testbed configuration in a common directory in CVS ● CVS branches: Production TB / Development TB ● Rollout: ● New version (CVS-tag) announced to site admins ● Site admins run update tools ● Install release ● Release version is published in Information Catalogue

Cracow Grid Workshop X# Deployment ● Pro's and Con's + centrally managable +/- LCFG + Rollout time est. 1d (15+ sites!) + Proven to work in CrossGrid - User, Developer depends on central instances - User must provide RPM packages => depends on 15 site admins to install it - Obey to strict policies

Cracow Grid Workshop LCG ESM Tools (LHC Experiment Software Management Tools) ● Designed to fit LHC Experiments needs: ● Experiment Software Managers (ESM) per VO ● Package/deplay/certify/support the Software ● Manage dependencies Local site admins ● Are independent from other experiments ● Installation per site ● Published installed versions in Information Catalogue ● Steer Job to software ● Central control over which software is installed

Cracow Grid Workshop LCG ESM Tools ● How it works ● Initial installation ● New VO per Experiment for software management ● only ESMs can write to VO directory ● For deployment of VO packages ● Installing a new Software package: ● ESM ● packages the software ● replicates install package to SEs ● steers job to site for ● installation/verification ● tagging a site as supporting the software in Info Catalogue ● User ● Specifies the required software in.jdl to steer job ● runs LCG tool to ensure SW installation on his WN ● runs his job

Cracow Grid Workshop LCG ESM Tools ● Pro's and Con's + Out in the wild and being used (est 70 sites 800 pkgs) Must scale well + centrally managable + User can steer job to software Useful for huge packages + Integrates well into the EDG MW +/- Installation tools depend on high level MW - User depends on ESM to support his software and version - ESM jobs conflict with user jobs - ESM is not a normal user

Cracow Grid Workshop Tank&Spark ● Designed to improve LCG-ESM shortcomings: ● Support VOMS ● Avoid NFS related problems ● Avoid ESM jobs in the queue with user jobs ● Automated update of the whole grid ● How it works? ● Tank daemon (webservice) is running on the CEs ● Sparks clients are running on the Wns ● Clients poll CEs for orders ● Clients rsync software from SE (put in place by the ESM, as before) ● Installation status is kept in MySQL-DB by Tank daemon ● Available software is published in Information Catalogue

Cracow Grid Workshop Tank&Spark ● Pro's and Con's + centrally Manageable + Scalability proven to 1000 nodes per site +/- Tool depends on high level MW ● Optimized for keeping network load low - User still depends on ● software and version specified by ESM / Experiment ● Newest EDG/LCG middleware ● To be deployed with LCG-2.4

Cracow Grid Workshop Alien/gLite packman ● Design goals ● Support the LHC Experiments ● Frequent releases ● Support individuals to install own software ● Steer job to site

Cracow Grid Workshop Alien/gLite packman ● How it works ● ESM Role present as well ● User can also do this ● Packaging / define dependencies in a metadata catalogue / write preinstall, prerun scripts ● One PackMan per site ● Shared filesystem holds installable packages (alien FC) ● Required packages are specified in.jdl ● WN triggers installation of required packages by contacting PackMan on CE ● PackMan ● installs in shared dir ● Returns a list of environment files for sourcing ● WN executes the job

Cracow Grid Workshop Alien/gLite packman ● Pro's and Con's + Aims at supporting individual users - Not yet deployed on a broad scale (=Not much known about it) -/+ User depends on a high level middleware - Depends on shared filesystem

Cracow Grid Workshop GAIT (Grid Application Installation Tool) ● Design goals ● Support individual users ● Ease of use ● Minimal dependencies ● On site admins ● On installed software ● On the underlying unix ● Don't waste resources ● Don't install twice ● Don't stay installed forever

Cracow Grid Workshop GAIT ● How it works ● Packager extends install_template.sh script to ● define an install function to software ● define a check function to verify software installation ● define an optional function run a testcase on the software ● This script ● Check in /opt, all homedirs, /tmp for “GAIT/ / ” ● If not found, try installation own homedir and /tmp ● If found it touches the installation for keeping it installed ● User ● ships install_*.sh with his job ● sources install_.sh, which ● makes sure the software is installed ● sets up environment variables properly ● runs his job ● Unused software can be expired by user or admin

Cracow Grid Workshop GAIT Example job.sh #deploy GAIT tar xzf GAIT.tgz #can b preinstalled source GAIT/scripts/install_[..].sh echo $GAIT_INSTALLDIR echo $PATH echo $LD_LIBRARY_PATH #run the software software job.jdl Executable = "job.sh"; StdOutput = "out"; StdError = "err"; InSandbox={"job.sh","GAIT.tgz"}; OutputSandbox = {"out","err"}; install_package sh source $INCLUDE_DIR/gait.sh.h function gait_install_package { SW_DIR=$1 cd $SW_DIR wget package tgz tar xzf package tgz } function gait_check_installation { SW_DIR=$1 test -e $SW_DIR/executable || -1 # TODO: check md5 hashs of files } function gait_run_testcase{optional} gait_run

Cracow Grid Workshop GAIT ● Pro's and Con's - Not many packages yet => Prototype for dynamic deployment of userspace software on the grid + No dependencies on ● Other Software (only plain unix-tools) ● Site admins ● Software Managers ● Not even a grid + Only little impact on user job + Easily interfaces to higher level tools

Cracow Grid Workshop Conclusions ● LCG-Experiment users are well supported ● MW deployment defined in X# Procedures ● Individual users are not well supported yet ● Three new tools are just available ● Packman ● Tank&Spark ● GAIT ● Check out available tools to simplify your deployment