LCG LCG-1 Deployment and usage experience Lev Shamardin SINP MSU, Moscow

Slides:



Advertisements
Similar presentations
LNL M.Biasotto, Roma, 22 novembre Some notes on WP6 INFN Testbed.
Advertisements

Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
Presenter Name Facility Name EDG Testbed Status Moving to Testbed Two.
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
NIKHEF Testbed 1 Plans for the coming three months.
McFarm: first attempt to build a practical, large scale distributed HEP computing cluster using Globus technology Anand Balasubramanian Karthik Gopalratnam.
DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
Institute for High Energy Physics ( ) NEC’2007 Varna, Bulgaria, September Activities of IHEP in LCG/EGEE.
New Cluster for Heidelberg TRD(?) group. New Cluster OS : Scientific Linux 3.06 (except for alice-n5) Batch processing system : pbs (any advantage rather.
Task 3.5 Tests and Integration ( Wp3 kick-off meeting, Poznan, 29 th -30 th January 2002 Santiago González de la.
Sergey Belov, LIT JINR 15 September, NEC’2011, Varna, Bulgaria.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
Sergey Belov, Tatiana Goloskokova, Vladimir Korenkov, Nikolay Kutovskiy, Danila Oleynik, Artem Petrosyan, Roman Semenov, Alexander Uzhinskiy LIT JINR The.
1 Deployment of an LCG Infrastructure in Australia How-To Setup the LCG Grid Middleware – A beginner's perspective Marco La Rosa
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
C. Loomis – Testbed Status – 28/01/2002 – n° 1 Future WP6 Tasks Charles Loomis January 28, 2002
Andrew McNab - Manchester HEP - 5 July 2001 WP6/Testbed Status Status by partner –CNRS, Czech R., INFN, NIKHEF, NorduGrid, LIP, Russia, UK Security Integration.
EDG LCFGng: concepts Fabric Management Tutorial - n° 2 LCFG (Local ConFiGuration system)  LCFG is originally developed by the.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
Partner Logo DataGRID WP4 - Fabric Management Status HEPiX 2002, Catania / IT, , Jan Iven Role and.
G RID M IDDLEWARE AND S ECURITY Suchandra Thapa Computation Institute University of Chicago.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
UCY HPCL Introduction to the CrossGrid Testbed George Tsouloupas UCY HPCL.
EU DataGrid segment in Russia. Testbed WP6. V.Ilyin 1, N. Kruglov 1, A. Kryukov 1, V. Korenkov 2, V. Kolosov 3, V. Mitsyn 2, L. Shamardin 1 1 SINP MSU.
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
InstantGrid: A Framework for On- Demand Grid Point Construction R.S.C. Ho, K.K. Yin, D.C.M. Lee, D.H.F. Hung, C.L. Wang, and F.C.M. Lau Dept. of Computer.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
Responsibilities of ROC and CIC in EGEE infrastructure A.Kryukov, SINP MSU, CIC Manager Yu.Lazin, IHEP, ROC Manager
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
Institute of High Energy Physics ( ) NEC’2005 Varna, Bulgaria, September Participation of IHEP in EGEE.
TERENA 2003, May 21, Zagreb TERENA Networking Conference, 2003 MOBILE WORK ENVIRONMENT FOR GRID USERS. TESTBED Miroslaw Kupczyk Rafal.
Development of Russian Grid Segment in the frames of EU DataGRID, LCG and EGEE projects V.A.Ilyin (SINP MSU), V.V.Korenkov (JINR, Dubna) NEC’2003, Varna.
INFSO-RI Enabling Grids for E-sciencE OSG-LCG Interoperability Activity Author: Laurence Field (CERN)
V.Ilyin, V.Gavrilov, O.Kodolova, V.Korenkov, E.Tikhonenko Meeting of Russia-CERN JWG on LHC computing CERN, March 14, 2007 RDMS CMS Computing.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Portal Update Plan Ashok Adiga (512)
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Construction methods and monitoring in meta-cluster systems LIT, JINR Korenkov V.V, Mitsyn V.V, Chkhaberidze D.V, Belyakov D.V.
Maite Barroso - 10/05/01 - n° 1 WP4 PM9 Deliverable Presentation: Interim Installation System Configuration Management Prototype
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Grid2Win : gLite for Microsoft Windows Roberto.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
Participation of JINR in CERN- INTAS project ( ) Korenkov V., Mitcin V., Nikonov E., Oleynik D., Pose V., Tikhonenko E. 19 march 2004.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
CERN 19/06/2002 Kickstart file generator Andrea Chierici (INFN-CNAF) Enrico Ferro (INFN-LNL) Marco Serra (INFN-Roma)
Tests at Saclay D. Calvet, A. Formica, Z. Georgette, I. Mandjavidze, P. Micout DAPNIA/SEDI, CEA Saclay Gif-sur-Yvette Cedex.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
Stephen Burke – Sysman meeting - 22/4/2002 Partner Logo The Testbed – A User View Stephen Burke, PPARC/RAL.
EGEE is a project funded by the European Union under contract IST Issues from current Experience SA1 Feedback to JRA1 A. Pacheco PIC Barcelona.
II EGEE conference Den Haag November, ROC-CIC status in Italy
Consorzio COMETA - Progetto PI2S2 UNIONE EUROPEA Grid2Win : gLite for Microsoft Windows Elisa Ingrà - INFN.
G. Russo, D. Del Prete, S. Pardi Kick Off Meeting - Isola d'Elba, 2011 May 29th–June 01th A proposal for distributed computing monitoring for SuperB G.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
The Beijing Tier 2: status and plans
The EU DataGrid Fabric Management Services
Presentation transcript:

LCG LCG-1 Deployment and usage experience Lev Shamardin SINP MSU, Moscow

LCG Lev Shamardin LCG Geography LCG covers a number of sites both in Europe and US LCG covers a number of sites both in Europe and US LCG-1 testbed covers 13 sites including SINP MSU LCG-1 testbed covers 13 sites including SINP MSU

LCG Lev Shamardin LCG-1 LCG-1 is the production version of the LCG software which is now installed on the 13 LCG-1 sites and running in production mode LCG-1 is the production version of the LCG software which is now installed on the 13 LCG-1 sites and running in production mode New LCG-2 is expected in production in November New LCG-2 is expected in production in November

LCG Lev Shamardin LCG-1 Architecture Minimal LCG-1 site must have a Computing Element with worker nodes and a Storage Element Minimal LCG-1 site must have a Computing Element with worker nodes and a Storage Element There are no limitations to the number of other components installed on each site There are no limitations to the number of other components installed on each site These components do not have to be registered in LCG-1 These components do not have to be registered in LCG-1 LCG-1 site MDS west MDS east RB LCG-1 site

LCG Lev Shamardin LCG-1 Architecture Current version LCG-1 is based on the EDG middleware with MDS information system Current version LCG-1 is based on the EDG middleware with MDS information system R-GMA is not yet stable enough for the production usage R-GMA is not yet stable enough for the production usage LCG-2 will be using new R-GMA information system LCG-2 will be using new R-GMA information system

LCG Lev Shamardin LCFGng and LCG-1 deployment The base middleware LCFGng configuration profiles are stored in the central LCG CVS repository The base middleware LCFGng configuration profiles are stored in the central LCG CVS repository Sites create their own site-specific profiles based on the CVS configuration Sites create their own site-specific profiles based on the CVS configuration No manual installation supported yet No manual installation supported yet LCG CVS Generic LCG-1 configuration LCG-1 Site Site-specific configuration LCG-1 SiteLCG-1 Site… LCG-1 Deployment group

LCG Lev Shamardin SINP MSU PC Farm 20 dual-CPU PIII nodes 20 dual-CPU PIII nodes Two 1.2 TB fileservers Two 1.2 TB fileservers Gigabit Ethernet uplinks to the fileservers Gigabit Ethernet uplinks to the fileservers Fast Ethernet links to the nodes Fast Ethernet links to the nodes Storage switch non-interactive nodes PC Farm batch master node

LCG Lev Shamardin Software installation system on SINP MSU PC Farm Based on Etherboot network boot package Based on Etherboot network boot package Nodes are installed with anaconda kickstart Nodes are installed with anaconda kickstart Nodes without a boot ROM are installed using a fake linux “kernel” or a boot disk Nodes without a boot ROM are installed using a fake linux “kernel” or a boot disk Supports both completely unattended automatic installation and manual installation Supports both completely unattended automatic installation and manual installation DHCP TFTP configuration server NFS node file server

LCG Lev Shamardin LCFGng support on SINP MSU PC Farm LCFGng server was installed for configuring and installing LCG-1 nodes LCFGng server was installed for configuring and installing LCG-1 nodes LCFGng enabled nodes configuration is controlled from the LCFGng server LCFGng enabled nodes configuration is controlled from the LCFGng server Nodes configuration and health status information provided with LCFGng components can be observed on the web Nodes configuration and health status information provided with LCFGng components can be observed on the web DHCP TFTP configuration server NFS node file server LCFGng server LCFG, HTTP

LCG Lev Shamardin Current LCG-1 status in SINP MSU For the LCG-1 testbed part of the PC Farm was configured as LCG-1 nodes For the LCG-1 testbed part of the PC Farm was configured as LCG-1 nodes These nodes are logically disconnected from the main farm These nodes are logically disconnected from the main farm Installed LCG-1 components are: Computing Element with several Worker Nodes, Storage Element, Resource Broker, BDII and User Interface Installed LCG-1 components are: Computing Element with several Worker Nodes, Storage Element, Resource Broker, BDII and User Interface MyProxy and MDS servers are coming soon MyProxy and MDS servers are coming soon Storage switch non-interactive nodes PC Farm batch master node LCG-1

LCG Lev Shamardin MSU PC Clusters LCG middleware installation on the SRCC MSU (Scientific Research Computing Center) parallel cluster LCG middleware installation on the SRCC MSU (Scientific Research Computing Center) parallel cluster Site specifics: Site specifics: Exotic batch system. Interface for the globus package is ready but was not completely tested Exotic batch system. Interface for the globus package is ready but was not completely tested Only manual node configuration can be done Only manual node configuration can be done Middleware for the worker nodes must be installed on the shared filesystem Middleware for the worker nodes must be installed on the shared filesystem Connect the Physical faculty and the faculty of Computing Mathematics and Cybernetics clusters to the SINP MSU Resource Broker Connect the Physical faculty and the faculty of Computing Mathematics and Cybernetics clusters to the SINP MSU Resource Broker Experience with SRCC is required, similar site-specific limitations Experience with SRCC is required, similar site-specific limitations

LCG Lev Shamardin Conclusion What difficulties will emerge if a site wants to install the LCG middleware What difficulties will emerge if a site wants to install the LCG middleware What LCG can give you right now What LCG can give you right now Future plans Future plans

LCG Lev Shamardin Difficulties At the moment the only documented way to install LCG software requires using the LCFGng configuration server At the moment the only documented way to install LCG software requires using the LCFGng configuration server Using LCFGng is not possible in a number of cases, in general due to the administrative reasons Using LCFGng is not possible in a number of cases, in general due to the administrative reasons Minimal stand-alone site must be running quite a big number of nodes supporting the infrastructure Minimal stand-alone site must be running quite a big number of nodes supporting the infrastructure The minimal set is: The minimal set is: Resource Broker Resource Broker BDII & MDS BDII & MDS Computing Element with Worker Nodes Computing Element with Worker Nodes Storage Element(s) Storage Element(s) MyProxy if long-time jobs support is required MyProxy if long-time jobs support is required This gives us at least 5 nodes will be supporting the infrastructure This gives us at least 5 nodes will be supporting the infrastructure

LCG Lev Shamardin What LCG-1 can give you right now Convenient way for job balancing between several sites Convenient way for job balancing between several sites Common way of user authentication and authorization for job submission Common way of user authentication and authorization for job submission Some basic accounting Some basic accounting Data replication Data replication

LCG Lev Shamardin Future plans SINP MSU Site is one of the sites participating in the LCG-1 SINP MSU Site is one of the sites participating in the LCG-1 New Russian sites will be connected to the SINP MSU Resource Broker in the nearest future New Russian sites will be connected to the SINP MSU Resource Broker in the nearest future IHEP Protvino will be connected soon IHEP Protvino will be connected soon ITEP and JINR Dubna in future ITEP and JINR Dubna in future Internal MSU sites (SRCC and others) as soon as manual installation is possible Internal MSU sites (SRCC and others) as soon as manual installation is possible