GRIF Status Michel Jouvin LAL / IN2P3

Slides:



Advertisements
Similar presentations
LAL Site Report Michel Jouvin LAL / IN2P3
Advertisements

LAL Site Report Michel Jouvin LAL / IN2P3
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Overview of LCG-France Tier-2s and Tier-3s Frédérique Chollet (IN2P3-LAPP) on behalf of the LCG-France project and Tiers representatives CMS visit to Tier-1.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Storage Survey and Recent Acquisition at LAL Michel Jouvin LAL / IN2P3
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Quantitative Methodologies for the Scientific Computing: An Introductory Sketch Alberto Ciampa, INFN-Pisa Enrico Mazzoni, INFN-Pisa.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
Preparation of KIPT (Kharkov) computing facilities for CMS data analysis L. Levchuk Kharkov Institute of Physics and Technology (KIPT), Kharkov, Ukraine.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
LAL Site Report Michel Jouvin LAL / IN2P3
Grid Applications for High Energy Physics and Interoperability Dominique Boutigny CC-IN2P3 June 24, 2006 Centre de Calcul de l’IN2P3 et du DAPNIA.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
LAL Site Report Michel Jouvin LAL / IN2P3
LAL Site Report Michel Jouvin LAL / IN2P3
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
WLCG GridKa+T2s Workshop Site Report --- Presenter, Site, Country.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Site Report --- Andrzej Olszewski CYFRONET, Kraków, Poland WLCG GridKa+T2s Workshop.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
11 November 2010 Natascha Hörmann Computing at HEPHY Evaluation 2010.
Cluster Configuration Update Including LSF Status Thorsten Kleinwort for CERN IT/PDP-IS HEPiX I/2001 LAL Orsay Tuesday, December 08, 2015.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
LCG and Tier-1 Facilities Status ● LCG interoperability. ● Tier-1 facilities.. ● Observations. (Not guaranteed to be wry, witty or nonobvious.) Joseph.
23 January 2007WLCG workshop, CERN System Management Working Group Alessandra Forti WLCG workshop CERN, 23 January 2007.
Materials for Report about Computing Jiří Chudoba x.y.2006 Institute of Physics, Prague.
CC - IN2P3 Site Report Hepix Fall meeting 2010 – Ithaca (NY) November 1st 2010
IRFU SITE REPORT Pierrick Micout CEA/DSM/IRFU/SEDI.
Evangelos Markatos and Charalampos Gkikas FORTH-ICS Athens, th Mar Institute of Computer Science - FORTH Christos.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
II EGEE conference Den Haag November, ROC-CIC status in Italy
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
November 28, 2007 Dominique Boutigny – CC-IN2P3 CC-IN2P3 Update Status.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
The status of IHEP Beijing Site WLCG Asia-Pacific Workshop Yaodong CHENG IHEP, China 01 December 2006.
CC-IN2P3: A High Performance Data Center for Research Dominique Boutigny February 2011 Toward a future cooperation with Israel.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Alice Operations In France
SuperB – INFN-Bari Giacinto DONVITO.
A Dutch LHC Tier-1 Facility
Service Challenge 3 CERN
Data Challenge with the Grid in ATLAS
Luca dell’Agnello INFN-CNAF
Simulation use cases for T2 in ALICE
Pierre Girard ATLAS Visit
DAPNIA Site report.
GRIF : an EGEE site in Paris Region
The LHCb Computing Data Challenge DC06
Presentation transcript:

GRIF Status Michel Jouvin LAL / IN2P3

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 Objectives Build a Tier2 facility for simulation and analysis in Paris Region Build a Tier2 facility for simulation and analysis in Paris Region 80% LHC 4 experiments, 20% EGEE and local 80% LHC 4 experiments, 20% EGEE and local LHC : analysis (2/3) and MC simulation (1/3) LHC : analysis (2/3) and MC simulation (1/3) Be ready at LHC startup (2 nd half of 2007) Be ready at LHC startup (2 nd half of 2007) Resource goals(end of 2007) Resource goals(end of 2007) CPU : 1500 kSI2K (1kSI2K ~ P4 Xeon 2,8 Ghz) CPU : 1500 kSI2K (1kSI2K ~ P4 Xeon 2,8 Ghz) Storage : 350 TB of disks (disk only, no MSS) Storage : 350 TB of disks (disk only, no MSS) Network : 10 Gb/s backbone inside Tier2, 1 Gb/s external link Network : 10 Gb/s backbone inside Tier2, 1 Gb/s external link

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 Members Project started by DAPNIA (CEA), LAL (IN2P3, Orsay) and LPNHE (IN2P3, Paris), Fall 2004 Project started by DAPNIA (CEA), LAL (IN2P3, Orsay) and LPNHE (IN2P3, Paris), Fall 2004 DAPNIA and LAL involved in Grid effort since beginning of EDG DAPNIA and LAL involved in Grid effort since beginning of EDG 3 EGEE contracts (2 for operation support) 3 EGEE contracts (2 for operation support) No lab big enough to run a T2 by itself No lab big enough to run a T2 by itself LLR (IN2P3, Palaiseau) and IPNO (IN2P3, Orsay) joined the project in Sept. 05 LLR (IN2P3, Palaiseau) and IPNO (IN2P3, Orsay) joined the project in Sept. 05 IPNO : nuclear physics (Alice + Agatha) IPNO : nuclear physics (Alice + Agatha) LLR : CMS LLR : CMS

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 Organization 1 EGEE/LCG site, distributed over all labs 1 EGEE/LCG site, distributed over all labs Computing and storage resources in each lab Computing and storage resources in each lab Computing rooms and financing Computing rooms and financing IPNO wil concentrate on non LHC resources funding IPNO wil concentrate on non LHC resources funding 1 Gb/s link for IPNO, LAL, LPNHE, “soon” for DAPNIA 1 Gb/s link for IPNO, LAL, LPNHE, “soon” for DAPNIA Technical Committee : people from every lab Technical Committee : people from every lab 5 FTE in 2005, 6-7 in 2006, more in FTE in 2005, 6-7 in 2006, more in 2007 Currently people involved (several part time) Currently people involved (several part time) M. Jouvin (chairman), P. Micout, P.F. Honoré… M. Jouvin (chairman), P. Micout, P.F. Honoré… Scientific Committee (fund raising) Scientific Committee (fund raising) J.P. Meyer (DAPNIA/Atlas, chairman), 1 person / lab J.P. Meyer (DAPNIA/Atlas, chairman), 1 person / lab

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 Finances Total budget estimated to 1,6 M€ ( ) Total budget estimated to 1,6 M€ ( ) 30% from Region council 30% from Region council 30% from National Research Agency (ANR) 30% from National Research Agency (ANR) 40% from the labs (CEA, CNRS, Paris6 university) 40% from the labs (CEA, CNRS, Paris6 university) No significant support from IN2P3 / LCG France (focused on T1) No significant support from IN2P3 / LCG France (focused on T1) ½ budget still uncertain… First answers soon… ½ budget still uncertain… First answers soon… Progressive investment : no HW replacement before 2009 Progressive investment : no HW replacement before : 150 K€, 2006 : 450 K€, 2007 : 1 M€ 2005 : 150 K€, 2006 : 450 K€, 2007 : 1 M€ If necessary, could use 2008 to spread the effort If necessary, could use 2008 to spread the effort : 300 K€/year expected from IN2P3/LCG France : 300 K€/year expected from IN2P3/LCG France

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 Current Status EGEE/LCG GRIF site created EGEE/LCG GRIF site created IN2P3-LAL decommissionned, resources moved to GRIF IN2P3-LAL decommissionned, resources moved to GRIF 2 sites with resources, 2 sites ordering 2 sites with resources, 2 sites ordering DAPNIA : 20 WNs CPUs, 12 TB, installation in progress DAPNIA : 20 WNs CPUs, 12 TB, installation in progress LAL : 26 WNs CPUs, 8 TB (SRM/DPM), LCG services LAL : 26 WNs CPUs, 8 TB (SRM/DPM), LCG services 4,5 TB on order 4,5 TB on order LPNHE : 15 WNs CPUs, 5 TB ordered soon LPNHE : 15 WNs CPUs, 5 TB ordered soon IPNO : 20 WN CPUs (dual core blades) IPNO : 20 WN CPUs (dual core blades) End of 2005 : 80 WNs CPUs, 25 TB End of 2005 : 80 WNs CPUs, 25 TB Separate CE/SE on each site Separate CE/SE on each site

12/10/2005GRIF Tier2 - HEPix - SLAC Main Activities… Setup of resources on each site Setup of resources on each site Global configuration consistency : Quattor choosen Global configuration consistency : Quattor choosen Flexible site customization inside a unique database Flexible site customization inside a unique database Setup of a multi-site technical team Setup of a multi-site technical team Tutorials for new sites administrators Tutorials for new sites administrators Sharing management load (ex : middleware upgrade) Sharing management load (ex : middleware upgrade) Write documentation for sharing information and expertise (Trac) Write documentation for sharing information and expertise (Trac)

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 … 2005 Main Activites Evaluate DPM as a storage solution Evaluate DPM as a storage solution Successful so far, easy to setup and manage Successful so far, easy to setup and manage Quattor component written to manage DPM configuration Quattor component written to manage DPM configuration Plan to evaluate a multi-site configuration Plan to evaluate a multi-site configuration Disk servers on several sites Disk servers on several sites Current lack of srmcp is a problem with CMS/Phedex Current lack of srmcp is a problem with CMS/Phedex Participation to LCG SC3 Participation to LCG SC3 Throughput phase : 35 MB/s sustained 4 days Throughput phase : 35 MB/s sustained 4 days Plan to join service phase mid-november Plan to join service phase mid-november

12/10/2005GRIF Tier2 - HEPix - SLAC : Mini Tier2 Main goal : setup 20+% of final configuration Main goal : setup 20+% of final configuration 300 WNs CPUs, 70 TB 300 WNs CPUs, 70 TB Exact size wil depend on fund rising success… Exact size wil depend on fund rising success… Focus Focus Muti-site or mono-site CE/SE resources Muti-site or mono-site CE/SE resources Final choice for batch scheduler : evaluation of LSF and SGE Final choice for batch scheduler : evaluation of LSF and SGE Final choice for SE architecture (DPM only, DPM + LUSTRE) Final choice for SE architecture (DPM only, DPM + LUSTRE) Setup of monitoring tools : Nagios ?, Lemon ?, others ? Setup of monitoring tools : Nagios ?, Lemon ?, others ? Integration with local operations on each site Integration with local operations on each site Miscellanous Miscellanous Continue active participitation to SC Continue active participitation to SC Evaluation of 10 Gb/s link feasibality and effectiveness Evaluation of 10 Gb/s link feasibality and effectiveness Computer rooms requirements (electrical power, air cooling…) Computer rooms requirements (electrical power, air cooling…)

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 Storage Challenge Efficient use and management of a large amount of storage seen as the main challenge Efficient use and management of a large amount of storage seen as the main challenge Access to data from CPUs, no staging Access to data from CPUs, no staging Decided to start partnership with HP on LUSTRE in the Grid (LCG) context Decided to start partnership with HP on LUSTRE in the Grid (LCG) context Performance with a large number of clients Performance with a large number of clients Geographically distributed LUSTRE configuration Geographically distributed LUSTRE configuration Replication of critical datas (metadatas) among sites Replication of critical datas (metadatas) among sites SRM and/or xrootd integration SRM and/or xrootd integration Funds requested to ANR, answer soon… Funds requested to ANR, answer soon… Uncertainty with HP troubles in France… Uncertainty with HP troubles in France…

12/10/2005GRIF Tier2 - HEPix - SLAC 2005 Batch Scheduler 1 unified T2 means 1 batch scheduler 1 unified T2 means 1 batch scheduler Required for a coherent view/publishing of resources Required for a coherent view/publishing of resources Main requirements Main requirements Efficient use of distributed resources Efficient use of distributed resources Handle running jobs, 10Kjobs in queues Handle running jobs, 10Kjobs in queues Torque may not be appropriate Torque may not be appropriate Scalability and rosbustness, lack of dynamic reconfiguration Scalability and rosbustness, lack of dynamic reconfiguration Looking at LSF Looking at LSF LAL has experience for its internal use (and contacts…) LAL has experience for its internal use (and contacts…) Multicluster may offer the flexibility for global unified resource but maintaining some job/resources affinity at each site Multicluster may offer the flexibility for global unified resource but maintaining some job/resources affinity at each site Evaluation to start soon : 1 cluster+CE per site + cross submission Evaluation to start soon : 1 cluster+CE per site + cross submission Other candidates : SGE, Condor ? Other candidates : SGE, Condor ?