20.10.20111 Mihnea Dulea, IFIN-HH Efficient Handling and Processing of PetaByte-Scale Data for the Grid Centers within the FR Cloud 1ST JOINT SYMPOSIUM.

Slides:



Advertisements
Similar presentations
National Grid's Contribution to LHCb IFIN-HH Serban Constantinescu, Ciubancan Mihai, Teodor Ivanoaica.
Advertisements

1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
IFIN-HH LHCB GRID Activities Eduard Pauna Radu Stoica.
GRIF Status Michel Jouvin LAL / IN2P3
Overview of LCG-France Tier-2s and Tier-3s Frédérique Chollet (IN2P3-LAPP) on behalf of the LCG-France project and Tiers representatives CMS visit to Tier-1.
Bondyakov A.S. Institute of Physics of ANAS, Azerbaijan JINR, Dubna.
11 September 2007Milos Lokajicek Institute of Physics AS CR Prague Status of the GRID in the Czech Republic NEC’2007.
Techniques for establishing and maintaining constant temperature in ICT systems in order to reduce energy consumption Mihail Radu Cătălin Truşcă, Ş. Albert,
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Green technology used for ATLAS processing Dr. Ing. Fărcaş Felix NATIONAL INSTITUTE FOR RESEARCH AND DEVELOPMENT OF ISOTOPIC AND MOLECULAR.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
E-Infrastructure hierarchy Networking and Computational facilities in Armenia ASNET AM Network Armenian National Grid Initiative Armenian ATLAS site (AM-04-YERPHI)
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Current Status of the Grid Computing for Physics in Romania Horia Hulubei National Institute of R&D in Physics and Nuclear Engineering (IFIN-HH)
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
Enabling Grids for E-sciencE EGEE III Security Training and Dissemination Mingchao Ma, STFC – RAL, UK OSCT Barcelona 2009.
Monitoring the Grid at local, national, and Global levels Pete Gronbech GridPP Project Manager ACAT - Brunel Sept 2011.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Site Report BEIJING-LCG2 Wenjing Wu (IHEP) 2010/11/21.
Enabling Grids for E-sciencE 1 EGEE III Project Prof.dr. Doina Banciu ICI Bucharest GRID activities in RoGrid Consortium.
Bob Jones Technical Director CERN - August 2003 EGEE is proposed as a project to be funded by the European Union under contract IST
Company LOGO “ALEXANDRU IOAN CUZA” UNIVERSITY OF IAŞI” Digital Communications Department Status of RO-16-UAIC Grid site in 2013 System manager: Pînzaru.
Analysis in STEP09 at TOKYO Hiroyuki Matsunaga University of Tokyo WLCG STEP'09 Post-Mortem Workshop.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
A. Hoummada May Korea Moroccan ATLAS GRID MAGRID Abdeslam Hoummada University HASSAN II Ain Chock B.P Maarif CASABLANCA - MOROCCO National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
11 November 2010 Natascha Hörmann Computing at HEPHY Evaluation 2010.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
Status Organization Overview of Program of Work Education, Training It’s the People who make it happen & make it Work.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
Materials for Report about Computing Jiří Chudoba x.y.2006 Institute of Physics, Prague.
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
CC - IN2P3 Site Report Hepix Fall meeting 2010 – Ithaca (NY) November 1st 2010
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks SA2 Networking support for EGEE III Xavier.
Connect communicate collaborate LHCONE European design & implementation Roberto Sabatino, DANTE LHCONE Meeting, Washington, June
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
II EGEE conference Den Haag November, ROC-CIC status in Italy
Update on CHEP from the Computing Speaker Committee G. Carlino (INFN Napoli) on behalf of the CSC ICB, October
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
Mihnea Dulea, IFIN-HH R-ECFA Meeting, National Physics Library IFIN-HH, Magurele Romanian participation in WLCG M. Dulea Elementary Particles.
The status of IHEP Beijing Site WLCG Asia-Pacific Workshop Yaodong CHENG IHEP, China 01 December 2006.
RECENT DEVELOPMENTS IN THE CONTRIBUTION OF DFCTI/IFIN-HH TO THE WLCG COLLABORATION Department of Computational Physics and Information Technologies (DFCTI)
EGI… …is a Federation of over 300 computing and data centres spread across 56 countries in Europe and worldwide …delivers advanced computing.
Bob Jones EGEE Technical Director
Accessing the VI-SEEM infrastructure
Grid Computing: Running your Jobs around the World
GRID OPERATIONS IN ROMANIA
Regional Operations Centres Core infrastructure Centres
Grid site as a tool for data processing and data analysis
Christos Markou Institute of Nuclear Physics NCSR ‘Demokritos’
High Energy Physics Computing Coordination in Pakistan
Romanian Sites Current Status
GRIF : an EGEE site in Paris Region
Presentation transcript:

Mihnea Dulea, IFIN-HH Efficient Handling and Processing of PetaByte-Scale Data for the Grid Centers within the FR Cloud 1ST JOINT SYMPOSIUM CEA-IFA HaPPSDaG - PROJECT PRESENTATION - - FIRST YEAR PROGRESS REPORT - M. Dulea National Institute for Nuclear Physics and Engineering 'Horia Hulubei' (IFIN-HH)

Mihnea Dulea, IFIN-HH  Computing support for LHC  Project topics  Project objectives and work planning  Framework agreements  General information  Project teams and infrastructure  First year results OVERVIEW

Mihnea Dulea, IFIN-HH COMPUTING SUPPORT for LHC - LCG LHC COMPUTING GRID LCG is a wide distributed array of computing resources that provides the computing support required for the storage, processing, simulation and analysis of the data gathered by the four major experiments performed at LHC. It consists of more than 140 computing centres and federations of centres from 35 countries. The resource centres are classified according to their size and functionality as Tier-0 CERN), Tier-1 (11 centres), and Tier-2. The centres are interconnected through a high-speed network (GEANT2 in EU). Current and activity related to LHC.

Mihnea Dulea, IFIN-HH COMPUTING SUPPORT - FR ATLAS FRENCH CLOUD Grid sites: CC-IN2P3 (Tier-1) Tier-2 centres:... (many) GRIF Grille de Recherche d'Ile de France computing grid in Paris region, joint initiative of CEA/IRFU + labs of CNRS/IN2P3 (6 sites) The sites are interconnected through dedicated 10 Gbps links; connected to the FR NREN: RENATER: Réseau national de télécommunications pour la technologie, l'enseignement et la recherche FR Cloud includes foreign grid centres from China, Japan, Romania

Mihnea Dulea, IFIN-HH COMPUTING SUPPORT - RO ROMANIAN TIER-2 FEDERATION RO-LCG Grid sites: IFIN-HH, 5 Grid sites (resource centres) ISS - Inst. for Space Sciences (2 sites) UPB - Univ. 'Politehnica' of Bucharest ITIM - NIRD in Molecular & Isotopic Technologies - Cluj UAIC, Alex. Ioan Cuza University - Iasi The sites are connected to the 10 Gbps backbone of the RO NREN - the Romanian Educational and Research Network RoEduNet 4 grid sites currently support ATLAS vo: RO-07-NIPNE, RO-02-NIPNE (IFIN-HH); RO-14-ITIM (Cluj), RO-16-UAIC (Iasi)

Mihnea Dulea, IFIN-HH PROJECT TOPIC Computing support for LHC experiments = provision of grid resources + services The overall support of LCG deployment and operation is provided from other funds (e.g. CONDEGRID project in RO). HAPPSDAG addressess specific ATLAS issues in order to optimize resource usage

Mihnea Dulea, IFIN-HH ATLAS ISSUES Generic requirements regarding - data transfer from Tier-1 to the associated Tier-2 sites (CC-IN2P3 => RO-LCG) - transfer of large files from SE to WN for each analysis job; consider many simultaneous jobs - transfer of log and results files from WN to SE; immediate transfer of log file to UI RO specific needs at the beginning of the project Grid cluster - analysis of the causes of the lower performance of RO-LCG sites before Oct elaborate and test technical solutions for performance improvement - ensure better communication and coordination between the RO sites and the FR-cloud partners - general measures for improving Tier1 - Tier2 interaction - elaborate general guidelines regarding the improvement in efficiency of the grid centers which are associated to ATLAS clouds Transfer paths from/to the Storage Element (SE)

Mihnea Dulea, IFIN-HH  Improve communication and coordination between GRIF/IN2P3 and RO sites (RO/FR)  Testing & improving quality of the FR - RO data link for large dataset transfers (RO/FR)  Implementation of specific measures for increasing ATLAS job load and storage performance on sites (RO)  Improving large dataset transfer between FR - RO and data analysis (RO/FR)  Contributing to grid monitoring and technical support within FR-cloud (RO)  Training regarding grid monitoring and support (FR => RO)  Dissemination (RO/FR) Strategic objective: provide means for improvement of the processing and handling of large data sets at the Tier2 centers which participate in the ATLAS experiment at the LHC computing support. (RO - case study) Specific objectives and partner contributions: PROJECT OBJECTIVES

Mihnea Dulea, IFIN-HH PLANNING of WORK  Stage 1 ( ) Analysis of Tier1-Tier2 communication  Stage 2 ( ) Studies and software tools for monitoring and operation of the FR Cloud - RO grid connection and job loading. Testing of data handling and processing.  Stage 3 ( ) Methods and procedures for improving the performance of the RO sites within the FR Cloud

Mihnea Dulea, IFIN-HH  General Cooperation Agreement for Scientific Research between CEA and IFA, signed in December Field of cooperation: Technologies for Information and Health - Topic proposed for 2010: Grid Technologies  Joint Call for proposals of joint R&D projects (May 2010) - IFIN-HH and IRFU submitted a proposal for a Joint Research and Development Projects  Cooperation Agreement in the Field of Scientific Research (AS) between CEA and IFIN-HH, ( ) - General Coordinators: Gerard Cognet (FR), Ioan Ursu (RO) - leading and coordinating the cooperation activities  Project Agreement (CEA, IFIN-HH) FRAMEWORK AGREEMENTS

Mihnea Dulea, IFIN-HH  RO Contract n° C1-06/2010, between IFA and IFIN-HH  Start date: 01/10/2010  Duration: 24 months  Funding of the RO part of the project: lei (~ €)  Funding of the FR part of the project: € GENERAL INFORMATION BUDGET RO (lei)CEA (Eur)RO (lei)CEA (Eur)RO (lei)CEA (Eur) Manpower Travels Others (Romanian Engineer staying at Saclay ) Others (French guests staying in Romania ) Others (equipment) Others (indirect costs) Total:

Mihnea Dulea, IFIN-HH PROJECT TEAMS Project coordinators: Jean-Pierre Meyer (FR), Mihnea Dulea (RO) Technical correspondents: Pierrick Micout (FR), Gabriel Stoicea (RO) FR team (CEA/IRFU) Eric LANÇON Pierrick MICOUT Christine LEROY Frédéric SCHAER Zoulikha GEORGETTE Adelino GOMEZ RO team (IFIN-HH) Serban Constantinescu Mihai Ciubancan Ionut Traian Vasile Camelia Mihaela Visan

CTI/DPETI 1200 (grid) (hpc) cores, 270 TB Centre for Informational Technologies (CTI) - IFIN-HH

Mihnea Dulea, IFIN-HH ANALYSIS of NETWORK INFRASTRUCTURE Objective: identify the weak points of the FR-RO data connection and adoption of measures for improving the transfer capacity of large datasets. Network structure: complex, various owners and administrators => more difficult to act Activities (RO+FR)  Testing connectivity & transport capacity with various tools  Finding routing paths and points of data traffic delay  Comparing performances of RO-CERN link with those of RO-IN2P3 SectionCentresAdministratorOwnerLocation IFIN-HH LANRO-02-NIPNE RO-07-NIPNE CTI/DPETIIFIN-HHMagurele IFIN - UPBUPBICOMMIFIN-HHUPB RoEduNetRO-14-ITIM RO-16-UAIC AARNIECMECTSRomania GEANT2In 34 EU states DANTEEU NRENsEUEU RENATERGRIF, IN2P3GIP RENATER France Conclusions: a) performance degradation at RoEduNet / GEANT2 interface b) bottlenecks on some of the RoEduNet routers

Mihnea Dulea, IFIN-HH IMPROVING POINT-TO-POINT TRAFFIC PERFORMANCES Requires close collaboration with network administrators along the RO-FR path Example: following bandwidth capacity and traffic analysis, a RoEduNet router was found, responsible of bottlneck. AARNIEC's intervention rised the available bandwidth to 700 Mbps (fig. below). Permanent monitoring required

Mihnea Dulea, IFIN-HH MONITORING TOOLS for DATA TRANSFER and STORAGE PERFORMANCE - 1 Development of software tools for monitoring of SE traffic (in/out) (adding data sent by daemons running on storage servers in a database + web interface for display) Tools developed in IFIN-HH; useful for FR partners too for monitoring RO sites.  Traffic from/to WNs and from/to external network Max at 5 Gbps Max at > 3 Gbps

Mihnea Dulea, IFIN-HH MONITORING TOOLS for DATA TRANSFER and STORAGE PERFORMANCE - 2  Traffic on gateway (in/out); SE extern throughput  Monitoring groups of running or pending jobs

Mihnea Dulea, IFIN-HH MONITORING TOOLS for DATA TRANSFER and STORAGE PERFORMANCE - 3  Accounting of running or pending jobs on CE or CREAM-CE

Mihnea Dulea, IFIN-HH IMPROVEMENT of SITE MONITORING and TECHNICAL SUPPORT Implementation of its own SAM (Service Availability Monitoring) system, that uses IFIN-HH grid infrastructure and a new monitoring vo - ifops. Results published using Nagios. Early notification of technical staff leads to improvement of availability of grid services Monitoring of CREAM-CE, tbit03.nipne.ro

Mihnea Dulea, IFIN-HH IMPROVEMENT and TESTS of SE-WN THROUGHPUT Adding more resources (WNs) doesn't always mean better results. Scalability is required Improvement of file transfer speed from SE to WN, required by analysis jobs (4-6 files 2-4 GB) Replacing the transfer to disk servers through Network File System (NFS) protocole by new DPM (Disk Pool Manager) disk storage servers. Higher transfer speed => no job exceeds the time limit => no cancellation Tests of the new configuration Time representation of the transfer speed (in Mbps) for 70 quasi-simultaneous jobs

Mihnea Dulea, IFIN-HH GLOBAL IMPROVEMENT of EFFICIENCY Mean efficiency of ATLAS job execution in 2011: 91% Monthly number of ATLAS jobs and number of ATLAS events processed in RO-LCG

Mihnea Dulea, IFIN-HH TRAINING REGARDING MONITORING AND TECHNICAL SUPPORT : training stage of C. Visan at CEA/IRFU, preparing later participation to monitoring and support activities for FR Cloud sites. Topics: - CEA/IRFU monitoring methods at site, VO, project levels; EGI/WLCG and LHC monitoring (Christine Leroy, Pierrick Micout ) - grid site usage (Georgette Zoulikha) - NAGIOS installing/configuration on virtual machines (Frederic Schaer) - job submission through Pathena (PanDA Athena), at LAL-Orsay (Laurent Duflot) - CACTI site monitoring (Victor Mendoza, Université Pierre et Marie Curie (UPMC)) - instructions for site and job monitoring in ADCoS (ATLAS Distributed Computing Operations Shift) and for support team of FR Cloud (Squad). (Sabine Crepe)

Mihnea Dulea, IFIN-HH MOBILITY  Kick-off meeting ( , Saclay)  Participation at the RO-LCG 2010 Conference, Bucharest (Christine Leroy, Sabine Crepe - IN2P3)  Participation of Gabriel Stoicea to the spring meeting of LCG-France ( )  Training - monitoring and support ( , Saclay), C.M. Visan

Mihnea Dulea, IFIN-HH BENEFITS CEA/IRFU  The results of the project contribute to global improvement of FR Cloud efficiency  Elaboration, in collaboration, of general guidelines for interaction between grid centres in ATLAS clouds, and  Using FR-RO interaction as a representative case study for sharing best practices with smaller sites IFIN-HH  General efficiency improvement of the activity of the RO sites  Better integration and visibility in the framework of the computing support for ATLAS collaboration  High-level training of RO technical staff

Mihnea Dulea, IFIN-HH PROSPECTS  Further development of methods and procedures for improving the performance of the RO sites within the FR Cloud  General guidelines regarding the improvement in efficiency of the grid centers which are associated to ATLAS clouds  HAPPSDAG workshop and technical meeting in Bucharest ( )  Participation of IFIN-HH to site and job monitoring in ADC shifts (ATLAS Distributed Computing) or in the monitoring team of FR Cloud.  Dissemination of results

Mihnea Dulea, IFIN-HH THANK YOU FOR YOUR ATTENTION ! Questions?