EGI-InSPIRE EGI-InSPIRE RI- 261323 www.egi.eu DDM Site Services winter release Fernando H. Barreiro Megino (IT-ES-VOS) ATLAS SW&C Week November 2010 1.

Slides:



Advertisements
Similar presentations
ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
Advertisements

CERN IT Department CH-1211 Geneva 23 Switzerland t ES 1 Alessandro Di Girolamo Experiment Support 4 July 2011 AMOD weekly report 27 th June.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
Staging to CAF + User groups + fairshare Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE Offline week,
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Summary of issues and questions raised. FTS workshop for experiment integrators Summary of use  Generally positive response on current state!  Now the.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
FZU participation in the Tier0 test CERN August 3, 2006.
ATLAS DQ2 Deletion Service D.A. Oleynik, A.S. Petrosyan, V. Garonne, S. Campana (on behalf of the ATLAS Collaboration)
Claudio Grandi INFN Bologna CMS Operations Update Ian Fisk, Claudio Grandi 1.
CERN IT Department CH-1211 Genève 23 Switzerland t EIS section review of recent activities Harry Renshall Andrea Sciabà IT-GS group meeting.
Introduction: Distributed POOL File Access Elizabeth Gallas - Oxford – September 16, 2009 Offline Database Meeting.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
Integrating HPC into the ATLAS Distributed Computing environment Doug Benjamin Duke University.
PanDA Summary Kaushik De Univ. of Texas at Arlington ADC Retreat, Naples Feb 4, 2011.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI AMOD report – Fernando H. Barreiro Megino CERN-IT-ES-VOS.
BNL DDM Status Report Hironori Ito Brookhaven National Laboratory.
CERN Using the SAM framework for the CMS specific tests Andrea Sciabà System Analysis WG Meeting 15 November, 2007.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Direct gLExec integration with PanDA Fernando H. Barreiro Megino CERN IT-ES-VOS.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Overview of STEP09 monitoring issues Julia Andreeva, IT/GS STEP09 Postmortem.
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Usage of virtualization in gLite certification Andreas Unterkircher.
DDM Monitoring David Cameron Pedro Salgado Ricardo Rocha.
Stefano Belforte INFN Trieste 1 Middleware February 14, 2007 Resource Broker, gLite etc. CMS vs. middleware.
Storage cleaner: deletes files on mass storage systems. It depends on the results of deletion, files can be set in states: deleted or to repeat deletion.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
ATLAS Bulk Pre-stageing Tests Graeme Stewart University of Glasgow.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
Distributed Transaction Management, Fall 2002Lecture 2 / Distributed Locking Jyrki Nummenmaa
INFSO-RI Enabling Grids for E-sciencE ATLAS DDM Operations - II Monitoring and Daily Tasks Jiří Chudoba ATLAS meeting, ,
EGI-InSPIRE EGI-InSPIRE RI DDM solutions for disk space resource optimization Fernando H. Barreiro Megino (CERN-IT Experiment Support)
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
The new FTS – proposal FTS status. EMI INFSO-RI /05/ FTS /05/ /05/ Bugs fixed – Support an SE publishing more than.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Service Availability Monitor tests for ATLAS Current Status Tests in development To Do Alessandro Di Girolamo CERN IT/PSS-ED.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Andrea Sciabà Hammercloud and Nagios Dan Van Der Ster Nicolò Magini.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Ops Portal New Requirements.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Data Management Highlights in TSA3.3 Services for HEP Fernando Barreiro Megino,
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
INFSO-RI Enabling Grids for E-sciencE FTS failure handling Gavin McCance Service Challenge technical meeting 21 June.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
1 November 17, 2005 FTS integration at CMS Stefano Belforte - INFN-Trieste CMS Computing Integration Coordinator CMS/LCG integration Task Force leader.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
Dynamic Data Placement: the ATLAS model Simone Campana (IT-SDC)
ATLAS Distributed Computing ATLAS session WLCG pre-CHEP Workshop New York May 19-20, 2012 Alexei Klimentov Stephane Jezequel Ikuo Ueda For ATLAS Distributed.
Distributed Data Management Miguel Branco 1 DQ2 status & plans BNL workshop October 3, 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
LHCb 2009-Q4 report Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.
GGUS summary (3 weeks) VOUserTeamAlarmTotal ALICE7029 ATLAS CMS LHCb Totals
PanDA Configurator and Network Aware Brokerage Fernando Barreiro Megino, Kaushik De, Tadashi Maeno 14 March 2015, US ATLAS Distributed Facilities Meeting,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Status of ARGUS support Peter Solagna – EGI.eu.
PD2P, Caching etc. Kaushik De Univ. of Texas at Arlington ADC Retreat, Naples Feb 4, 2011.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Update on Service Availability Monitoring (SAM) Marian Babik, David Collados,
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Feedback to sites from the VO auger Jiří Chudoba (Institute of Physics and.
ATLAS Use and Experience of FTS
Summary on PPS-pilot activity on CREAM CE
Status of the CERN Analysis Facility
BNL FTS services Hironori Ito.
Data Management cluster summary
X in [Integration, Delivery, Deployment]
lundi 25 février 2019 FTS configuration
Status and plans for bookkeeping system and production tools
Presentation transcript:

EGI-InSPIRE EGI-InSPIRE RI DDM Site Services winter release Fernando H. Barreiro Megino (IT-ES-VOS) ATLAS SW&C Week November

EGI-InSPIRE RI New release of Site Services planned for this week on preproduction Running on Functional Tests since mid October with exception of some of the new features Heavy update: Many new features and bug fixes –Some to be tuned and improved after production experience Introduction 2

EGI-InSPIRE RI Breaking cloud boundaries and improving source selection 3

EGI-InSPIRE RI Tier1 Tier2 Tier1’ Tier2’ Cloud1 Cloud2 Breaking cloud boundaries 4 Current Computing Model does not foresee direct T2 cloud1  T2’ cloud2 communication But we need to break the cloud boundaries! –User analysis with jobs running in different clouds and collected in one place –PD2P Possibilities: –FTS STAR-Channel transfers –Multihopping (not natively supported in FTS, done via DaTRI atm) Proposed solution inside Site Services: Automatic path selection: cost estimation of STAR-channel and multihop transfers based on FTS statistics In case of multihop: Submission of child subscriptions through Tier1 SCRATCHDISKs Proposed solution inside Site Services: Automatic path selection: cost estimation of STAR-channel and multihop transfers based on FTS statistics In case of multihop: Submission of child subscriptions through Tier1 SCRATCHDISKs

EGI-InSPIRE RI The setup 5

EGI-InSPIRE RI Django, Matplotlib, jquery Putting final touches to first version FTS performance monitoring Andrii Thykonov 6

EGI-InSPIRE RI Some example measurements Source*Destination*Small files (0-100MB) avg(std) Medium files (0.1-1GB) avg(std) UTA_SWT2IN2P3-LAPP0.9MB/s (38KB/s) 5 file transfers 5.4MB/s (0.9MB/s) 5 files transfers UTA_SWT2BNL-OSG20.8MB/s (41KB/s) 5 file transfers 3.7MB/s (0.4MB/s) 5 file transfers BNL-OSG2IN2P3-CC0.1MB/s (81KB/s) 1656 file transfers 7.8MB/s (0.7MB/s) 22 file transfers IN2P3-CCIN2P3-LAPP0.7MB/s (473KB/s) 320 file transfers 1.2MB/s (14KB/s) 157 file transfers * Always DATADISK spacetoken Multihopping does not seem to make sense for small and medium files unless STAR channels collapse 7 Transfer times of small files are completely dominated by SRM interactions

EGI-InSPIRE RI Other improvements 8

EGI-InSPIRE RI Priority of activities can now be configured Files will be partitioned into FTS jobs with equal priority  Priority lanes for express file replication (e.g. for DBRelease replication) FTS priority lanes 9

EGI-InSPIRE RI Before: Site exclusion prevented new subscriptions to be processed, but subscriptions already in the queue of the system were generating FTS/staging jobs that were condemned to fail. Now: In addition, site exclusion will avoid generating new FTS/staging jobs with already queued files. 2 cases: –Destination gets blacklisted: Just wait –Source gets blacklisted: Look for different sources to get the subscriptions moving. Management of excluded sites in SS 10

EGI-InSPIRE RI Changed the polling methods: gfal_ls (aka SRM_ls) replaced by gfal_prestagestatus –Polling of requests instead of polling of single files –An aborted request by the server is noticed immediately. Before we would wait for several hours until timeout 2.Fixed “mystical” bug that was preventing the correct filling of the internal SS staging queues (symptom: staging at sites being stuck) Improvements in the stager 11

EGI-InSPIRE RI Automatic restarting after reboot –Not new - had to be understood after SLC5 migration –Init.d script improved –Rebooting campaigns should not be an issue anymore On demand file callbacks: –Analysis tools can know immediately successfully copied/staged files and are able to start releasing jobs And many other smaller fixes and features… Others 12

EGI-InSPIRE RI Work in progress: ActiveMQ integration 13

EGI-InSPIRE RI ActiveMQ integration for messaging Site Services send many different callbacks to dashboard and analysis tools. Examples: –Dataset content –File events (Transferring, copied, registered, failed…) –Subscription events (Queued, completed, canceled, broken…) Message queues will allow to send once and listen by everyone interested Testing environment set up Martin will take care of the development (See Donal’s talk for some more information on ActiveMQ) ActiveMQ integration 14

EGI-InSPIRE RI Simone Campana Stephane Jezequel Vincent Garonne Andrii Thykonov David Tuckett I.Ueda CREDITS 15