1. 2 Overview Extremely short summary of the physical part of the conference (I am not a physicist, will try my best) Overview of the Grid session focused.

Slides:



Advertisements
Similar presentations
1 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna A. Fanfani University of Bologna MC Production System & DM catalogue.
Advertisements

CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Stuart K. PatersonCHEP 2006 (13 th –17 th February 2006) Mumbai, India 1 from DIRAC.Client.Dirac import * dirac = Dirac() job = Job() job.setApplication('DaVinci',
A tool to enable CMS Distributed Analysis
Analysis demos from the experiments. Analysis demo session Introduction –General information and overview CMS demo (CRAB) –Georgia Karapostoli (Athens.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Physicists's experience of the EGEE/LCG infrastructure usage for CMS jobs submission Natalia Ilina (ITEP Moscow) NEC’2007.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
Enabling Grids for E-sciencE Overview of System Analysis Working Group Julia Andreeva CERN, WLCG Collaboration Workshop, Monitoring BOF session 23 January.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Julia Andreeva CERN (IT/GS) CHEP 2009, March 2009, Prague New job monitoring strategy.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
The first year of LHC physics analysis using the GRID: Prospects from ATLAS Davide Costanzo University of Sheffield
IST E-infrastructure shared between Europe and Latin America High Energy Physics Applications in EELA Raquel Pezoa Universidad.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
PanDA Update Kaushik De Univ. of Texas at Arlington XRootD Workshop, UCSD January 27, 2015.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Overview of STEP09 monitoring issues Julia Andreeva, IT/GS STEP09 Postmortem.
…building the next IT revolution From Web to Grid…
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
A PanDA Backend for the Ganga Analysis Interface J. Elmsheuser 1, D. Liko 2, T. Maeno 3, P. Nilsson 4, D.C. Vanderster 5, T. Wenaus 3, R. Walker 1 1: Ludwig-Maximilians-Universität.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Monitoring for CCRC08, status and plans Julia Andreeva, CERN , F2F meeting, CERN.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
The GridPP DIRAC project DIRAC for non-LHC communities.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
ANALYSIS TOOLS FOR THE LHC EXPERIMENTS Dietrich Liko / CERN IT.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Shifters Jamboree Kaushik De ADC Jamboree, CERN December 4, 2014.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
CERN - IT Department CH-1211 Genève 23 Switzerland t Grid Reliability Pablo Saiz On behalf of the Dashboard team: J. Andreeva, C. Cirstoiu,
The GridPP DIRAC project DIRAC for non-LHC communities.
WMS baseline issues in Atlas Miguel Branco Alessandro De Salvo Outline  The Atlas Production System  WMS baseline issues in Atlas.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
1 Tutorial:Initiation a l’Utilisation de la Grille EGEE/LCG, June 5-6 N. De Filippis CMS tools for distributed analysis N. De Filippis - LLR-Ecole Polytechnique.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
U.S. ATLAS Grid Production Experience
INFN-GRID Workshop Bari, October, 26, 2004
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Simulation use cases for T2 in ALICE
LCG middleware and LHC experiments ARDA project
Monitoring of the infrastructure from the VO perspective
The LHCb Computing Data Challenge DC06
Presentation transcript:

1

2 Overview Extremely short summary of the physical part of the conference (I am not a physicist, will try my best) Overview of the Grid session focused on the distributed analysis

3 The Standard Model of Fundamental Interactions Something one may be proud of D. Kazakov

4 Physics beyond the SM Low Energy Supersymmetry Extra gauge bosons Axions Extra dimensions Deviation from Unitarity triangle Modification of Newton law Free quarks New forces / particles Violation of Baryon number Violation of Lepton number Monopoles Violation of Lorentz invariance Compositeness Not found so far … D. Kazakov

5 The SM Higgs Boson If it is there we may see it soon Indirect limit from radiative corrections Direct limit from Higgs non observation at LEP II (CERN) D. Kazakov

6 Some new results

7 But they correspond to the Standard Model

8 Conference Summary Valery Rubakov

9 “Blessed are those who believe and yet have not seen” “Blessed are those who believe and yet have not seen” Instead of Conclusion St. John, XX, 29 “Блаженны не видевшие и уверовавшие” “Блаженны не видевшие и уверовавшие” Св. Иоанн, XX, 29 Dmitry Kazakov

10 Grid session First time Grid distributed analysis session was included in the ICHEP scientific program Talks on the status of the distributed analysis on the Grid from LHC experiments (ATLAS, ALICE and CMS). Yet more plans and intentions than status reports LCG status talk, network report, common software talk

11 Distributed Analysis Challenges Distributed production is now routinely done in HENP For MC production and reprocessing of data - not yet LHC scale Scale: few TB’s of data generated/processed daily in ATLAS Scope: organized activity, managed by experts Lessons learned from production Robust software systems to automatically recover from grid failures Robust site services – with hundreds of sites, there are daily failures Robust data management – pre-location of data, cataloguing, transfers Distributed analysis is in early stages of testing Moving from Regional Analysis Center model (ex. D0) to fully distributed analysis model – computing on demand Presents new challenges, in addition to those faced in production Chaotic by nature – hundreds of users, random fluctuations in demand Robustness becomes even more critical – software, sites, services Kaushik De

12 Divide and Conquer Experiments optimize/factorize both data and resources Data factorization Successive processing steps lead to compressed physics objects End user does physics analysis using physics objects only Limited access to detailed data for code development, calibration Periodic centralized reprocessing to improve analysis objects Resource factorization Tiered model of data location and processors Higher tiers hold archival data and perform centralized processing Middle tiers for MC generation and some (re)processing Middle and lower tiers play important role in distributed analysis Regional centers are often used to aggregate nearby resources Kaushik De

13 Example from D0 from A. Boehnlein Kaushik De

14 Common features Becoming pragmatic Organization of the analysis workflow is more and more driven by the data management organization - Sending analysis jobs close to the input data sets - Merging MC and Analysis output into file blocks allowing to organize data access in a more optimal way (ATLAS,CMS) - Aiming to decrease a load on the central data management services (catalogue) and work load management services (RB)  Common system for distributed analysis and production (Alien, Panda)  Sort of central queue for the Grid (Alien, Dirac, Atlas Production System tried for analysis) Though this approach is not shared by all experiments  Develop Job Submission Tools which should provide for the user simple interface to the Grid (Ganga, Crab, Panda)

15 Job submission tools User Interface Logging and bookkeeping Computer sites WMS Information System VOMS Getting proxy Submitting a job Register Checking job status Experiment software VO specific Grid flavour specific Experiment Data Managent LHC specific applications Job Submission tool Talks to experiment DM to find out where data is and how to split user task Implements task splitting Does packaging of the user code and libraries Generates executable shell Generates Grid submission instructions Submits all jobs belonging to a task Check status of jobs belonging to a given task and retrieves the job output File Catalogue Error recovery Job submission tools should hide from the user complexity of dealing with the distributed computing facility, providing simple and user-friendly interface Experiments develop different solutions Examples: GANGA for ATLAS and LHCb CRAB and ASAP for CMS

16 Job Management: Productions (ATLAS) Once we have data distributed in the correct way we can rework the distributed production system to optimise job distribution, by sending jobs to the data (or as close as possible to them) This was not the case previously, as jobs were sent to free CPUs and had to copy the input file(s) to the local WN, from wherever in the world the data happened to be Next: make better use of the task and dataset concepts A “task” acts on a dataset and produces more datasets Use bulk submission functionality to send all jobs of a given task to the location of their input datasets Minimize the dependence on file transfers and the waiting time before execution Collect output files belonging to the same dataset to the same SE and transfer them asynchronously to their final locations David Constanzo

17

18 Analysis statistics for CMS Widely used by CMS physics community Plot shows CMS analysis jobs submitted via Crab for the period distributed by site: ~83K jobs, 50 users over 85 sites

19 Interactive analysis in Alice A user starts ROOT session on a laptop The analysis macros are started from the ROOT command line The data files on the GRID are accessed using ROOT (AliEn) UI (via xrootd) The results are stored locally or can be registered on the GRID (AliEn file catalogue) If the data files are stored on a cluster, the interactive analysis is done in parallel using PROOF I. Belikov

20 Common concerns Robustness of data management critically important (Kaushik De, Atlas) Problem diagnosis and debugging (Ian Fisk, CMS) Do not yet distinguish activities for prioritization (Atlas and CMS) Need for increase reliability of the Grid and experiment infrastructure (Atlas and CMS)

21 Conclusions Scale of resources and users unprecedented at LHC. Enabling of the distributed analysis is a serious challenge for the LHC experiments A big progress is already done in this direction. Experiments defined their computing models. Experiment data management and work load management systems are mostly in place Still a lot of work has to be done to ensure that both Grid and experiment specific services and infrastructures do provide a required level of scalability, reliability and performance