1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
Oliver Gutsche - CMS / Fermilab Analyzing Millions of Gigabyte of LHC Data for CMS - Discover the Higgs on OSG.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
1 Data Storage MICE DAQ Workshop 10 th February 2006 Malcolm Ellis & Paul Kyberd.
Exploiting the Grid to Simulate and Design the LHCb Experiment K Harrison 1, N Brook 2, G Patrick 3, E van Herwijnen 4, on behalf of the LHCb Grid Group.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Sung-Won Lee 1 Study of Jets Production Association with a Z boson in pp Collision at 7 and 8 TeV with the CMS Detector Kittikul Kovitanggoon Ph. D. Thesis.
03/27/'07T. ISGC20071 Computing GRID for ALICE in Japan Hiroshima University Takuma Horaguchi for the ALICE Collaboration
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
Preparation of KIPT (Kharkov) computing facilities for CMS data analysis L. Levchuk Kharkov Institute of Physics and Technology (KIPT), Kharkov, Ukraine.
LHCb computing in Russia Ivan Korolko (ITEP Moscow) Russia-CERN JWGC, October 2005.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
Introduction to CMS computing J-Term IV 8/3/09 Oliver Gutsche, Fermilab.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
A.Golunov, “Remote operational center for CMS in JINR ”, XXIII International Symposium on Nuclear Electronics and Computing, BULGARIA, VARNA, September,
Meeting, 5/12/06 CMS T1/T2 Estimates à CMS perspective: n Part of a wider process of resource estimation n Top-down Computing.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
1 Development of charm quark tagger for supersymmetric particle search at the CMS detector Kittikul Kovitanggoon, Burin Asavapibhop, Narumon Suwonjandee.
WLCG and the India-CERN Collaboration David Collados CERN - Information technology 27 February 2014.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
Tiers and GRID computing 김 민 석 ( 성균관대 )
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
tons, 150 million sensors generating data 40 millions times per second producing 1 petabyte per second The ATLAS experiment.
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
The ATLAS Computing & Analysis Model Roger Jones Lancaster University ATLAS UK 06 IPPP, 20/9/2006.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Collaborative Research Projects in Australia: High Energy Physicists Dr. Greg Wickham (AARNet) Dr. Glenn Moloney (University of Melbourne) Global Collaborations.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
National e-Science Infrastructure Consortium of THAILAND
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Grid site as a tool for data processing and data analysis
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
A highly reliable data center network topology Tier 1 at JINR
evoluzione modello per Run3 LHC
LHC DATA ANALYSIS INFN (LNL – PADOVA)
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALICE Computing Upgrade Predrag Buncic
LHC Collisions.
CERN, the LHC and the Grid
The ATLAS Computing Model
LHCb thinking on Regional Centres and Related activities (GRIDs)
The LHC Computing Grid Visit of Professor Andreas Demetriou
Presentation transcript:

1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science and High Performance Computing (eHPC2015) Big data management at CMS collaboration with worldwide LHC computing grid

2 2 Outline

3 3 Introductions CMS has had a distributed computing model motivated by various factors. The large quantity of data and computing requirement encouraged distributed resources from a facility infrastructure point of view. Ability to leverage resources at labs and university. Hardware, expertise, infrastructure. Benefits of providing local control of some resources. Ability to secure local funding sources. ~20% of the resources are located at CERN, 40% at T1s, and 40% at T2s, Relies on the development of tools to make transparent access to the resources. Efficient distributed computing services. Can only be successful with sufficient networking between facilities. Availability of high performance networks has made the distributed model feasible.

4 4 Large Hadron Collider (LHC) 27 km in circumference To collide rotating beams of protons or heavy ions Maximum energy of proton- proton collisions at = 14 TeV and 4 x cm -2 s -1 In 2011, collision at = 7 TeV and 4 x cm -2 s -1 In 2012, collision at = 8 TeV and 7.7 x cm -2 s -1 In 2015, expect collision at = 13 TeV and 22.8 x cm - 2 s -1 CMS ALICE ATLAS LHCb

5 5 Compact Muon Solenoid (CMS)

6 6 CMS Collisions

7 7 CMS Collision Data ● CMS detectors are gigantic digital cameras that can identify various elementary particles from the millions of collisions per second. ● Decay particles from each collisions will be: Recorded the passage of each particle through various sub-detectors as a series of electronic signals. Sent the data to the CERN Data Centre (DC) for digital reconstruction. Reconstructed digitized summary as a `collision event’. Data from the CMS experiments will be distributed around the globe with the Worldwide LHC Computing Grid (WLCG) project that is built and maintained for data storage and provides analysis infrastructure for the entire CMS community Thailand Involved in WLCG: Tier-2/Tier-3 computing centres of CMS Thailand [T2_TH_CUNSTDA and T3_TH_CHULA].

8 8 CMS Physics and Event Rates Design Luminosity (L) = cm -2 s pp events/25 ns xing ~ 1 GHz input rate “Good” events contain ~ 20 bkg. Events 1 kHz W events 10 Hz top events < 104 detectable Higgs decays/year Can store ~ 300 Hz events Select in stages Level-1 Triggers 1 GHz to 100 kHz High Level Triggers 100 kHz to 300 Hz

9 9 CMS Physics and Event Rates

10 CMS Data Flow

11 Worldwide LHC Computing Grid

12 Tier-0 The first tier in the CMS model, for which there is only one site, CERN, is known as Tier-0 (T0). The T0 performs several functions. The standard workflow is as follows: 1.accepts RAW data from the CMS Online Data Acquisition and Trigger System (TriDAS) 2.repacks the RAW data received from the DAQ into primary datasets based on trigger information 3.archives the repacked RAW data to tape 4.distributes RAW data sets among the next tier stage resources (Tier-1) so that two copies are saved 5.performs PromptCalibration in order to get the calibration constants needed to run the reconstruction 6.feeds the RAW datasets to reconstruction 7.performs prompt first pass reconstruction which writes the RECO and Analysis Object Data (AOD) extraction 8.distributes the RECO datasets among Tier-1 centers, such that the RAW and RECO match up at each Tier-1 9.distributes full AOD to all Tier-1 centers The T0 does not provide analysis resources and only operates scheduled activities.

13 Tier-1 There is a set of thirteen Tier-1 (T1) sites, which are large centers in CMS collaborating countries (large national labs, e.g. FNAL, and RAL). Tier-1 sites will in general be used for large-scale, centrally organized activities and can provide data to and receive data from all Tier-2 sites. Each T1 center: 1.receives a subset of the data from the T0 related to the size of the pledged resources in the WLCG MOU 2.provides tape archive of part of the RAW data (secure second copy) which it receives as a subset of the datasets from the T0 3.provides substantial CPU power for scheduled: re-reconstruction, skimming, calibration, and AOD extraction 4.stores an entire copy of the AOD 5.distributes RECOs, skims and AOD to the other T1 centers and CERN as well as the associated group of T2 centers 6.provides secure storage and redistribution for MC events generated by the T2's

14 Tier-2 A more numerous set of smaller Tier-2 (T2) centers but with substantial CPU resources. T2 provide: 1.services for local communities 2.grid-based analysis for the whole experiment (Tier-2 resources available to whole experiment through the grid) 3.Monte Carlo simulation for the whole experiment T2 centers rely upon T1s for access to large datasets and for secure storage of the new data (generally Monte Carlo) produced at the T2. The MC production in Tier-2's will in general be centrally organized, with generated MC samples being sent to an associated Tier-1 site for distribution among the CMS community. All other Tier-2 activities will be user driven, with data placed to match resources and needs: tape, disk, manpower, and the needs of local communities. The Tier-2 activities will be organized by the Tier-2 authorities in collaboration with physics groups, regional associations and local communities.

15 Thailand at CERN WLCG Tier-2/Tier-3 computing centres of CMS Thailand [T2_TH_CUNSTDA and T3_TH_CHULA] CMS Chulalongkorn University, Bangkok ALICE King Mongkut's University of Technology Thonburi (KMUTT), Bangkok Thai Microelectronics Center (TMEC), Muang Chachoengsao Suranaree University of Technology, Nakhon Ratchasima CERN and Thailand: relations/nms/thailand.html

16 Tier-2/Tier-3 Monitoring Monitoring Tier-2/Tier-3 of CMS Thailand that are part of worldwide LHC computing grid (WLCG). Duties for T2 TH CUNSTDA and T3 TH CHULA Periodical investigation of CMS site readiness, availability and reliability status. To maintain adequate network bandwidth and high PhEDEx – CMS data transfers. Contribute to preserve memory for analysis operations, data operations, and local activities. WLCG squid monitoring for server traffic volume, HTTP hits/requests, cached objects.

17 Tier-2/Tier-3 Monitoring

18 Tier-2/Tier-3 Monitoring

19 Tier-3 Maintenance Working with IBM system x3755 M3, x3550 M4, BladeCenter H Ensuring suitable working condition of Tier-3 system, user machines and web-server hosts, etc. Management of user accounts and corresponding data. Serving e-science etc. etc.

20 Conclusions

21 Acknowledgments This research is supported by Rachadapisek Sompote Fund for Postdoctoral Fellowship, Chulalongkorn University. Department of Physics, Faculty of Science, Chulalongkorn University for financial support. The CMS collaboration. All eHPC 2015 staffs for organizing this event.