Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.

Slides:



Advertisements
Similar presentations
1 AMY Detector (eighties) A rather compact detector.
Advertisements

31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
First results from the ATLAS experiment at the LHC
Southwest Tier 2 Center Status Report U.S. ATLAS Tier 2 Workshop - Harvard Mark Sosebee for the SWT2 Center August 17, 2006.
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
Laura Gilbert Particle Physics and Grid – A Virtualisation Study.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.
HPC at IISER Pune Neet Deo System Administrator
Particle Physics and the Grid Randall Sobie Institute of Particle Physics University of Victoria Motivation Computing challenge LHC Grid Canadian requirements.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Cluster currently consists of: 1 Dell PowerEdge Ghz Dual, quad core Xeons (8 cores) and 16G of RAM Original GRIDVM - SL4 VM-Ware host 1 Dell PowerEdge.
LOGO Scheduling system for distributed MPD data processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
TechFair ‘05 University of Arlington November 16, 2005.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Data Import Data Export Mass Storage & Disk Servers Database Servers Tapes Network from CERN Network from Tier 2 and simulation centers Physics Software.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
High Energy Physics at UTA UTA faculty Andrew Brandt, Kaushik De, Andrew White, Jae Yu along with many post-docs, graduate and undergraduate students investigate.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
The Computing System for the Belle Experiment Ichiro Adachi KEK representing the Belle DST/MC production group CHEP03, La Jolla, California, USA March.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Computing for LHCb-Italy Domenico Galli, Umberto Marconi and Vincenzo Vagnoni Genève, January 17, 2001.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Status of UTA IAC + RAC Jae Yu 3 rd DØSAR Workshop Apr. 7 – 9, 2004 Louisiana Tech. University.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
Indexing and Selection of Data Items Using Tag Collections Sebastien Ponce CERN – LHCb Experiment EPFL – Computer Science Dpt Pere Mato Vila CERN – LHCb.
Rob Allan Daresbury Laboratory NW-GRID Training Event 25 th January 2007 Introduction to NW-GRID R.J. Allan CCLRC Daresbury Laboratory.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
High Energy Physics & Computing Grids TechFair Univ. of Arlington November 10, 2004.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
Computational Research in the Battelle Center for Mathmatical medicine.
Outline  Higgs Particle Searches for Origin of Mass  Grid Computing  A brief Linear Collider Detector R&D  The  The grand conclusion: YOU are the.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
ATLAS Midwest Tier2 University of Chicago Indiana University Rob Gardner Computation and Enrico Fermi Institutes University of Chicago WLCG Collaboration.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Materials for Report about Computing Jiří Chudoba x.y.2006 Institute of Physics, Prague.
Atlas Software Structure Complicated system maintained at CERN – Framework for Monte Carlo and real data (Athena) MC data generation, simulation and reconstruction.
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
UTA Site Report Jae Yu UTA Site Report 7 th DOSAR Workshop Louisiana State University Apr. 2 – 3, 2009 Jae Yu Univ. of Texas, Arlington.
Joint Institute for Nuclear Research Synthesis of the simulation and monitoring processes for the data storage and big data processing development in physical.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
6th DOSAR Workshop University Mississippi Apr. 17 – 18, 2008
Grid site as a tool for data processing and data analysis
U.S. ATLAS Tier 2 Computing Center
Southwest Tier 2 Center Status Report
5th DOSAR Workshop Louisiana Tech University Sept. 27 – 28, 2007
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Southwest Tier 2.
ALICE Computing Model in Run3
LHC Collisions.
High Energy Physics at UTA
Nuclear Physics Data Management Needs Bruce G. Gibbard
Support for ”interactive batch”
High Energy Physics at UTA
Presentation transcript:

Computing Issues for the ATLAS SWT2

What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University of Oklahoma and Langston University One of five Tier 2 centers in the U.S. Builds, operates and maintains computing resources for U.S. ATLAS

ATLAS ATLAS is a particle detector being constructed as part of the Large Hadron Collider (LHC) LHC is in Geneva, Switzerland at CERN Collides Protons at 14 TeV with 40 MHz proton bunch crossing rate The goal is produce ~100 Hz event recording 1.5 PB raw data / year

Monte-Carlo Simulations in ATLAS Why simulate? –Trigger refinement –Reconstruction refinement –Understanding new phenomena Four step process –Generation –Simulation –Digitization –Reconstruction

MC Generation Generation step (Pythia is one example) –Work from a description of what type of event to generate Physics process Initial Energy Selection Criteria –Software uses MC techniques to generate candidate events –If events don’t meet criteria, more events generated

Simulation Uses generator output (particles/momenta) Physical description of detector (materials/environment)– Based on GEANT Produces accurate depiction of how the particles move through the various parts of the detector, losing/depositing energy in the process This step consumes the most CPU time

Digitization Introduces detector measurements Simulates behavior of detector electronics Input is taken from the simulation step Output is description of the event as seen by the measurement systems Output “looks like” real data from detector

Reconstruction Reconstructs physics objects from the raw data Same code is used for MC data and real data Input is taken from digitization step (or detector) Output is the data physicists use for analysis

Computing Event simulation is computationally intensive Simulation step takes a few minutes up to an hour per event Large need for simulated data during the lifetime of the experiment. Yields large need for compute cycles

Computing (cont.) At full luminosity raw data will be ~1.5 PB/year Data for analysis will be order of magnitude smaller (but retained) Most ATLAS physicists need access to this data Yields large need for storage resources

ATLAS Tier Model ATLAS is coping with “tiered” structure. Tier 0 (CERN) Tier 1 National Facility (Brookhaven National Laboratory) Tier 2 Regional Facility (SWT2, NET2, AGLT2, MWT2, SLAC) Tier3 University Facility (e.g. DPCC, HPC)

Purpose of a Tier 2 Perform Monte-Carlo based simulations of collision events Perform Reprocessing, converting raw data to physics description Provide resources for user analysis

Distributed Computing in ATLAS Data handling is central issue Move jobs or data Production versus Analysis –Production is CPU intensive –Analysis is I/O intensive How to manage user access Grid Computing

How To Satisfy demands Balance cost / performance SMP vs Commodity processor SMP’s –Great for memory intensive multi-threaded applications –Expensive Computing Clusters –Less expensive –Less performance for multi-threaded applications Can be improved by spending money on interconnect network

UTA Computing Clusters DPCC (Joint project between Physics and CSE) –Installed 2003 –~80 Node cluster (160 processors) –50TB Storage UTA_SWT2 –Installed 2006 –160 Node Cluster (320 processors) –16 TB Storage UTA_CPB –Being installed now –50 Node (200 processors) –60TB Storage

What Makes a Computing Cluster Head node(s) –Allows interactive login, compilation Worker nodes –Provide computing cycles, may not be accessible Network –Can be most important aspect depending on computing model Storage access –NFS, Global file systems Batch system –Controls access to worker nodes by scheduling work

Networking Cost vs. Performance for application –Main issue is communication latency Low cost (Ethernet) –Suitable for single threaded applications –Can be used for multi-threaded applications –Higher latency High cost (Myricom, Infiniband) –Low latency –Best suited for improving multi-threaded apps

DPCC Resources 80 worker nodes –Dual 2.4 / 2.66 GHz Xeon processors –2GB RAM –60/80 GB local Disk 50 TB RAID storage –11 raid servers (3 x 1.5 TB RAID Cards) 1000 Mb/s Ethernet internal network

DPCC Diagram

UTA_SWT2 Resources Wholly operated by UTA for ATLAS 160 Worker nodes –Dual 3.2 GHz Xeon EM64T processors –4GB RAM –160 GB local disk 16 TB Storage –DataDirect S2A3000 SAN –6 I/O servers running IBRIX Dual internal networks –Management (100 Mbs) –Storage (1000 Mbs)

UTA_SWT2

UTA_CPB Being constructed now 50 Worker nodes –Dual dual-core 2.4 GHz Opteron (2216) –8 GB RAM –80 GB Local Storage 60 TB Storage –6 I/O servers with attached storage –10 Dell MD1000 Storage Units (7.5 TB RAW) Single 1000 Mb/s Network Will likely be supplemented this year with additional storage

UTA_CPB

Summary 520 dedicated cores + ~100 additional available 90 TB disk space + ~30TB additional available