HEP and Data Grids (Aug. 4-5, 2001)Paul Avery1 High Energy Physics and Data Grids Paul Avery University of Florida

Slides:



Advertisements
Similar presentations
International Grid Communities Dr. Carl Kesselman Information Sciences Institute University of Southern California.
Advertisements

1 AMY Detector (eighties) A rather compact detector.
First results from the ATLAS experiment at the LHC
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
Welcome to CERN Accelerating Science and Innovation 2 nd March 2015 – Bidders Conference – DO-29161/EN.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
US-CMS Meeting (May 19, 2001)Paul Avery1 US-CMS Meeting (UC Riverside) May 19, 2001 Grids for US-CMS and CMS Paul Avery University of Florida
The LHC Computing Grid Project Tomi Kauppi Timo Larjo.
POLITEHNICA University of Bucharest California Institute of Technology National Center for Information Technology Ciprian Mihai Dobre Corina Stratan MONARC.
The CMS Muon Detector Thomas Hebbeker Aachen July 2001 Searching for New Physics with High Energy Muons.
HEP Prospects, J. Yu LEARN Strategy Meeting Prospects on Texas High Energy Physics Network Needs LEARN Strategy Meeting University of Texas at El Paso.
Exploiting the Grid to Simulate and Design the LHCb Experiment K Harrison 1, N Brook 2, G Patrick 3, E van Herwijnen 4, on behalf of the LHCb Grid Group.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
International collaboration in high energy physics experiments  All large high energy physics experiments today are strongly international.  A necessary.
1 Chasing the Higgs boson with a worldwide distributed trigger system Sander Klous NIKHEF VENI proposal 2006.
Recipe for the Universe in the High School Classroom Randy Ruchti Department of Physics University of Notre Dame.
1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
CHEP 2000 (Feb. 7-11)Paul Avery (Data Grids in the LHC Era)1 The Promise of Computational Grids in the LHC Era Paul Avery University of Florida Gainesville,
Rackspace Analyst Event Tim Bell
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
Ruth Pordes, Fermilab CD, and A PPDG Coordinator Some Aspects of The Particle Physics Data Grid Collaboratory Pilot (PPDG) and The Grid Physics Network.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
GriPhyN EAC Meeting (Jan. 7, 2002)Carl Kesselman1 University of Southern California GriPhyN External Advisory Committee Meeting Gainesville,
GriPhyN EAC Meeting (Apr. 12, 2001)Paul Avery1 University of Florida Opening and Overview GriPhyN External.
Brussels Grid Meeting (Mar. 23, 2001)Paul Avery1 University of Florida Extending the Grid Reach in Europe.
What is the Higgs??? Prof Nick Evans University of Southampton.
Search for a Z′ boson in the dimuon channel in p-p collisions at √s = 7TeV with CMS experiment at the Large Hadron Collider Search for a Z′ boson in the.
Collider Detector at Fermilab Sung-hyun chang High Energy Physics lab. KNU.
Internet 2 Workshop (Nov. 1, 2000)Paul Avery (The GriPhyN Project)1 The GriPhyN Project (Grid Physics Network) Paul Avery University of Florida
…building the next IT revolution From Web to Grid…
Silicon Module Tests The modules are tested in the production labs. HIP is is participating in the rod quality tests at CERN. The plan of HIP CMS is to.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
PPDGLHC Computing ReviewNovember 15, 2000 PPDG The Particle Physics Data Grid Making today’s Grid software work for HENP experiments, Driving GRID science.
GriPhyN EAC Meeting (Jan. 7, 2002)Paul Avery1 Integration with iVDGL è International Virtual-Data Grid Laboratory  A global Grid laboratory (US, EU, Asia,
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
WelcomeWelcome CSEM – CERN Day 23 rd May 2013 CSEM – CERN Day 23 rd May 2013 to Accelerating Science and Innovation to Accelerating Science and Innovation.
Searching for New Matter with the D0 Experiment Todd Adams Department of Physics Florida State University September 19, 2004.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
LHC Computing, CERN, & Federated Identities
John Womersley 1/13 Fermilab’s Future John Womersley Fermilab May 2004.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
LHC Computing, SPC-FC-CC-C; H F Hoffmann1 CERN/2379/Rev: Proposal for building the LHC computing environment at CERN (Phase 1) Goals of Phase.
05 Novembre years of research in physics European Organization for Nuclear Research.
Germany and CERN / June 2009Germany and CERN | May Welcome - Willkommen CERN: to CERN: Accelerating Science and Innovation Professor Wolfgang A.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
WLCG – Status and Plans Ian Bird WLCG Project Leader openlab Board of Sponsors CERN, 23 rd April 2010.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Hall D Computing Facilities Ian Bird 16 March 2001.
] Open Science Grid Ben Clifford University of Chicago
The CMS Experiment at LHC
The 5 minutes tour of CERN The 5 minutes race of CERN
The 5 minutes tour of CERN The 5 minutes race of CERN
LHC DATA ANALYSIS INFN (LNL – PADOVA)
Status of CMS and the Austrian Contribution to the Trigger System
CERN, the LHC and the Grid
LHC Computing Grid Project
LHC Computing Grid Project
Presentation transcript:

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery1 High Energy Physics and Data Grids Paul Avery University of Florida US/UK Grid Workshop San Francisco August 4-5, 2001

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery2 Essentials of High Energy Physics è Better name  “Elementary Particle Physics”  Science: Elementary particles, fundamental forces e   udud cscs tbtb è Goal  unified theory of nature  Unification of forces (Higgs, superstrings, extra dimensions, …)  Deep connections to large scale structure of universe  Large overlap with astrophysics, cosmology, nuclear physics Quarks Leptons ParticlesForces Strong  gluon Electro-weak  , W , Z 0 Gravity  graviton

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery m~ 10 eV>300,000 Y m MeV - GeV m>> GeV~10 -6 sec m ~ 100 GeV~ sec Quantum Mechanics Atomic physics Quantum Electro Dynamics Nuclei, Hadrons Symmetries, Field theories Quarks. Gauge theories 1990 LEP3 families, Precision Electroweak m~10 2 GeV Origin of masses The next step... ~ sec2007 LHCHiggs ? Supersymmetry ? 1970  83 SPS ElectroWeak unification, QCD ~ 3 min m~10 16 GeV~ secProton Decay ? Underground GRAND Unified Theories ? m~10 19 GeV (Planck scale) ~ sec??Quantum Gravity? Superstrings ? The Origin of the Universe 1994 TevatronTop quark ue + Z e - u HEP Short History + Frontiers

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery4 HEP Research è Experiments primarily accelerator based  Fixed target, colliding beams, special beams è Detectors  Small, large, general purpose, special purpose è … but wide variety of other techniques  Cosmic rays, proton decay, g-2, neutrinos, space missions è Increasing scale of experiments and laboratories  Forced on us by ever higher energies  Complexity, scale, costs  large collaborations  International collaborations are the norm today  Global collaborations are the future (LHC) LHC discussed in next few slides

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery5 The CMS Collaboration Member States Non-Member States Total USA Member States Total USA 50 Non-Member States Number of Scientists Number of Laboratories Slovak Republic CERN France Italy UK Switzerland USA Austria Finland Greece Hungary Belgium Poland Portugal Spain Pakistan Georgia Armenia Ukraine Uzbekistan Cyprus Croatia China Turkey Belarus Estonia India Germany Korea Russia Bulgaria China (Taiwan) 1809 Physicists and Engineers 31 Countries 144 Institutions Associated Institutes Number of Scientists Number of Laboratories 36 5

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery6 CERN LHC site CMS Atlas LHCb ALICE

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery7 High Energy Physics at the LHC “Compact” Muon Solenoid at the LHC (CERN) Smithsonian standard man

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery8 Particle Proton  Proton 2835 bunch/beam Protons/bunch10 11 Beam energy7 TeV (7x10 12 ev) Luminosity10 34 cm  2 s  1 Crossing rate40 MHz (every 25 nsec) Collision rate~10 9 Hz Parton (quark, gluon) Proton Selection: 1 in l l jet Bunch New physics rate ~ 10  5 Hz Collisions at LHC (2007  ?) (Average ~20 Collisions/Crossing)

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery9 HEP Data è Scattering is principal technique for gathering data  Collisions of beam-beam or beam-target particles  Typically caused by a single elementary interaction  But also background collisions  obscures physics è Each collision generates many particles: “Event”  Particles traverse detector, leaving electronic signature  Information collected, put into mass storage (tape)  Each event is independent  trivial computational parallelism è Data Intensive Science  Size of raw event record: 20KB  1MB  10 6  10 9 events per year  0.3 PB per year (2001)BaBar (SLAC)  1 PB per year (2005)CDF, D0 (Fermilab)  5 PB per year (2007)ATLAS, CMS (LHC)

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery10 Data Rates: From Detector to Storage Level 1 Trigger: Special Hardware 40 MHz ~1000 TB/sec 75 KHz 75 GB/sec 5 KHz 5 GB/sec Level 2 Trigger: Commodity CPUs 100 Hz 100 MB/sec Level 3 Trigger: Commodity CPUs Raw Data to storage Physics filtering

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery11 LHC Data Complexity è “Events” resulting from beam-beam collisions:  Signal event is obscured by 20 overlapping uninteresting collisions in same crossing  CPU time does not scale from previous generations

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery12 40M events/sec, selectivity: 1 in Example: Higgs Decay into 4 Muons

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery Physicists 150 Institutes 32 Countries LHC Computing Challenges è Complexity of LHC environment and resulting data è Scale: Petabytes of data per year (100 PB by ~2010) Millions of SpecInt95s of CPU è Geographical distribution of people and resources

Transatlantic Net WG (HN, L. Price) Tier0 - Tier1 BW Requirements [*] è [*] Installed BW in Mbps. Maximum Link Occupancy 50%; work in progress

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery15 Hoffmann LHC Computing Report 2001 Tier0 – Tier1 link requirements (1) Tier1  Tier0 Data Flow for Analysis Gbps (2) Tier2  Tier0 Data Flow for Analysis Gbps (3) Interactive Collaborative Sessions (30 Peak) Gbps (4) Remote Interactive Sessions (30 Flows Peak) Gbps (5) Individual (Tier3 or Tier4) data transfers 0.8 Gbps Limit to 10 Flows of 5 Mbytes/sec each TOTAL Per Tier0 - Tier1 Link Gbps  Corresponds to ~10 Gbps Baseline BW Installed on US-CERN Link  Adopted by the LHC Experiments (Steering Committee Report)

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery16 LHC Computing Challenges è Major challenges associated with:  Scale of computing systems  Network-distribution of computing and data resources  Communication and collaboration at a distance  Remote software development and physics analysis Result of these considerations: Data Grids

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery17 Tier0 CERN Tier1 National Lab Tier2 Regional Center (University, etc.) Tier3 University workgroup Tier4 Workstation Global LHC Data Grid Hierarchy Tier 1 T Tier 0 (CERN) Key ideas: è Hierarchical structure è Tier2 centers è Operate as unified Grid

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery18 Example: CMS Data Grid Tier2 Center Online System CERN Computer Center > 20 TIPS USA Center France Center Italy Center UK Center Institute Institute ~0.25TIPS Workstations, other portals ~100 MBytes/sec 2.5 Gbits/sec Mbits/sec Bunch crossing per 25 nsecs. 100 triggers per second Event is ~1 MByte in size Physicists work on analysis “channels”. Each institute has ~10 physicists working on one or more channels Physics data cache ~PBytes/sec 2.5 Gbits/sec Tier2 Center ~622 Mbits/sec Tier 0 +1 Tier 1 Tier 3 Tier 4 Tier2 Center Tier 2 Experiment CERN/Outside Resource Ratio ~1:2 Tier0/(  Tier1)/(  Tier2) ~1:1:1

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery19 Tier1 and Tier2 Centers è Tier1 centers  National laboratory scale: large CPU, disk, tape resources  High speed networks  Many personnel with broad expertise  Central resource for large region è Tier2 centers  New concept in LHC distributed computing hierarchy  Size  [national lab * university] 1/2  Based at large University or small laboratory  Emphasis on small staff, simple configuration & operation è Tier2 role  Simulations, analysis, data caching  Serve small country, or region within large country

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery20 LHC Tier2 Center (2001) Router FEth FEth Switch GEth Switch Data Server >1 RAID Tape WAN Hi-speed channel

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery21 è Buy late, but not too late: phased implementation  R&D Phase  Implementation Phase  R&D to develop capabilities and computing model itself  Prototyping at increasing scales of capability & complexity 1.4 years 1.2 years 1.1 years 2.1 years Hardware Cost Estimates

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery22 HEP Related Data Grid Projects è Funded projects  GriPhyNUSANSF, $11.9M + $1.6M  PPDG IUSADOE, $2M  PPDG IIUSADOE, $9.5M  EU DataGridEU$9.3M è Proposed projects  iVDGLUSANSF, $15M + $1.8M + UK  DTFUSANSF, $45M + $4M/yr  DataTagEUEC, $2M?  GridPPUKPPARC, > $15M è Other national projects  UK e-Science (> $100M for )  Italy, France, (Japan?)

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery23 (HEP Related) Data Grid Timeline Q2 00 Q3 00 Q4 00 Q1 01 Q2 01 Q3 01 GriPhyN approved, $11.9M+$1.6M Outline of US-CMS Tier plan Caltech-UCSD install proto-T2 Submit GriPhyN proposal, $12.5M Submit iVDGL preproposal EU DataGrid approved, $9.3M 1 st Grid coordination meeting Submit PPDG proposal, $12M Submit DTF proposal, $45M Submit iVDGL proposal, $15M PPDG approved, $9.5M 2 nd Grid coordination meeting iVDGL approved? DTF approved? DataTAG approved Submit DataTAG proposal, $2M

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery24 Coordination Among Grid Projects è Particle Physics Data Grid (US, DOE)  Data Grid applications for HENP  Funded 1999, 2000 ($2M)  Funded ($9.4M)  è GriPhyN (US, NSF)  Petascale Virtual-Data Grids  Funded 9/2000 – 9/2005 ($11.9M+$1.6M)  è European Data Grid (EU)  Data Grid technologies, EU deployment  Funded 1/2001 – 1/2004 ($9.3M)   HEP in common  Focus: infrastructure development & deployment  International scope  Now developing joint coordination framework GridPP, DTF, iVDGL  very soon?

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery25 Data Grid Management

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery26 PPDG BaBar Data Management BaBar D0 CDF Nuclear Physics CMSAtlas Globus Users SRB Users Condor Users HENP GC Users CMS Data Management Nuclear Physics Data Management D0 Data Management CDF Data Management Atlas Data Management Globus Team Condor SRB Team HENP GC

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery27         EU DataGrid Project

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery28 PPDG and GriPhyN Projects è PPDG focus on today’s (evolving) problems in HENP  Current HEP:BaBar, CDF, D0  Current NP:RHIC, JLAB  Future HEP:ATLAS, CMS è GriPhyN focus on tomorrow’s solutions  ATLAS, CMS, LIGO, SDSS  Virtual data, “Petascale” problems (Petaflops, Petabytes)  Toolkit, export to other disciplines, outreach/education è Both emphasize  Application sciences drivers  CS/application partnership (reflected in funding)  Performance è Explicitly complementary

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery29 University CPU, Disk, Users PRIMARY SITE Data Acquisition, Tape, CPU, Disk, Robot Satellite Site Tape, CPU, Disk, Robot Satellite Site Tape, CPU, Disk, Robot University CPU, Disk, Users University CPU, Disk, Users Satellite Site Tape, CPU, Disk, Robot Resource Discovery, Matchmaking, Co-Scheduling/Queueing, Tracking/Monitoring, Problem Trapping + Resolution PPDG Multi-site Cached File Access System

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery30 GriPhyN: PetaScale Virtual-Data Grids Virtual Data Tools Request Planning & Scheduling Tools Request Execution & Management Tools Transforms Distributed resources (code, storage, CPUs, networks) è Resource è Management è Services Resource Management Services è Security and è Policy è Services Security and Policy Services è Other Grid è Services Other Grid Services Interactive User Tools Production Team Individual Investigator Workgroups Raw data source ~1 Petaflop ~100 Petabytes

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery31 Virtual Data in Action è Data request may  Compute locally  Compute remotely  Access local data  Access remote data è Scheduling based on  Local policies  Global policies  Cost Major facilities, archives Regional facilities, caches Local facilities, caches Item request

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery32 GriPhyN Goals for Virtual Data è Transparency with respect to location  Caching, catalogs, in a large-scale, high-performance Data Grid è Transparency with respect to materialization  Exact specification of algorithm components  Traceability of any data product  Cost of storage vs CPU vs networks è Automated management of computation  Issues of scale, complexity, transparency  Complications: calibrations, data versions, software versions, … Explore concept of virtual data and its applicability to data-intensive science

HEP and Data Grids (Aug. 4-5, 2001)Paul Avery33 Data Grid Reference Architecture