Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University.

Slides:



Advertisements
Similar presentations
Your university or experiment logo here What is it? What is it for? The Grid.
Advertisements

Particle physics – the computing challenge CERN Large Hadron Collider –2007 –the worlds most powerful particle accelerator –10 petabytes (10 million billion.
HEPiX Edinburgh 28 May 2004 LCG les robertson - cern-it-1 Data Management Service Challenge Scope Networking, file transfer, data management Storage management.
October 2011 David Toback, Texas A&M University Research Topics Seminar 1 David Toback January 2015 Big Computing and the Mitchell Institute for Fundamental.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
The Biggest Experiment in History. Well, a tiny piece of it at least… And a glimpse 12bn years back in time To the edge of the observable universe So.
Oliver Gutsche - CMS / Fermilab Analyzing Millions of Gigabyte of LHC Data for CMS - Discover the Higgs on OSG.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
Alain Romeyer - Dec Grid computing for CMS What is the Grid ? Let’s start with an analogy How it works ? (Some basic ideas) Grid for LHC and CMS.
Nordic Data Grid Facility NDGF – Paula Eerola, paula.eerola [at] hep.lu.se paula.eerola [at] hep.lu.sepaula.eerola [at] hep.lu.se 1st Iberian.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Other servers Java client, ROOT (analysis tool), IGUANA (CMS viz. tool), ROOT-CAVES client (analysis sharing tool), … any app that can make XML-RPC/SOAP.
welcome undergraduate physicists High Energy
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
HEP Prospects, J. Yu LEARN Strategy Meeting Prospects on Texas High Energy Physics Network Needs LEARN Strategy Meeting University of Texas at El Paso.
October 2011 David Toback, Texas A&M University Research Topics Seminar 1 David Toback Texas A&M University Research Topics Seminar September 2012 Cosmology.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Dark Matter Masses of Galaxies Gravity and Light Black Holes What is Dark Matter?
Copyright © 2010 Platform Computing Corporation. All Rights Reserved.1 The CERN Cloud Computing Project William Lu, Ph.D. Platform Computing.
RomeWorkshop on eInfrastructures 9 December LCG Progress on Policies & Coming Challenges Ian Bird IT Division, CERN LCG and EGEE Rome 9 December.
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
CERN TERENA Lisbon The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
The Large Hadron Collider is the world's largest and highest-energy particle accelerator.
1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
Recreating the Big Bang with the World’s Largest Machine Prof Peter Watkins Head of Particle Physics Group The University of Birmingham Admissions Talk.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
…building the next IT revolution From Web to Grid…
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
Working Together Scientific Collaboration or Conspiracy?
Your university or experiment logo here What is it? What is it for? The Grid.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
Brookhaven Science Associates U.S. Department of Energy USATLAS Tier 1 & 2 Networking Meeting Scott Bradley Manager, Network Services 14 December 2005.
Outline  Higgs Particle Searches for Origin of Mass  Grid Computing  A brief Linear Collider Detector R&D  The  The grand conclusion: YOU are the.
Heritage The first electron synchrotron in Europe was constructed in Glasgow in the 1950s E = 300 MeV.
Heritage The first electron synchrotron in Europe was constructed in Glasgow in the 1950s E = 300 MeV.
LHC Computing, CERN, & Federated Identities
1 (Brief) Introductory Remarks On Behalf of the U.S. Department of Energy ESnet Site Coordinating Committee (ESCC) W.Scott Bradley ESCC Chairman
August 28, 2003APAN, Logistical Networking WS DiDaS Distributed Data Storage Ludek Matyska Masaryk University, Institute of Comp. Sci. and CESNET, z.s.p.o.
05 Novembre years of research in physics European Organization for Nuclear Research.
Storage Management on the Grid Alasdair Earl University of Edinburgh.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Collaborative Research Projects in Australia: High Energy Physicists Dr. Greg Wickham (AARNet) Dr. Glenn Moloney (University of Melbourne) Global Collaborations.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
Hall D Computing Facilities Ian Bird 16 March 2001.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
European Organization for Nuclear Research
CERN presentation & CFD at CERN
Understanding the nature of matter -
LHC DATA ANALYSIS INFN (LNL – PADOVA)
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
LHC Collisions.
Where do we go from here? Dark Energy
CERN, the LHC and the Grid
High Energy Physics at UTA
High Energy Physics at UTA
Presentation transcript:

Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University

Outline How Strange is the Universe? 5 Modern Mysteries. In trying to resolve these mysteries, particle physicists face a significant data logistics problem. Solution should be flexible enough to encourage the creative approaches that will maximize productivity. REDDnet breaks “data-tethered” compute model, allows unfettered access w/o strong central control.

Is the Universe Even Stranger Than We Have Imagined? One piece of evidence: rotational velocities of stars in galaxies Pick a star, how fast is it moving around galactic center? Mass of galaxy is much, much larger than you get by counting the stars in the galaxy 1st Year Physics!

We Don’t Know What The Majority of Matter in the Universe Is. This “extra” matter is 90% of the Universe! Conventional explanations have mostly been ruled out – Planets, dust, … Most of the matter in the Universe is probably an exotic form of matter — heretofore unknown! But there is a good chance particle physicists will make some soon at the LHC at CERN! ~10% normal matter 90% “other” matter

5 Mysteries for a New Millennium What is the majority of matter in the universe made of? Does space have more than three dimensions? Where is all the anti-matter created by the Big Bang? What is this bizarre thing called “Dark Energy?” Why do things have mass?

Answering These Questions Presents Many Challenges… Experiments require significant infrastructure, large collaborations 2500 Physicists! CERN Large Hadron Collider: 2007 Start 27 km tunnel in Switzerland & France (100 m below ground) CMS

Petascale Computing Required 2008: ~50,000 8 GHz P4s CMS will generate Petabytes of data per year and require Petaflops of CPU… But physics is done in small groups, geographically distributed

Distributed Resources, People Why Distributed Resources? Sociology Politics Funding To maximize the quality and rate of scientific discovery, all physicists must have equal ability to access and analyze the experiment's data… CMS Collaboration: >37 Countries, >163 Institutes

LHC Data Grid Hierarchy Tier 1 Online System CERN Center PBs of Disk; Tape Robot FNAL Tier1 IN2P3 Tier1 INFN Tier1 RAL Tier1 Institute Workstations/Laptops ~ MBs 10 Gbps 1 to 10 Gbps ~PByte/sec Gbps Tier2 Center 1-10 Gbps Tier 0 +1 Tier 3 Tier 4 Caltech Tier2 Tier 2 Experiment >10 Tier1 and ~100 Tier2 Centers UERJ Tier2 Physics data cache Vanderbilt Tier3 The small Analysis Groups doing the physics: work at the Tier 3/4 Level.

Data Logistics Yin and Yang Uncertainty reigns at the most important level — where the physics will get done. Physicists will evolve novel use cases that will not jive with expectations or any plans/rules/edicts. High Level Control Infrastructure Ready?TestedUse Cases Tier 0 Strong, Centralized MostMuchUnderstood Tier 4AnarchyLittle/NoneNone?????

Use Cases: What we Do Know Physicists will: need access to TB Data Sets for short term periods. run over this data many times, refining, improving their analysis. use local computing resources where they may not have much storage available. make “opportunistic use” of compute resources at Tier 3 sites and Grid sites. perform “production runs” at Tier 2 sites.

REDDnet at Tier 3 Opportunistic computing vs data-tethered computing –CMS has no formal solution for Tier 3 storage –Compute on resources — even those where data not hosted On-demand working storage –improve data logistics –Acts local — familiar user tools Demonstrate at a Tier 3 –Performance –Reliability –… and convenience

REDDnet SC06 Depots Near Term Plan of Work Provide T3 scratch space Host/mirror popular datasets on REDDnet Participate in Data and Service Challenges –Summer 07 Challenge Starting Soon –Network and Data Transfer Load tests Integrate with existing CMS tools Develop a Tier 3 Analysis environment –Initial small test community –Test with individual analyses –Run on the Grid