Computing at Fermilab David J. Ritchie Computing Division May 2, 2006.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
October 2011 David Toback, Texas A&M University Research Topics Seminar 1 David Toback January 2015 Big Computing and the Mitchell Institute for Fundamental.
The Biggest Experiment in History. Well, a tiny piece of it at least… And a glimpse 12bn years back in time To the edge of the observable universe So.
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
Oliver Gutsche - CMS / Fermilab Analyzing Millions of Gigabyte of LHC Data for CMS - Discover the Higgs on OSG.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
HEP Prospects, J. Yu LEARN Strategy Meeting Prospects on Texas High Energy Physics Network Needs LEARN Strategy Meeting University of Texas at El Paso.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
LHC’s Second Run Hyunseok Lee 1. 2 ■ Discovery of the Higgs particle.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Chapter 4 COB 204. What do you need to know about hardware? 
25 February 2000Tim Adye1 Using an Object Oriented Database to Store BaBar's Terabytes Tim Adye Particle Physics Department Rutherford Appleton Laboratory.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
TechFair ‘05 University of Arlington November 16, 2005.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
Supercomputing in High Energy Physics Horst Severini OU Supercomputing Symposium, September 12-13, 2002.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Fermilab June 29, 2001Data collection and handling for HEP1 Matthias Kasemann Fermilab Overview of Data collection and handling for High Energy Physics.
F Computing at Fermilab Fermilab Onsite Review Stephen Wolbers August 7, 2002.
Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University.
Fermi National Accelerator Laboratory SC2006 Fermilab Data Movement & Storage Multi-Petabyte tertiary automated tape store for world- wide HEP and other.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
Silicon Module Tests The modules are tested in the production labs. HIP is is participating in the rod quality tests at CERN. The plan of HIP CMS is to.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
A Brief Introduction to Quarknet and Particle Physics Mike Strauss.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
High Energy Physics & Computing Grids TechFair Univ. of Arlington November 10, 2004.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
11 November 2010 Natascha Hörmann Computing at HEPHY Evaluation 2010.
FSU Experimental HEP Faculty Todd Adams Susan Blessing Harvey Goldman S Sharon Hagopian Vasken Hagopian Kurtis Johnson Harrison Prosper Horst Wahl.
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
LHC Computing, CERN, & Federated Identities
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
HIGH ENERGY PHYSICS DATA WHERE DATA COME FROM, WHERE THEY NEED TO GO In their quest to understand the fundamental nature of matter and energy, Fermilab.
Introducing High Energy Physics Computing to the Secondary Classroom
Detector building Notes of our discussion
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Dagmar Adamova, NPI AS CR Prague/Rez
The LHC Computing Grid Visit of Her Royal Highness
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Southwest Tier 2.
LHC Collisions.
Using an Object Oriented Database to Store BaBar's Terabytes
Development of LHCb Computing Model F Harris
The LHC Computing Grid Visit of Professor Andreas Demetriou
Presentation transcript:

Computing at Fermilab David J. Ritchie Computing Division May 2, 2006

May 2006David J. Ritchie:CD-doc-800-v1 2 Computing At Fermilab is Essential To: Control the Accelerator. Record the Detector information from the Collisions. Reconstruct the Particles’ Parameters. Globally Distribute the Data and Calibrations. Physics results from the Particles. Estimate the Errors. Publish our Discoveries. Simulate everything to check the answers. Discuss and Analyze across Physicists Worldwide

May 2006David J. Ritchie:CD-doc-800-v1 3 Computing At Fermilab is At the Edge Respond to Accelerator time scales of Pico- Seconds (10** -12 ) –Special-purpose Custom Built “Processors” with DSPs etc Collect data at Gigahertz (10**9 words/second). –Use specially built parallel optic fibers from DSPs to 100MByte buffers. Filter information in real time –~150 PCs operating in parallel to read and compute on the data in real time before. Record data to tape at 20MBytes/sec all the time the accelerator is working. … and …

May 2006David J. Ritchie:CD-doc-800-v1 4 Computer Security The Dark Side of the Force The Internet is now ~100 Million computers/users: “one in a million” events happen every day. With the net ‘somewhere on the net’ is always just next door. We are learning how to live in a world where not everyone is friendly -- but most are. A recent demonstration of an unpatched Mac showed it was hacked in 2 minutes. We use automated scanning software to detect problems and always follow up on any inkling of a problem.

May 2006David J. Ritchie:CD-doc-800-v1 5 Accelerator Controls: Measuring where the Beam is

May 2006David J. Ritchie:CD-doc-800-v1 6 How Do You Do Science? (elementary school version) 1.Observe phenomena. 2.Develop a hypothesis. 3.Use hypothesis to make predictions. 4.Devise experiment to look for predictions. 5.Obtain results demonstrating (or not) predictions. 6.Draw conclusions about correctness of hypothesis. 7.If correct, add to accumulated hypotheses which are the theory. If not, revise hypothesis and … Repeat

May 2006David J. Ritchie:CD-doc-800-v1 7 How Do You Do Science? (Large Scale Science Version) Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan and Design Experiment Build / Install Equipment Acquire / Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat

May 2006David J. Ritchie:CD-doc-800-v1 8 Computing Needs Phenomena and Hypothesis Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: access previous results PC’s (Windows, Linux), UNIX, Macs Printers – color and B/W World-wide networking MSWord, TeX, Acrobat Web preprint repositories –Arxiv.org at Cornell U.Arxiv.org –Spires at Stanford U.Spires

May 2006David J. Ritchie:CD-doc-800-v1 9 Computing Needs Organize Collaboration Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: PC’s (Windows, Linux), UNIX, Macs World-wide networking Web Pages –DZero “Top Group”DZero “Top Group” –MINOSMINOS

May 2006David J. Ritchie:CD-doc-800-v1 10 Computing Needs Propose Experiment Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: with colleagues show the feasibility of a proposal. PC’s (Windows, Linux), UNIX, Macs World-wide networking Web pages –Dark Energy Camera ProjectDark Energy Camera Project –Nova ProposalNova Proposal Simulations…

May 2006David J. Ritchie:CD-doc-800-v1 11 Simulations are done throughout the experiment lifetime -- before, during, after data taking. e.g. CDF Silicon Vertex Detector A 3D modeling program simulating particles in the detector.

May 2006David J. Ritchie:CD-doc-800-v1 12 Computing Needs Get Approved and Funded Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Acquire / Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: Project Management Funds come from DOE, NSF, NASA, Congress, Taxpayer—YOU. PC’s (Windows, Linux, Macs) Printers – color and B/W World-wide networking MSProject, Power Point… Proposal review… –NUMI Proposal ReviewNUMI Proposal Review

May 2006David J. Ritchie:CD-doc-800-v1 13 Computing Needs Plan / Design, Build / Install Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Acquire / Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: simulation, sensor design, engineering software…, Unusual sensors Custom electronics Computing at its core: microprocessors, silicon logic

May 2006David J. Ritchie:CD-doc-800-v1 14 Unusual Sensors

May 2006David J. Ritchie:CD-doc-800-v1 15 Computing Needs Acquire / Record / Store Data Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Acquire / Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: Acquire, record, store Use sensors and custom electronics

May 2006David J. Ritchie:CD-doc-800-v1 16 Acquire, Record, Store Data

May 2006David J. Ritchie:CD-doc-800-v1 17 Recording Data Record data (“events”) –The sensors and custom electronics convert the voltages and currents into 300,000 bytes of 1’s and 0’s: Now, we are in the computer realm. –Get an event every few hundred microseconds. –Raw data rate — hundred’s of GB/second. Either trigger on interesting events or filter out the uninteresting ones – ~10 MB/s. What gets through, one records to disk and eventually to tape.

May 2006David J. Ritchie:CD-doc-800-v1 18 Computing Needs Analyze Data Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Acquire / Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: Reconstruct 1’s and 0’s into tracks and particle identification Do physics analysis on resulting collections of tracks Means: PC farms, lots and lots Disk backed up by tape with robotic tape mounts Enstore Usage.Enstore Usage.

May 2006David J. Ritchie:CD-doc-800-v Programmers to make 1,000,000 lines of code to get to the results (ack. L Sexton-Kennedy)

May 2006David J. Ritchie:CD-doc-800-v1 20 Computing Needs Obtain Results Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: Do additional physics analysis. Means: Use analysis farms with their disk caching capabilities to speed data analysis. CDF Usage

ROOT An object oriented HEP analysis framework.

May 2006David J. Ritchie:CD-doc-800-v1 22 A Framework Are often histograms Results CDF Tops the Top World Average provides utilities and services.

May 2006David J. Ritchie:CD-doc-800-v1 23 And in AstroPhysics

May 2006David J. Ritchie:CD-doc-800-v1 24 To Public Data Release

May 2006David J. Ritchie:CD-doc-800-v1 25 Computing Needs Publish Conclusions Identify Phenomena Develop Hypothesis Organize Collaboration Propose Experiment Get Approved Obtain Funding Plan / Design Experiment Build / Install Equipment Record / Store Data Analyze Data Obtain Results Publish Conclusions Each stage of scientific process has different computing needs. Repeat Need: Show evidence of progress in scientific work. Means: Publish papers and conferences.Publish papers and conferences. Top Quark Mass

May 2006David J. Ritchie:CD-doc-800-v1 26 Mass of the Top Quark

May 2006David J. Ritchie:CD-doc-800-v1 27 Vital Statistics FNAL has >10,000 network connected devices They are connected by >1,000 miles of cabling There is more than 200TB of disk spinning now We have over 4 PB of data on tape Half of it is less than 2 years old

May 2006David J. Ritchie:CD-doc-800-v1 28 Moore’s Law Data Volume doubles every 2.4 years Lots of Data!

May 2006David J. Ritchie:CD-doc-800-v Petabytes of Data stored in the Fermilab Robots 4 Petabytes = 4,000 TB = 4,000,000 Gigabytes or about 100,000 Disks on your PC. Equivalent to a stack of CDs nearly 10 times as high as the Eiffel Tower Data In Fermilab Robots Eiffel Tower

May 2006David J. Ritchie:CD-doc-800-v1 30 Computing Facilities The activities map into facilities at the laboratory: –Networks –Mass Storage robotics / tape drives –Large computing farms –Databases –Operations –Support

May 2006David J. Ritchie:CD-doc-800-v1 31 D0 computing systems MBps 100+ MBps 400+ MBps “data-logger” “farm” “central-analysis” Enstore Mass Storage System “linux-analysis -clusters” “linux-build-cluster” “clueD0” ~100 desktops “d0-test” and “sam-cluster”

May 2006David J. Ritchie:CD-doc-800-v1 32 No Data Center is an Island Run II -- D0 planned for Regional Analysis Centers (RAC’s) since 1998: –Distributed analysis around the globe (arbitrary imagined distribution) UO UA CINVESTAV Rice FSU LTU UTA

May 2006David J. Ritchie:CD-doc-800-v1 33 And for the LHC in 2007 the CMS Experiment - example of a global community grid Germany Taiwan UK Italy Data & jobs moving locally, regionally & globally within CMS grid. Transparently across grid boundaries from campus to the world. Florida CERN Caltech Wisconsin UCSD France Purdue MIT UNL

May 2006David J. Ritchie:CD-doc-800-v1 34 US CMS Computing Fermilab is the host lab of U.S. CMS experiment which will begin taking data at CERN in Geneva, Switzerland in ~2007 Fermilab hosts the project management for the U.S. CMS Software and Computing Program in DOE U.S. Physicists will participate in this research.

May 2006David J. Ritchie:CD-doc-800-v1 35 The US CMS center at FNAL transfers data to 39 sites worldwide in CMS global Xfer challenge. Peak Xfer rates of ~5Gbps are reached. CMS Data Xfer FNAL to World

May 2006David J. Ritchie:CD-doc-800-v1 36 ~45 Sites Working together as a Coherent Computing Facility ~20 Research Groups Sharing the Sites to Run Compute Intensive Jobes 2500 jobs

May 2006David J. Ritchie:CD-doc-800-v1 37 Collaborations across Continents and across Disciplines coarse-grained particle ion channel simulator based on the Boltzmann Transport Monte Carlo methodology.

May 2006David J. Ritchie:CD-doc-800-v1 38 The Grid is for Everyone. How will it be used?

May 2006David J. Ritchie:CD-doc-800-v1 39 Acknowledgements As always in a laboratory, one gains from the work of one’s colleagues. This talk is no exception. I would especially like to complain about Ruth Pordes who was meant to give this talk!.