Scientific Computing for SLAC Science Bebo White Stanford Linear Accelerator Center October 2006.

Slides:



Advertisements
Similar presentations
1 The Evolution of ESnet (Summary) William E. Johnston ESnet Manager and Senior Scientist Lawrence Berkeley National Laboratory.
Advertisements

Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
Data Analysis in High Energy Physics, Weird or Wonderful? Richard P. Mount Director: Scientific Computing and Computing Services Stanford Linear Accelerator.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Network Policies and Procedures Presentation for DoE Office of Assurance Cybersecurity Review visit to SLAC August 2005.
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
Scientific Computing at SLAC Richard P. Mount Director: Scientific Computing and Computing Services DOE Review June 15, 2005.
HEP Prospects, J. Yu LEARN Strategy Meeting Prospects on Texas High Energy Physics Network Needs LEARN Strategy Meeting University of Texas at El Paso.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
CHEP 2004 September 2004Richard P. Mount, SLAC Huge-Memory Systems for Data-Intensive Science Richard P. Mount SLAC CHEP, September 29, 2004.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Fabric Management for CERN Experiments Past, Present, and Future Tim Smith CERN/IT.
25 February 2000Tim Adye1 Using an Object Oriented Database to Store BaBar's Terabytes Tim Adye Particle Physics Department Rutherford Appleton Laboratory.
Particle Physics and the Grid Randall Sobie Institute of Particle Physics University of Victoria Motivation Computing challenge LHC Grid Canadian requirements.
Division Report Computing Center CHEN Gang Computing Center Oct. 24, 2013 October 24 ,
CERN TERENA Lisbon The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
May Richard P. Mount, SLAC Advanced Computing Technology Overview Richard P. Mount Director: Scientific Computing and Computing Services Stanford.
Data GRID Activity in Japan Yoshiyuki WATASE KEK (High energy Accelerator Research Organization) Tsukuba, Japan
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University.
High Energy Physics Data Management Richard P. Mount Stanford Linear Accelerator Center DOE Office of Science Data Management Workshop, SLAC March 16-18,
Lessons Learned from Managing a Petabyte Jacek Becla Stanford Linear Accelerator Center (SLAC) Daniel Wang now University of CA in Irvine, formerly SLAC.
ATLAS Computing at SLAC Future Possibilities Richard P. Mount Western Tier 2 Users Forum April 7, 2009.
Astronomy, Petabytes, and MySQL MySQL Conference Santa Clara, CA April 16, 2008 Kian-Tat Lim Stanford Linear Accelerator Center.
HEP and Non-HEP Computing at a Laboratory in Transition Richard P. Mount Director: Scientific Computing and Computing Services Stanford Linear Accelerator.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
National HEP Data Grid Project in Korea Kihyeon Cho Center for High Energy Physics (CHEP) Kyungpook National University CDF CAF & Grid Meeting July 12,
PPDGLHC Computing ReviewNovember 15, 2000 PPDG The Particle Physics Data Grid Making today’s Grid software work for HENP experiments, Driving GRID science.
Scientific Computing at SLAC: The Transition to a Multiprogram Future Richard P. Mount Director: Scientific Computing and Computing Services Stanford Linear.
SLAC Status, Les CottrellESnet International Meeting, Kyoto July 24-25, 2000 SLAC Update Les Cottrell & Richard Mount July 24, 2000.
ICFA/SCIC Aug '05 SLAC Site Report 1 SLAC Site Report Les Cottrell, Gary Buhrmaster, Richard Mount, SLAC For ICFA/SCIC meeting 8/22/05
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Xrootd Proxy Service Andrew Hanushevsky Heinz Stockinger Stanford Linear Accelerator Center SAG September-04
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
US ATLAS Tier 1 Facility Rich Baker Deputy Director US ATLAS Computing Facilities October 26, 2000.
HEP and Non-HEP Computing at a Laboratory in Transition Richard P. Mount Director: Scientific Computing and Computing Services Stanford Linear Accelerator.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
GDB meeting - Lyon - 16/03/05 An example of data management in a Tier A/1 Jean-Yves Nief.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
PetaCache: Data Access Unleashed Tofigh Azemoon, Jacek Becla, Chuck Boeheim, Andy Hanushevsky, David Leith, Randy Melen, Richard P. Mount, Teela Pulliam,
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
05/14/04Larry Dennis, FSU1 Scale of Hall D Computing CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental.
Hall D Computing Facilities Ian Bird 16 March 2001.
GridPP10 Meeting CERN June 3 rd 2004
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Understanding the nature of matter -
UK GridPP Tier-1/A Centre at CLRC
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Particle Physics at KISTI
Nuclear Physics Data Management Needs Bruce G. Gibbard
SLAC B-Factory BaBar Experiment WAN Requirements
Using an Object Oriented Database to Store BaBar's Terabytes
Presentation transcript:

Scientific Computing for SLAC Science Bebo White Stanford Linear Accelerator Center October 2006

2 Scientific Computing The relationship between Science and the components of Scientific Computing

3 Drivers for SLAC Computing Computing to enable today’s data- intensive science –Clusters, interconnects, networks, mass storage, etc. Computing research to prepare for tomorrow’s challenges –Massive memory, low latency, petascale databases, detector simulation, etc.

4 SLAC Scientific Computing Science GoalsComputing Techniques BaBar Experiment (winds down ) Measure billions of collisions to understand matter-antimatter asymmetry (why matter exists today) High-throughput data processing, trivially parallel computation, heavy use of disk and tape storage. Intercontinental distributed computing. ATLAS Experiment and Experimental HEP Analyze petabytes of data to understand the origin of mass High-throughput data processing. trivially parallel computation, heavy use of disk and tape storage. Intercontinental distributed computing. Accelerator Science Simulate accelerator behavior before construction and during operation Parallel computation, visual analysis of large data volumes Particle Astrophysics (mainly simulation) Star formation in the early universe, colliding black holes, … Parallel computation (SMP and cluster), visual analysis of growing volumes of data Particle Astrophysics Major Projects (GLAST, LSST …) Analyze terabytes to petabytes of data to understand the dark matter and dark energy riddles High-throughput data processing, very large databases, visualization Photon Science Femtosecond x-ray pulses, “ultrafast” science, structure of individual molecules … High throughput data analysis and large-scale simulation New Architectures for SLAC Science Radical new approaches to computing for Stanford-SLAC data- intensive science Current focus: massive solid-state storage for high-throughput, low- latency data analysis

5 Data Challenge in High Energy Physics 2006 example SLAC Online System Selection and Compression ~10TB/s Raw data written to tape:10MB/s Simulated and derived data: 20 MB/s International network data flow to “Tier A Centers” 50 MB/s (400Mb/s)

6 Tier 1 Online System Event Reconstruction France Germany Institute ~0.25TIPS ~100 MBps ~ Gbps Mbps Physics data cache ~PBps ~ Gbps Tier 0 +1 Tier 3 Tier 4 Tier physicists in 31 countries are involved in this 20- year experiment in which DOE is a major player. Grid infrastructure spread over the US and Europe coordinates the data analysis Analysis Italy FermiLab, USA Data Challenge in High Energy Physics: CERN / LHC High Energy Physics Data 2008 onwards Event Simulation CERN LHC CMS detector 12,500 tons, $700M Gbps

7 Client Disk Server Tape Server SLAC-BaBar Computing Fabric IP Network (Cisco) 120 dual/quad CPU Sun/Solaris ~700 TB Sun RAID arrays (FibreChannel +some SATA) 1700 dual CPU Linux (over 3700 cores) 25 dual CPU Sun/Solaris 40 STK 9940B 6 STK 9840A 6 STK Powderhorn over 1 PB of data HEP-specific ROOT software (Xrootd) + Objectivity/DB object database some NFS HPSS + SLAC enhancements to ROOT and Objectivity server code

8 Used/Required Space

9 ESnet: Source and Destination of the Top 30 Flows, Feb Terabytes/Month Fermilab (US)  WestGrid (CA) SLAC (US)  INFN CNAF (IT) SLAC (US)  RAL (UK) Fermilab (US)  MIT (US) SLAC (US)  IN2P3 (FR) IN2P3 (FR)  Fermilab (US) SLAC (US)  Karlsruhe (DE) Fermilab (US)  Johns Hopkins LIGO (US)  Caltech (US) LLNL (US)  NCAR (US) Fermilab (US)  SDSC (US) Fermilab (US)  Karlsruhe (DE) LBNL (US)  U. Wisc. (US) Fermilab (US)  U. Texas, Austin (US) BNL (US)  LLNL (US) Fermilab (US)  UC Davis (US) Qwest (US)  ESnet (US) Fermilab (US)  U. Toronto (CA) BNL (US)  LLNL (US) CERN (CH)  BNL (US) NERSC (US)  LBNL (US) DOE/GTN (US)  JLab (US) U. Toronto (CA)  Fermilab (US) NERSC (US)  LBNL (US) CERN (CH)  Fermilab (US) DOE Lab-International R&E Lab-U.S. R&E (domestic) Lab-Lab (domestic) Lab-Comm. (domestic)

10 Growth and Diversification Continue shared cluster growth as much as possible Increasing MPI (parallel) capacity and support (astro, accelerator, and more) Grid interfaces and support (Atlas et.al) Large SMPs (Astro) Visualization

11 Research - PetaCache The PetaCache architecture aims at revolutionizing the query and analysis of scientific databases with complex structure –Generally this applies to feature databases (terabytes-petabytes) rather than bulk data (petabytes-exabytes) The original motivation comes from HEP –Sparse (~random) access to tens of terabytes today, petabytes tomorrow –Access by thousands of processors today, tens of thousands tomorrow

12 Latency Ideal

13 Latency Current Reality

14 Latency Practical Goal

15 PetaCache Summary Data-intensive science increasingly requires low-latency access to terabytes or petabytes Memory is one key: –Commodity DRAM today (increasing total cost by ~2x) –Storage-class memory (whatever that will be) in the future Revolutions in scientific data analysis will be another key –Current HEP approaches to data analysis assume that random access is prohibitively expensive –As a result, permitting random access brings much-less-than- revolutionary immediate benefit Use the impressive motive force of a major HEP collaboration with huge data-analysis needs to drive the development of techniques for revolutionary exploitation of an above-threshold machine

16 Research – Very Large Databases 10-year, unique experience with VLDB –Designing, building, deploying, and managing peta-scale production datasets/database – BaBar – 1.4 PB –Assisting LSST (Large Synoptic Survey Telescope) in solving data-related challenges (effort started 4Q 2004)

17 LSST – Data Related Challenges (1/2) Large volumes –7 PB/year (image and catalog data) –500 TB/year (database) Todays VLDBs ~10s TB range High availability –Petabytes -> 10s of 1000s of disks -> daily disk failures Real time requirement –Transient alerts generated in < 60 sec

18 LSST – Data Related Challenges (2/2) Spatial and temporal aspects –Most surveys focus on a single dimension All data made public with minimal delay –Wide range of users – professional and amateur astronomers, students, general public

19 VLDB Work by SCCS Prototyping at SCCS Close collaboration with key MySQL developers Working closely with world-class database gurus

20 Research – Geant4 A toolkit simulating elementary particles passing through and interacting with matter, and modeling the detector apparatus measuring the passage of elementary particles and recording the energy and dose deposition Geant4 is developed and maintained by an international collaboration –SLAC is the second largest center next to CERN

21 Acknowledgements Richard Mount, Director, SCCS Chuck Boeheim, SCCS Randall Melen, SCCS

WWW 2008 April 21-25, 2008 Beijing, China

23 Host Institution and Partners Beihang University –School of Computer Science Tsing-Hua University, Peking University, Chinese Academy of Sciences, … Microsoft Research Asia City Government of Beijing (pending)

24 BICC: Beijing International Convention Center

25 Key Personnel General Chairs: –Jinpeng Huai, Beihang University –Robin Chen, AT&T Labs Conference Vice Chair: –Yunhao Liu, HKUST Local volunteers –6-10 grad students led by Dr. Zongxia Du –In cooperation with John Miller (TBD) IW3C2 Liaison: Ivan Herman PCO: two candidates under consideration

26 Local Organizing Committee Composition of Local Organizing Committee: –Vincent Shen, The HK University of Science and TechnologyThe HK University of Science and Technology –Zhongzhi Shi, Chinese Academy of SciencesChinese Academy of Sciences –Hong Mei, Peking UniversityPeking University –Dianfu Ma, Beihang UniversityBeihang University –Guangwen Yang, Tsinghua UniversityTsinghua University –Hsiao-Wuen Hon, Microsoft Research Asia –Minglu Li, Shanghai Jiao Tong UniversityShanghai Jiao Tong University –Hai Jin, Huazhong University of Science and TechnologyHuazhong University of Science and Technology –… and Chinese Internet/Software/Telecom companies

27