Download presentation
Presentation is loading. Please wait.
Published byHerbert Robinson Modified over 9 years ago
1
Computing Sciences NERSC High Energy and Nuclear Physics Computing Group Craig E. Tull HCG/NERSC/LBNL 2005 Science Colloquium Series DOE - August 23, 2005 NERSC High Energy and Nuclear Physics Computing Group Craig E. Tull HENPCG/NERSC/LBNL 2006 Director’s Review of Computing LBNL - September 22, 2006
2
Computing Sciences HENP Computing Group Group Leader: Craig Tull 1 Senior Scientist, 4 Scientists, 6 Engineers, 1 Postdoc Embedded software professionals w/ science backgrounds Provide computing systems & support for science collaborations Mostly software focus, but with a track record for integrated software and hardware systems Scientists focus on science and requirements on the software rather than detailed design or implementation. Leave non-scientific code to computing professionals with expertise and time to apply software process. Management and Leadership Roles Software and Technology Development Software Engineering Best Practices and Support
3
Computing Sciences HENP Environment Large, distributed collaborations are the norm ~2000 scientists, from ~150 institutions in ~50 countries Scientists require equal access to data and resources Very long time duration of projects & software Detectors take 5-10 years to design and build Detectors have an operational lifetime of 5-20 years 10 to 30 year Project lifetimes Strong focus on robust, maintainable software supporting graceful evolution of components & protocols Commodity computing (Intel, Linux) Polite parallelism/Partitioning of calculations Data Intensive (100's TB => 1,000's TB) The World is Networked Scientists are developers and not just users Many skill levels from Wizard to Neophyte Issues of scaling are sociological as well as technical
4
Computing Sciences HENPC Group Role Within a Project Software professionals with scientific background as full collaborators Establishment of a software development process Adapt CSE methodologies and processes to HENP environment Object Oriented, Architectural Aspects of Unified Software Development Process (USDP) and Extreme Programming (XP) Design and implementation of software development infrastructure Code repository, release build, test, and distribution systems Design and implementation of major software components Control frameworks Online experiment control Data persistency frameworks Physics toolkits Training and mentoring Tutorials, code guidelines, requirement/design/code reviews, etc. Re-engineering of existing designs Provide expertise to improve robustness, performance, maintainability
5
Computing Sciences HENPC Group Role across Projects Promote a common culture Best practices, open source, architecture, code reuse Develop and integrate tools that support these best practices Explore and integrate new technologies Object Oriented Database Systems CORBA based distributed systems GRID integration C++, Java, Python J2EE, JBoss, JMX Generate an Institutional knowledge base User Requirements Architectural decomposition Components Leverage coupling between NERSC and Physical Sciences at LBNL
6
Computing Sciences HENPC Projects (2004-2006)
7
Computing Sciences Experiments & Projects (2004-2006) ATLASLHC accelerator at CERN, Geneva Software lead, Chief Architect, Core software & Athena Gaudi framework BaBarPEP-II collider at SLAC, Stanford Conditions Database, IceCubeNeutrino detector at South Pole Chief Architect, Experiment Control, Build Environment, Offline Framework MajoranaNeutrinoless double beta decay, LDRD GEANT4 build system, GEANT4 Geometry Database SNAPSupernova satellite Java Simulation Framework GUPFSGlobal Unified Parallel File System Management and Scientific Liason SNFSuper Nova Factory, Telescope, Hawaii Lisp-based Observation Scheduling Software PPDGParticle Physics Data Grid Replica catalogs technical survey, Security & VO roles
8
Computing Sciences The ATLAS Experiment: A large HEP project
9
Computing Sciences LHC √ s = 14 TeV (7 times higher than Tevatron/Fermilab) → search for new massive particles up to m ~ 5 TeV L design = 10 34 cm -2 s -1 (>10 2 higher than Tevatron/Fermilab) → search for rare processes with small (N = L ) ALICE : heavy ions ATLAS and CMS : pp, general purpose ATLAS and CMS : pp, general purpose 27 km ring used for e + e - LEP machine in 1989-2000 Start : Summer 2007 pp LHCb : pp, B-physics
10
Computing Sciences ATLAS Length : ~ 46 m Radius : ~ 12 m Weight : ~ 7000 tons ~ 10 8 electronic channels ~ 3000 km of cables Tracking (| |<2.5, B=2T) : -- Si pixels and strips -- Transition Radiation Detector (e/ separation) Calorimetry (| |<5) : -- EM : Pb-LAr -- HAD: Fe/scintillator (central), Cu/W-LAr (fwd) Muon Spectrometer (| |<2.7) : air-core toroids with muon chambers
11
Computing Sciences ATLAS Collaboration 34 Countries 151 Institutions 1770 Scientific Authors Albany, Alberta, NIKHEF Amsterdam, Ankara, LAPP Annecy, Argonne NL, Arizona, UT Arlington, Athens, NTU Athens, Baku, IFAE Barcelona, Belgrade, Bergen, Berkeley LBL and UC, Bern, Birmingham, Bonn, Boston, Brandeis, Bratislava/SAS Kosice, Brookhaven NL, Bucharest, Cambridge, Carleton, Casablanca/Rabat, CERN, Chinese Cluster, Chicago, Clermont-Ferrand, Columbia, NBI Copenhagen, Cosenza, INP Cracow, FPNT Cracow, Dortmund, JINR Dubna, Duke, Frascati, Freiburg, Geneva, Genoa, Glasgow, LPSC Grenoble, Technion Haifa, Hampton, Harvard, Heidelberg, Hiroshima, Hiroshima IT, Indiana, Innsbruck, Iowa SU, Irvine UC, Istanbul Bogazici, KEK, Kobe, Kyoto, Kyoto UE, Lancaster, Lecce, Lisbon LIP, Liverpool, Ljubljana, QMW London, RHBNC London, UC London, Lund, UA Madrid, Mainz, Manchester, Mannheim, CPPM Marseille, Massachusetts, MIT, Melbourne, Michigan, Michigan SU, Milano, Minsk NAS, Minsk NCPHEP, Montreal, FIAN Moscow, ITEP Moscow, MEPhI Moscow, MSU Moscow, Munich LMU, MPI Munich, Nagasaki IAS, Naples, Naruto UE, New Mexico, Nijmegen, BINP Novosibirsk, Ohio SU, Okayama, Oklahoma, LAL Orsay, Oslo, Oxford, Paris VI and VII, Pavia, Pennsylvania, Pisa, Pittsburgh, CAS Prague, CU Prague, TU Prague, IHEP Protvino, Ritsumeikan, UFRJ Rio de Janeiro, Rochester, Rome I, Rome II, Rome III, Rutherford Appleton Laboratory, DAPNIA Saclay, Santa Cruz UC, Sheffield, Shinshu, Siegen, Simon Fraser Burnaby, Southern Methodist Dallas, NPI Petersburg, Stockholm, KTH Stockholm, Stony Brook, Sydney, AS Taipei, Tbilisi, Tel Aviv, Thessaloniki, Tokyo ICEPP, Tokyo MU, Tokyo UAT, Toronto, TRIUMF, Tsukuba, Tufts, Udine, Uppsala, Urbana UI, Valencia, UBC Vancouver, Victoria, Washington, Weizmann Rehovot, Wisconsin, Wuppertal, Yale, Yerevan
12
Computing Sciences ATLAS Computing Characteristics Large, complex detector ~10 8 channels Long lifetime Project started in 1992, first data in 2007, last data 2027? 320 MB/sec raw data rate ~3 PB/year Large, geographically dispersed collaboration 1770 people, 151 institutions, 34 countries Most are, or will become, software developers Programming abilities range from Wizard to Neophyte Scale and complexity reflected in software ~1000 packages, ~8000 C++ classes, ~5M lines of code ~70% code is algorithmic (written by physicists) ~30% infrastructure, framework (written by sw engineers) HENPC responsible for large portion of this software Provide robustness but plan for evolution Requires enabling technologies Requires management & coherency
13
Computing Sciences Software Methodology Object-Oriented using C++ as programming language Some wrapped FORTRAN and Java Python as interactive & configuration language Heavy use of components behind abstract interfaces Support multiple implementations Robustness & evolution Lightweight development process Emphasis on automation and feedback rather than very formal process Previous attempt at developing a software system had failed due to a too rigorous software process decoupled from developers Make it easy for developers to do the “right thing” Some requirements/design reviews Sub-system “functionality” reviews 2 weeks each Focus on client viewpoint
14
Computing Sciences Event Store Athena/Gaudi Component Model Converter Algorithm StoreGateSvc Persistency Service Data Files Algorithm StoreGateSvc Persistency Service Data Files Detector Store Message Service JobOptions Service Particle Prop. Service Other Services Histogram Service Application Manager Converter Event LoopMgr Auditors Persistency Service Data Files Histogram Store Sequencer
15
Computing Sciences Athena Components Algorithms Provide basic per-event processing Share a common interface (state machine) Sequencer is type of Algorithm that sequences/filters other Algorithms Tools More specialized but more flexible than Algorithms Services E.g. Particle Properties, Random Numbers, Histogramming Data Stores (blackboards) Data registered by one Algorithm/Tool can be retrieved by another Multiple stores handle different lifetimes (per event, per job, etc.) Stores accessed via Services (e.g. StoreGateSvc) Converters Transform data from one representation to another e.g. transient/persistent Properties Adjustable parameters of components Can be modified at run-time to configure job
16
Computing Sciences ATLAS Computing Organization
17
Computing Sciences HENPC Group within ATLAS David Quarrie Software Project Lead Paolo Calafiura Chief Architect Core Services Group Lead Wim Lavrijsen Python configuration and support tools Charles Leggett Athena/Gaudi framework support Martin Woudstra Integration of Athena with production system
18
Computing Sciences FY06-07 Major Activities Computing support for ATLAS Detector Commissioning Electronics & detector calibrations Cosmic ray tests Computing System Commissioning (CSC) Commissioning of the Computing and Software system itself High Level Trigger Large Scale Tests Offline software components used in HLT Athena framework Tracking & calorimetry algorithms
19
Computing Sciences ATLAS (some highlights) Management & Leadership Roles ATLAS Software Project Lead: David Quarrie ATLAS Chief Architect: Paolo Calafiura Previously D.Quarrie US ATLAS Core Software Manager: Paolo Calafiura Previously D.Quarrie, C.Tull Software and Technology Development Athena Control & Analysis Framework PyROOT: Introspection-driven ROOT Python interface StoreGate: Object Transient Store Software Engineering Best Practices and Support Nightly Build & Release Campaign Dozens of tutorials and extensive documentation ASK (Athena Startup Kit): Robust GUI for build system Hephaestus: Low-overhead Memory Profiler
20
Computing Sciences IceCube Management & Leadership Roles Software Architect: Simon Patton Experiment Control: Simon Patton Software and Technology Development Ice Tray: Component-based Analysis Framework JBoss/JMX Control Architecture Hierarchical State Machine Web Portal interface Software Engineering Best Practices and Support BFD (Baseline File Development): UML based develop, build, release system. Tutorials & Developer Support
21
Computing Sciences IceCube Computing System Architecture and Development (Simon Patton) Strong coherent vision for all IceCube software. Laying out "best practices" to follow to ensure good code. Development environment, tools supporting best practices. Experiment Control (Simon Patton, Chris Day, Akbar Mohktarani) Layered State Machine control of components of data flow Uses J2EE, JBoss/JMX Heirarchical State Machine
22
Computing Sciences SNAP Management & Leadership Roles Computing Framework Architect: Gary Kushner SNAP Collaboration Systems Manager Group Represent SNAP computing w/ Bill Carithers Software and Technology Development Computing Framework & Monte Carlo (Java-based) Simulate the Universe being observed Simulate the Instrumentation and Detecton Process Simulate the Extraction of Cosmological Parameter from Mission Data Software Engineering Best Practices and Support Ant-based build environment Shrink-wrapped deployment package (out-of-box experience) Redesign/implementation of Physics Codes: up to X15 speedup
23
Computing Sciences Supernova Mission Simulator With our Monte Carlo: —Have simulated SNAP with detector characteristics and observing program —Have simulated other potential experiments including ground- based instruments —Use state-of-the-art SNe models that simulate SNe population drift with redshift —Included systematic effects and calibration errors —Can generate error ellipses for cosmological parameters —Can optimize SNAP
24
Computing Sciences High Level CS Architecture
25
Computing Sciences BaBar Management & Leadership Roles Previously Chief Architect: David Quarrie Previously Database Head: Simon Patton Software and Technology Development Conditions DataBase (CDB): Igor Gaponenko & Akbar Mokhtarani Historically: Object Oriented Database (Objectivity) Offline & Online software & General applications Software Engineering Best Practices and Support Refactorizing all BaBar database applications to newer persistency technologies (ROOT I/O & MySQL) Expert-level support for distributed database management & tools Consultation on Database Software
26
Computing Sciences CDB Concepts : Scope & Ownership Diagram DATABASE ORIGIN P-LAYOUT (2-D SPACE) PARTITION VIEW FOLDER CONFIGURATION PHYSICAL CONDITION (2-D SPACE) REVISION VISIBLE INTERVAL ORIGINAL INTERVAL USER DATA OBJECT MIR uses owns provides scope for
27
Computing Sciences Majorana Software and Technology Development Centralized MySQL database for MaGe (GEANT4) material properties and geometry Schema, API, implementation, and support Software Engineering Best Practices and Support Unified software management to ensure code integrity across participating institutions Incremental build and test system General application and development support Geant4, Root, CLHEP, Dawn, etc
28
Computing Sciences “MaGe” Simulation Package. Framework uses powerful object-oriented and abstraction capabilities of C++ and STL for flexibility Gerda-related detector geometries Majorana-related detector geometries MaGe Geant 4/ ROOT Event Generators Common geometries Physics processes Majorana-related output Gerda-related output
29
Computing Sciences MaGe Activities Previous Activities: Characterization of radioactive backgrounds in conceptual design for NuSAG proposal. Interpretation of results from highly segmented detector at MSU. TUNL-FEL Run Charge Distribution in 0 -decay Current Activities: Full characterization of Majorana reference design and optimization of segmentation scheme. Neutron background from muons in rock Alpha contamination on surfaces. Pulse-generation. Gerda Posters at Neutrino04, TAUP05
30
Computing Sciences Challenges & Opportunities Many other science disciplines are growing in size, distribution, and data volume HENP lessons learned should be leveraged Non-HENP techniques of interest to HENP New HENP experiments/projects: SNAP, Majorana, Daya Bay, Gretina "Old" experiments/projects: BaBar, ATLAS, IceCube Lack of any base funding for HENPC: Problems with long-term stability & predictability Great difficulty jump-starting new projects Common use software, libraries & tools a "volunteer" effort
31
Computing Sciences Summary of HENPC Group A small group with involvement in many projects Have had a major impact in the field Leaders in use of Object Oriented programming and software process Leaders in use of Object Oriented Databases Control Frameworks in use by many experiments Not just those we’ve worked directly on Demonstrated ability to leverage our expertise and empower large, dispersed, software teams Demonstrated ability to design and implement large scale scientific computing systems Keeping abreast of modern computing techniques and tools, but delivering concrete, robust software for production use.
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.