HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Grid Computing for HEP L. E. Price Argonne National Laboratory HEP-CCC Meeting CERN,

Slides:



Advertisements
Similar presentations
1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
Advertisements

U.S. Department of Energy’s Office of Science Basic Energy Sciences Advisory Committee Dr. Daniel A. Hitchcock October 21, 2003
High Performance Computing Course Notes Grid Computing.
Foundations for an LHC Data Grid Stu Loken Berkeley Lab.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Planning for a Western Analysis Facility Richard P. Mount Planning for a Western Analysis FacilityPage 1.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Assessment of Core Services provided to USLHC by OSG.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
CERN TERENA Lisbon The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
Ohio State University Department of Computer Science and Engineering 1 Cyberinfrastructure for Coastal Forecasting and Change Analysis Gagan Agrawal Hakan.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
CHEP 2000 (Feb. 7-11)Paul Avery (Data Grids in the LHC Era)1 The Promise of Computational Grids in the LHC Era Paul Avery University of Florida Gainesville,
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Ruth Pordes, Fermilab CD, and A PPDG Coordinator Some Aspects of The Particle Physics Data Grid Collaboratory Pilot (PPDG) and The Grid Physics Network.
August 26, 1999: MONARC Regional Reps Meeting Harvey Newman (CIT) MONARC Second Regional Centre Representatives Meeting Harvey B. Newman (Caltech) CERN.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
GriPhyN EAC Meeting (Jan. 7, 2002)Carl Kesselman1 University of Southern California GriPhyN External Advisory Committee Meeting Gainesville,
Perspectives on Grid Technology Ian Foster Argonne National Laboratory The University of Chicago.
1 Computing Challenges for the Square Kilometre Array Mathai Joseph & Harrick Vin Tata Research Development & Design Centre Pune, India CHEP Mumbai 16.
The Earth System Grid (ESG) Computer Science and Technologies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
…building the next IT revolution From Web to Grid…
Middleware Camp NMI (NSF Middleware Initiative) Program Director Alan Blatecky Advanced Networking Infrastructure and Research.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
Breakout # 1 – Data Collecting and Making It Available Data definition “ Any information that [environmental] researchers need to accomplish their tasks”
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
DoE NGI Program PI Meeting, October 1999Particle Physics Data Grid Richard P. Mount, SLAC Particle Physics Data Grid Richard P. Mount SLAC Grid Workshop.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
PPDGLHC Computing ReviewNovember 15, 2000 PPDG The Particle Physics Data Grid Making today’s Grid software work for HENP experiments, Driving GRID science.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
Open Science Grid & its Security Technical Group ESCC22 Jul 2004 Bob Cowles
July 26, 1999MONARC Meeting CERN MONARC Meeting CERN July 26, 1999.
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
NORDUnet NORDUnet e-Infrastrucure: Grids and Hybrid Networks Lars Fischer CTO, NORDUnet Fall 2006 Internet2 Member Meeting, Chicago.
U.S. ATLAS Computing Facilities DOE/NFS Review of US LHC Software & Computing Projects Bruce G. Gibbard, BNL January 2000.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
DOE/NSF Quarterly review January 1999 Particle Physics Data Grid Applications David Malon Argonne National Laboratory
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
Hall D Computing Facilities Ian Bird 16 March 2001.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Clouds , Grids and Clusters
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Nuclear Physics Data Management Needs Bruce G. Gibbard
Presentation transcript:

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Grid Computing for HEP L. E. Price Argonne National Laboratory HEP-CCC Meeting CERN, November 12, 1999

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL The Challenge Providing rapid access to event samples and subsets from massive datastores, from 100s of Terabytes in 2000 to 100 Petabytes by Transparent access to computing resources, throughout the U.S., and throughout the World The extraction of small or subtle new physics signals from large and potentially overwhelming backgrounds Enabling access to the data, and to the rest of the physics community, across and ensemble of networks of varying capability and reliability, using heterogeneous computing resources

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Achieving a Balance Proximity of the data to central computing and data handling resources Proximity of frequently accessed data to the users, to be processed in desktops, local facilities, or regional centers Making efficient use of limited network bandwidth; especially transoceanic Making appropriate use of regional and local computing and data handling Involving scientists and students in each world region in the physics analysis

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Need for optimization Meeting the demands of hundreds of users who need transparent access to local and remote data in disk caches and tape stores Prioritizing hundreds to thousands of requests from the local and remote communities Structuring and organizing the data; providing the tools for locating, moving, and scheduling data transport between tape and disk and across networks Ensuring that the overall system is dimensioned correctly to meet the aggregate need

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Science and Massive Datasets Massive dataset generation the new norm in science –High Energy Physics –Nuclear Physics –LIGO –Automated astronomical scans (e.g., Sloan Digital Sky Survey) –The Earth Observing System (EOS) –The Earth System Grid –Geophysical data (e.g., seismic) –Satellite weather image analysis –The Human Brain Project (time series of 3-D images) –Protein Data Bank –The Human Genome Project –Molecular structure crystallography data

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Proposed Solution A data analysis grid for High Energy Physics Tier 1 T CERN T

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Analogy to Computing Grid Because the resources needed to solve complex problems are rarely collocated Topic of intensive CS research for a number of years already Computing (or data) resources from a “plug on the wall”

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Why a Hierarchical Data Grid? Physical –Appropriate resource use  data proximity to users & labs –Efficient network use  local > regional > national > oceanic –Scalable growth  avoid bottlenecks Human –Central lab cannot manage / help / care about 1000s of users –Cleanly separates functionality of different resource types –University/regional computing complements national labs  funding agencies –Easier to leverage resources, maintain control, assert priorities at regional/local level –Effective involvement of scientists and students independently of location

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Logical Steps toward Data Grid Production Basic Research Testbeds Design/Optimization (Pre)                             

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL U.S. Grid Technology Projects LHC, GriPhyN Clipper/ NGI-PPDG                           Apogee PASS/Globus/HENP-GC /MONARC/GIOD/Nile

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL In Progress Laboratory and experiment-specific development, deployment and operation (hardware and software); Tool development in HENP, Computer Science, Industry; The Particle Physics Data Grid: –NGI-funded project aiming (initially) at jump- starting the exploitation of CS and HENP software components to make major improvements in data access. Business as usual

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Proposals being Developed GriPhyN: Grid Physics Networking –Targeted at NSF; –Focus on the long-term university-based grid infrastructure for major physics and astronomy experiments. APOGEE: A Physics-Optimized Grid Environment for Experiments –Targeted at DoE HENP (and/or DoE SSI); –Focus on medium to long-term software needs for HENP distributed data management; –Initial focus on instrumentation, modeling and optimization.

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL PPDG, APOGEE and GriPhyN A coherent program of work; Substantial common management proposed; A focus for HENP collaboration with Computer Science and Industry; PPDG/Apogee will create “middleware” needed by data-intensive science including LHC. (Synergy but no overlap with CMS/Atlas planning.)

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Data Grid Projects in Context Construction and Operation of HENP Data Management an Analysis Systems Tiers 0/1 >> $20M/yr of existing funding at HENP labs. e.g. SLAC FY1999 ~$7M equipment for BaBar (of which < $2M physics CPU); ~$3M labor, M&S.

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Data Grid Projects in Context Construction and Operation of HENP Data Management and Data Analysis Systems at DoE Laboratories Tiers 0/1 GriPhyN HENP Data Manage- ment at Major University Centers Tier 2 Draft proposal for NSF funding: $5-$16M/year $16M = $8M hardware $5M labor/R&D $3M network

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Data Grid Projects in Context Construction and Operation of HENP Data Management and Data Analysis Systems at DoE Laboratories Tiers 0/1 GriPhyN HENP Data Manage- ment at Major University Centers Tier 2 OO Databases and Analysis Tools Resource Management Tools Metadata Catalogs WAN Data Movers Mass Storage Management Systems Matchmaking Widely Applicable Technolgy and Computer Science (not only from HENP; 100s of non-HEP FTEs)

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Data Grid Projects in Context Construction and Operation of HENP Data Management and Data Analysis Systems at DoE Laboratories Tiers 0/1 GriPhyN HENP Data Manage- ment at Major University Centers Tier 2 OO Databases and Analysis Tools Resource Management Tools Metadata Catalogs WAN Data Movers Mass Storage Management Systems Matchmaking PPDG Particle Physics Data Grid NGI Project Large-scale tests/service focused on use of existing components

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Data Grid Projects in Context Construction and Operation of HENP Data Management and Data Analysis Systems at DoE Laboratories Tiers 0/1 GriPhyN HENP Data Manage- ment at Major University Centers Tier 2 OO Databases and Analysis Tools Resource Management Tools Metadata Catalogs WAN Data Movers Mass Storage Management Systems Matchmaking PPDG Particle Physics Data Grid NGI Project Unified Project Management Optimization and Evaluation Instrumentation Modeling and Simulation A new level of rigor as the foundation for future progress APOGEE

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Data Grid Projects in Context Construction and Operation of HENP Data Management and Data Analysis Systems at DoE Laboratories Tiers 0/1 GriPhyN HENP Data Manage- ment at Major University Centers Tier 2 OO Databases and Analysis Tools Resource Management Tools Metadata Catalogs WAN Data Movers Mass Storage Management Systems Matchmaking PPDG Particle Physics Data Grid NGI Project Unified Project Management Optimization and Evaluation Instrumentation Modeling and Simulation APOGEE R&D + Contacts with CS/Industry  Long-term Goals Testbeds

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Overall Program Goal A Coordinated Approach to the Design and Optimization of a Data Analysis Grid for HENP Experiments

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Particle Physics Data Grid Universities, DoE Accelerator Labs, DoE Computer Science Funded by DoE-NGI at $1.2M for first year

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL PPDG Collaborators Particle Accelerator Computer Physics Laboratory Science ANLXX LBNL XX BNLXX x CaltechXX FermilabXX x Jefferson LabXX x SLACXX x SDSCX WisconsinX

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL First Year PPDG Deliverables Implement and Run two services in support of the major physics experiments at BNL, FNAL, JLAB, SLAC: –“High-Speed Site-to-Site File Replication Service”; Data replication up to 100 Mbytes/s –“Multi-Site Cached File Access Service”: Based on deployment of file-cataloging, and transparent cache-management and data movement middleware First Year: Optimized cached read access to file in the range of 1-10 Gbytes, from a total data set of order One Petabyte Using middleware components already developed by the Proponents

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL PPDG Site-to-Site Replication Service u Network Protocols Tuned for High Throughput u Use of DiffServ for (1) Predictable high priority delivery of high - bandwidth data streams (2) Reliable background transfers u Use of integrated instrumentation to detect/diagnose/correct problems in long-lived high speed transfers [NetLogger + DoE/NGI developments] u Coordinated reservation/allocation techniques for storage-to-storage performance SECONDARY SITE CPU, Disk, Tape Robot PRIMARY SITE Data Acquisition, CPU, Disk, Tape Robot

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL PPDG Multi-site Cached File Access System University CPU, Disk, Users PRIMARY SITE Data Acquisition, Tape, CPU, Disk, Robot Satellite Site Tape, CPU, Disk, Robot Satellite Site Tape, CPU, Disk, Robot University CPU, Disk, Users University Users Satellite Site Tape, CPU, Disk, Robot

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL PPDG Middleware Components

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL APOGEE Focus on Instrumentation and Modeling Planned proposal to DOE Originally targeted at SSI Roughly the same collaborators as PPDG Intended to be the next step after PPDG

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Understanding Complex Systems (Writing into the BaBar Object Database at SLAC) Aug. 1: ~4.7 Mbytes/s Oct. 1: ~28 Mbytes/s

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL APOGEE Manpower Requirements (FTE) FY00FY01FY02FY03FY04 Instrumentation Low-level data capture Filtering and collecting agents Data analysis and presentation HENP workload profiling Simulation Framework design and development User workload simulation Component simulations (network, mass-storage system, object DB etc.) Site simulation packages111 Instrumentation/Simulation Testbed Instrumentation of existing experiment(s) (e.g.PPDG) Acquire and simulate performance measurements Acquire user workload profile Test prediction and optimization Evaluation and Optimization Quantify evolving needs of physics (including site policies etc.) Develop metrics for usefulness of data management facilities Optimize model systems Long-Term Strategy (Towards "Virtual Data") Tracking and testing HENP/CS/Industry developments Development projects in collaboration with HENP/CS/Industry Project Management (APOGEE and PPDG) Project leader (physicist) Lead computer scientist TOTALS

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL APOGEE Funding Needs $k$k$k$k$k FY00FY01FY02FY03FY04 Manpower Instrumentation Simulation Instrumentation/Simulation Testbed Evaluation and Optimization Long-Term Strategy (Towards "Virtual Data") Project Management (APOGEE and PPDG) Commercial Software Testbed hardware (in addition to parasitic use of production systems) Workstations, M&S, Travel TOTALS

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL GriPhyN Proposal Addresses several massive dataset problems  ATLAS, CMS  LIGO  Sloan Digital Sky Survey (SDSS)  Tier 2 computing centers (university based)  Hardware  commodity CPU / disk / tape  System support Networking Transatlantic link to CERN  "high-speed" Tier 2 backbone  multi-gigabit/sec R&D  Leverage Tier 2 + existing resources into Grid  Computer Science partnership, software

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL GriPhyN Goals Build production grid Exploit all computing resources most effectively Enable US physicists to participate fully in LHC program (also LIGO, SDSS)  Eliminate disadvantage of not being at CERN  Early physics analysis at LHC startup  Maintain and extend US leadership Build collaborative infrastructure for students & faculty  Training ground for next generation leaders

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Tier 2 Regional Centers Total number »20  ATLAS:6  CMS:6  LIGO:5  SDSS2  3 Flexible architecture and mission  complements national labs  Intermediate-level data handling  Makes possible regional collaborations  Well-suited to universities (training, mentoring and education) Scale: Tier2 = (university * laboratory) 1/2  1 scenario:  Tier 2 = Tier 1  Tier 2  20% Tier 1

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL GriPhyN Funding (Very Rough)

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL R&D Proposal $15M (Jan. 1999) R&D goals (complementary to APOGEE / PPDG)  Data, resource management over wide area  Fault-tolerant distributed computing over LAN  High-speed networks, as they relate to data management  Grid testbeds (with end-users)  Simulations crucial to success  MONARC group  With APOGEE / PPDG  Leverage resources available to us  Strong connections with Computer Science people  Existing R&D projects  Commercial connections

HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Grid Computing: Conclusions HENP at the frontier of Information Technology –Collaboration with Computer Science; –Collaboration with industry; –Outreach to other sciences. –Visibility (and scrutiny) of HENP computing; Enabling revolutionary advances in data analysis in the LHC era –Increasing the value of the vital investment in experiment-specific data-analysis software