Download presentation
Presentation is loading. Please wait.
Published byTyler White Modified over 9 years ago
1
Experiment Requirements for Global Infostructure Irwin Gaines FNAL/DOE
2
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Outline Recall partnership principles LHC computing model CMS and ATLAS grid prototyping Categories of work packages Contributors
3
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Agreement on 5 principles: The cost and complexity of 21 st Century Science requires the creation of advanced and coherent global Infostructure (information infrastructure). The construction of a coherent Global Infostructure for Science requires definition and drivers from Global Applications (that will also communicate with each other) Further, forefront Information Technology must be incorporated into this Global Infostructure for the Applications to reach their full potential for changing the way science is done. LHC is a near term Global Application requiring advanced and un-invented Infostructure and is ahead in planning compared to many others. U.S. agencies must work together for effective U.S. participation on Global scale infostructure, and the successful execution of the LHC program in a 4 way agency partnership, with international cooperation in view.
4
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure LHC as exemplar of global science Project already involves scientists (and funding agencies) from all over the world High visibility science Experiments already making good use of prototype grids Sociological (as well as technical) reasons for decentralized computing systems Recognized challenge of accumulating sufficient resources
5
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure LHC Global Science LHC is most exciting, challenging, and relevant science Challenges scientifically, technically, culturally, managerially Collaboration open and fair access and sharing of data, tools, ideas unique opportunities for discovery to small and remote groups Data and Information vast data beyond technical capabilities of any single organization revolutionary new applications of new tools of information technology Globalization building truly global (science) communities acquiring data centrally, analyzing data globally, like a large corporation Opportunities to advance Information Technology Relevant to Science at Large
6
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Centres taking part in the LCG-1 around the world around the clock
7
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure LHC Computing Model Distributed model from the start (distributed resources + coherent global access to data) Must support Production (reconstruction, simulation) Scheduled, predictable, batch Run by experiment or physics group Highly compute intensive, accesses predictable data sets Data Analysis (including calibration and monitoring) Random, chaotic, often interactive Run by individuals and small groups Mostly data intensive, accesses random data Highly collaborative Code development and testing Highly interactive Highly collaborative
8
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure LHC Computing Facilities Model
9
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Adoption of Grids by LHC Experiments Already some major successes: CMS and ATLAS production running Good collaborations with computer scientists: iVDGL, GriPhyN, PPDG, EDG… LHC Computing Grid Project (LCG) We now have a scientific community that understands the components and value of grid computing.
10
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Lawrence Berkeley National Laboratory Brookhaven National Laboratory Indiana University Boston University Argonne National Laboratory U Michigan University of Texas at Arlington Oklahoma University US -ATLAS testbed launched February 2001 ATLAS Grid Testbed Sites
11
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure l Fermilab –1+5 PIII dual 0.700 GHz processor machines l Caltech –1+3 AMD dual 1.6 GHz processor machines l San Diego –1+3 PIV single 1.7 GHz processor machines l Florida –1+5 PIII dual 1 GHz processor machines l Rice –1+? machines l Wisconsin –5 PIII single 1 GHz processor machines l Total: l ~40 1 GHz dedicated processors UCSD Florida Wisconsin Caltech Fermilab US-CMS Development Grid Testbed Rice
12
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure US-CMS Integration Grid Testbed l Fermilab (Tier1) –40 dual 0.750 GHz processor machines l Caltech (Tier2) –20 dual 0.800 GHz processor machines –20 dual 2.4 GHz processor machines l San Diego (Tier2) –20 dual 0.800 GHz processor machines –20 dual 2.4 GHz processor machines l Florida (Tier2) –40 dual 1 GHz processor machines l CERN (LCG Tier0 site) –36 dual 2.4 GHz processor machines l Total: l 240 0.85 GHz processors: Red Hat 6 l 152 2.4 GHz processors: Red Hat 7 UCSD Florida Caltech Fermilab CERN
13
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure “work packages” for LHC computing HW infrastructure Distributed computing infrastructure Grid services Experiment software Collaboration tools Support services
14
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Hardware Infrastructure Tier 0 at CERN Compute elements Storage elements Mass storage Tier 1 national regional centers Tier 2 regional centers Local computing resources
15
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Distributed computing infrastructure Networking Intercontinental Regional wide area Local “end user” connections Servers for distributed computing Metadata servers Resources brokers Monitoring centers
16
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Grid Services Low level middleware (casual user doesn’t see this layer) Application specific middleware (service built on top of low level MW with flexible user interfaces and higher level functionality) Modeling and monitoring Trouble shooting and fault tolerance Distributed Data Analysis Environment Grid hardware for Research and development of tools Deployment and integration Production
17
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure CMS Approach to R&D, Integration, Deployment prototyping, early roll out, strong QC/QA & documentation, tracking of external “practices”
18
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Experiment Software core SW detector specific applications physics analysis support analysis group support Some of this software suitable for common development
19
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Collaboration tools phone conferencing video conferencing Remote informal interaction (virtual coffee break) Document sharing Collaborative software development Collaborative data analysis Telepresence Remote control of experiment
20
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Support Services Training and documentation Information Services User support (help desk): 24x7
21
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Grid Middleware I 1. User Management 1.1 Registration of users as members of a virtual organization (VO) (including subgroup credentials within the VO) 1.2 Authentication of users 1.3 Authorization of users for particular tasks
22
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Grid Middleware II 2. Resource Management 2.1 Resource declaration (making resources available to the Grid) 2.2 Resource discovery 2.3 Resource assignment tools (eg, these CPUs are only available for experiment A, only for physicists from country B, only for physics group C, etc) 2.4 Prioritization tools
23
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Grid Middleware III 3. Job Management 3.1 Job Submission 3.2 job monitoring 4. Data management 4.1 Data replication 4.2 Data access 4.3 data set management 4.4 Data movement/job movement/data re- creation decisions
24
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Production Grids middleware support error recovery robustness 24x7 operation monitoring and system usage optimization strategy and policy for resource allocation authentication and authorization simulation of grid operations tools for optimizing distributed systems
25
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Blue Ribbon Panel on Cyberinfrastructure
26
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Contributors Funding Agencies: Base Program Funding Agencies: LHC Research Program (LHC Software & Computing Projects) US Funding Agencies: networks and infrastructure CERN Tier 0/1 facilities at CERN Networking and infrastructure LCG Project Other collaborating countries funding agencies DOE/NSF Computatational Science Research Program
27
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Who contributes where? Hardware Infrastructure Distributed Computing Infrastructure Grid Services Experiment Software Collaboration Tools Support Services US Base Program US Research Program DOE/NSF Networking CERN Other countries DOE/NSF CS * ********** * ********** ************* ******
28
7-Feb-0302 3 rd NSF/DOE meeting on partnerships for global inforstructure Proposal for further action Form small working group (representatives from experiments, both agencies, physics and CS side) to flesh out workplans and “sign up” for tasks (Road Map to Global Infostructure): report back in < 1 month Meeting soon in Europe
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.