Brussels Grid Meeting (Mar. 23, 2001)Paul Avery1 University of Florida Extending the Grid Reach in Europe.

Slides:



Advertisements
Similar presentations
CERN STAR TAP June 2001 Status of the EU DataGrid Project Fabrizio Gagliardi CERN EU-DataGrid Project Leader June 2001
Advertisements

International Grid Communities Dr. Carl Kesselman Information Sciences Institute University of Southern California.
Particle physics – the computing challenge CERN Large Hadron Collider –2007 –the worlds most powerful particle accelerator –10 petabytes (10 million billion.
Fighting Malaria With The Grid. Computing on The Grid The Internet allows users to share information across vast geographical distances. Using similar.
INFSO-RI Enabling Grids for E-sciencE The EGEE project Fabrizio Gagliardi Project Director EGEE CERN, Switzerland Research Infrastructures.
High Performance Computing Course Notes Grid Computing.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
EGEE is proposed as a project funded by the European Union under contract IST The EGEE International Grid Infrastructure and the Digital Divide.
US-CMS Meeting (May 19, 2001)Paul Avery1 US-CMS Meeting (UC Riverside) May 19, 2001 Grids for US-CMS and CMS Paul Avery University of Florida
CERN Krakow 2001 F. Gagliardi - CERN/IT 1 RTD efforts in Europe by Kyriakos Baxevanidis Foster cohesion, interoperability, cross- fertilization of knowledge,
1 Distributed, Internet and Grid Computing. 2 Distributed Computing Current supercomputers are too expensive ASCI White (#1 in TOP500) costs more than.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Knowledge Environments for Science: Representative Projects Ian Foster Argonne National Laboratory University of Chicago
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
HEP Prospects, J. Yu LEARN Strategy Meeting Prospects on Texas High Energy Physics Network Needs LEARN Strategy Meeting University of Texas at El Paso.
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
International collaboration in high energy physics experiments  All large high energy physics experiments today are strongly international.  A necessary.
CANS Meeting (December 1, 2004)Paul Avery1 University of Florida UltraLight U.S. Grid Projects and Open Science Grid Chinese American.
Peer to Peer & Grid Computing Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science The University.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
CERN TERENA Lisbon The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
WG Goals and Workplan We have a charter, we have a group of interested people…what are our plans? goalsOur goals should reflect what we have listed in.
From GEANT to Grid empowered Research Infrastructures ANTONELLA KARLSON DG INFSO Research Infrastructures Grids Information Day 25 March 2003 From GEANT.
A short introduction to the Worldwide LHC Computing Grid Maarten Litmaath (CERN)
CHEP 2000 (Feb. 7-11)Paul Avery (Data Grids in the LHC Era)1 The Promise of Computational Grids in the LHC Era Paul Avery University of Florida Gainesville,
Jarek Nabrzyski, Ariel Oleksiak Comparison of Grid Middleware in European Grid Projects Jarek Nabrzyski, Ariel Oleksiak Poznań Supercomputing and Networking.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
GriPhyN EAC Meeting (Jan. 7, 2002)Carl Kesselman1 University of Southern California GriPhyN External Advisory Committee Meeting Gainesville,
GriPhyN EAC Meeting (Apr. 12, 2001)Paul Avery1 University of Florida Opening and Overview GriPhyN External.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Perspectives on Grid Technology Ian Foster Argonne National Laboratory The University of Chicago.
Bob Jones Technical Director CERN - August 2003 EGEE is proposed as a project to be funded by the European Union under contract IST
HEP-CCC Meeting, November 1999Grid Computing for HEP L. E. Price, ANL Grid Computing for HEP L. E. Price Argonne National Laboratory HEP-CCC Meeting CERN,
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
…building the next IT revolution From Web to Grid…
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
E-Science Research Councils awarded e-Science funds ” science increasingly done through distributed global collaborations enabled by the Internet, using.
Middleware Camp NMI (NSF Middleware Initiative) Program Director Alan Blatecky Advanced Networking Infrastructure and Research.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
3 December 2015 Examples of partnerships and collaborations from the Internet2 experience Interworking2004 Ottawa, Canada Heather Boyles, Internet2
GriPhyN EAC Meeting (Jan. 7, 2002)Paul Avery1 Integration with iVDGL è International Virtual-Data Grid Laboratory  A global Grid laboratory (US, EU, Asia,
29/1/2002A.Ghiselli, INFN-CNAF1 DataTAG / WP4 meeting Cern, 29 January 2002 Agenda  start at  Project introduction, Olivier Martin  WP4 introduction,
1CHEP2000 February 2000F. Gagliardi EU HEP GRID Project Fabrizio Gagliardi
Fire Emissions Network Sept. 4, 2002 A white paper for the development of a NSF Digital Government Program proposal Stefan Falke Washington University.
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
LHC Computing, CERN, & Federated Identities
CERN The GridSTART EU accompany measure Fabrizio Gagliardi CERN
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
Open Science Grid in the U.S. Vicky White, Fermilab U.S. GDB Representative.
LHC Computing, SPC-FC-CC-C; H F Hoffmann1 CERN/2379/Rev: Proposal for building the LHC computing environment at CERN (Phase 1) Goals of Phase.
GRIDSTART Brussels 20/9/02 1www.gridstart.org GRIDSTART and European activities Dr Francis Wray EPCC The University of Edinburgh.
LHC Computing at RAL PPD Dave Newbold RAL PPD / University of Bristol The LHC computing challenge PPD and the Grid Computing for physics PPD added value.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
Hall D Computing Facilities Ian Bird 16 March 2001.
] Open Science Grid Ben Clifford University of Chicago
Clouds , Grids and Clusters
Long-term Grid Sustainability
LHC Computing, RRB; H F Hoffmann
Presentation transcript:

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery1 University of Florida Extending the Grid Reach in Europe Brussels, Mar. 23, Global Data Grids The Need for Infrastructure

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery2 Global Data Grid Challenge “Global scientific communities, served by networks with bandwidths varying by orders of magnitude, need to perform computationally demanding analyses of geographically distributed datasets that will grow by at least 3 orders of magnitude over the next decade, from the 100 Terabyte to the 100 Petabyte scale.”

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery3 Data Intensive Science: è Scientific discovery increasingly driven by IT  Computationally intensive analyses  Massive data collections  Rapid access to large subsets  Data distributed across networks of varying capability è Dominant factor: data growth (1 Petabyte = 1000 TB)  2000~0.5 Petabyte  2005~10 Petabytes  2010~100 Petabytes  2015~1000 Petabytes? How to collect, manage, access and interpret this quantity of data?

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery4 Data Intensive Disciplines è High energy & nuclear physics è Gravity wave searches (e.g., LIGO, GEO, VIRGO) è Astronomical sky surveys (e.g., Sloan Sky Survey) è Global “Virtual” Observatory è Earth Observing System è Climate modeling è Geophysics

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery5 Data Intensive Biology and Medicine è Radiology data è X-ray sources (APS crystallography data) è Molecular genomics (e.g., Human Genome) è Proteomics (protein structure, activities, …) è Simulations of biological molecules in situ è Human Brain Project è Global Virtual Population Laboratory (disease outbreaks) è Telemedicine è Etc. Commercial applications not far behind

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery6 The Large Hadron Collider at CERN “Compact” Muon Solenoid at the LHC Standard man

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery Physicists 150 Institutes 32 Countries LHC Computing Challenges è Complexity of LHC environment and resulting data è Scale: Petabytes of data per year (100 PB by 2010) è Global distribution of people and resources CMS Experiment

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery8 Global LHC Data Grid Hierarchy Tier 1 T Tier 0 (CERN) Tier0 CERN Tier1 National Lab Tier2 Regional Center at University Tier3 University workgroup Tier4 Workstation GriPhyN: è R&D è Tier2 centers è Unify all IT resources

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery9 Global LHC Data Grid Hierarchy Tier2 Center Online System CERN Computer Center > 20 TIPS France Center USA Center Italy Center UK Center Institute Institute ~0.25TIPS Workstations, other portals ~100 MBytes/sec Gb/sec Mbits/sec Bunch crossing per 25 nsecs. 100 triggers per second Event is ~1 MByte in size Physicists work on analysis “channels”. Each institute has ~10 physicists working on one or more channels Physics data cache ~PBytes/sec Gb/sec Tier2 Center ~622 Mbits/sec Tier 0 +1 Tier 1 Tier 3 Tier 4 Tier2 Center Experiment Tier 2

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery10 Global Virtual Observatory Source Catalogs Image Data Specialized Data: Spectroscopy, Time Series, Polarization Information Archives: Derived & legacy data: NED,Simbad,ADS, etc Discovery Tools: Visualization, Statistics Standards Multi-wavelength astronomy, Multiple surveys

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery11 GVO: The New Astronomy è Large, globally distributed database engines  Integrated catalog and image databases  Multi-Petabyte data size  Gbyte/s aggregate I/O speed per site è High speed (>10 Gbits/s) backbones  Cross-connecting, correlating the major archives è Scalable computing environment  100s–1000s of CPUs for statistical analysis and discovery

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery12 Infrastructure for Global Grids

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery13 Grid Infrastructure è Grid computing sometimes compared to electric grid  You plug in to get resource (CPU, storage, …)  You don’t care where resource is located è This analogy might have an unfortunate downside You might need different sockets!

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery14 Role of Grid Infrastructure è Provide essential common Grid infrastructure  Cannot afford to develop separate infrastructures è Meet needs of high-end scientific collaborations  Already international and even global in scope  Need to share heterogeneous resources among members  Experiments drive future requirements è Be broadly applicable outside science  Government agencies: National, regional (EU), UN  Non-governmental organizations (NGOs)  Corporations, business networks (e.g., supplier networks)  Other “virtual organizations” è Be scalable to the Global level  But EU + US is a good starting point

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery15 A Path to Common Grid Infrastructure è Make a concrete plan è Have clear focus on infrastructure and standards è Be driven by high-performance applications è Leverage resources & act coherently è Build large-scale Grid testbeds è Collaborate with industry

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery16 Building Infrastructure from Data Grids è 3 Data Grid projects recently funded è Particle Physics Data Grid (US, DOE)  Data Grid applications for HENP  Funded 2000, 2001  è GriPhyN (US, NSF)  Petascale Virtual-Data Grids  Funded 9/2000 – 9/2005  è European Data Grid (EU)  Data Grid technologies, EU deployment  Funded 1/2001 – 1/2004   HEP in common  Focus: infrastructure development & deployment  International scope

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery17 Background on Data Grid Projects è They support several disciplines  GriPhyN:CS, HEP (LHC), gravity waves, digital astronomy  PPDG:CS, HEP (LHC + current expts), Nuc. Phys., networking  DataGrid:CS, HEP, earth sensing, biology, networking è They are already joint projects  Each serving needs of multiple constituencies  Each driven by high-performance scientific applications  Each has international components  Their management structures are interconnected è Each project developing and deploying infrastructure  US$23M (additional proposals for US$35M) What if they join forces?

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery18 A Common Infrastructure Opportunity è GriPhyN + PPDG + EU-DataGrid + national efforts  France, Italy, UK, Japan è Have agreed to collaborate, develop joint infrastructure  Initial meeting March 4 in Amsterdam to discuss issues  Future meetings in June, July è Preparing management document  Joint management, technical boards + steering committee  Coordination of people, resources  An expectation that this will lead to real work è Collaborative projects  Grid middleware  Integration into applications  Grid testbed: iVDGL  Network testbed (Foster): T 3 = Transatlantic Terabit Testbed

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery19 iVDGL è International Virtual-Data Grid Laboratory  A place to conduct Data Grid tests at scale  A concrete manifestation of world-wide grid activity  A continuing activity that will drive Grid awareness  A basis for further funding è Scale of effort  For national, international scale Data Grid tests, operations  Computationally and data intensive computing  Fast networks è Who  Initially US-UK-EU  Other world regions later  Discussions w/ Russia, Japan, China, Pakistan, India, South America

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery20 iVDGL Parameters è Local control of resources vitally important  Experiments, politics demand it  US, UK, France, Italy, Japan,... è Grid Exercises  Must serve clear purposes  Will require configuration changes  not trivial  “Easy”, intra-experiment tests first (10-20%, national, transatlantic)  “Harder” wide-scale tests later (50-100% of all resources) è Strong interest from other disciplines  Our CS colleagues (wide scale tests)  Other HEP + NP experiments  Virtual Observatory (VO) community in Europe/US  Gravity wave community in Europe/US/(Japan?)  Bioinformatics

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery21 Revisiting the Infrastructure Path è Make a concrete plan  GriPhyN + PPDG + EU DataGrid + national projects è Have clear focus on infrastructure and standards  Already agreed  COGS (Consortium for Open Grid Software) to drive standards? è Be driven by high-performance applications  Applications are manifestly high-perf: LHC, GVO, LIGO/GEO/Virgo, …  Identify challenges today to create tomorrow’s Grids

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery22 Revisiting the Infrastructure Path (cont) è Leverage resources & act coherently  Well-funded experiments depend on Data Grid infrastructure  Collab. with national laboratories: FNAL, BNL, RAL, Lyon, KEK, …  Collab. with other Data Grid projects: US, UK, France, Italy, Japan  Leverage new resources: DTF, CAL-IT 2, …  Work through Global Grid Forum è Build and maintain large-scale Grid testbeds  iVDGL  T 3 è Collaboration with industry  next slide è EC investment in this opportunity  Leverage and extend existing projects, worldwide expertise  Invest in testbeds  Work with national projects (US/NSF, UK/PPARC, …) Part of same infrastructure

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery23 Collaboration with Industry è Industry efforts are similar, but only in spirit  ASP, P2P, home PCs, …  IT industry mostly has not invested in Grid R&D  We have different motives, objectives, timescales è Still many areas of common interest  Clusters, storage, I/O  Low cost cluster management  High-speed, distributed databases  Local and wide-area networks, end-to-end performance  Resource sharing, fault-tolerance, … è Fruitful collaboration requires clear objectives è EC could play important role in enabling collaborations

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery24 Status of Data Grid Projects è GriPhyN  US$12M funded by NSF/ITR 2000 program (5 year R&D)  2001 supplemental funds requested for initial deployments  Submitting 5-year proposal ($15M) to NSF  Intend to fully develop production Data Grids è Particle Physics Data Grid  Funded in 1999, 2000 by DOE ($1.2 M per year)  Submitting 3-year Proposal ($12M) to DOE Office of Science è EU DataGrid  10M Euros funded by EU (3 years, 2001 – 2004)  Submitting proposal in April for additional funds è Other projects?

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery25 Grid References è Grid Book  è Globus  è Global Grid Forum  è PPDG  è EU DataGrid  è GriPhyN 

Brussels Grid Meeting (Mar. 23, 2001)Paul Avery26 Summary è Grids will qualitatively and quantitatively change the nature of collaborations and approaches to computing è Global Data Grids provide challenges needed to build tomorrows Grids è We have a major opportunity to create common infrastructure è Many challenges during the coming transition  New grid projects will provide rich experience and lessons  Difficult to predict situation even 3-5 years ahead