Download presentation
Presentation is loading. Please wait.
1
SweGrid in practice, Grid(s) in Sweden TREFpunkt Karlshamn, 20-21 April,2005 Balázs Kónya Lund University NorduGrid Collaboration The Grid, as seen by Ursula Wilby, Sydsvenskan 10.2.2002
2
2 outline Grid Computing as of Today The Grid Vision How Grids work Production Grids SweGrid Hardware Operation & Services Middleware Users & Applications (Grids in Scandinavia) Summary
3
3 Some history: TREFpunkt 2002 Expression of interest: Development of GRID testbed in Sweden (Swedish informal GRID consortium, T Ekelöf). SWEGRID has been put into production Design and implementation of the NorduGrid Middleware Architecture (started February 2002) Advanced Resource Connector (ARC) from the NorduGrid Collaboration has become one of the major grid middlewares
4
4 Grid Computing as of Today: An old vision... The “Grid book” 1998: Computational Power Grid A future infrastructure of computing and data management a new utility, next to the existing water, heating, electricity,..., Computing from the tap source: IBM
5
5 Grid computing as of Today: Turning into reality... Europe is building its E- Infrastructure FP5: ~50 million Euro FP6: ~140 million Euro Grids become part of the European computing infrastructure Production quality Grids are being used on daily basis to solve scientific problems Vision: the grid layer should be seamlessly “integrated” with the network 5 “The Grid, for Europe, is far more than resource sharing. It is a big step forward to build the Cyberinfrastructure for a united research community tackling the grand challenges of our universe. It is a coordinated, single economic engine preparing to compete with Asia and the United States.” Wolfgang Gentzsch
6
6 Grid Computing as of Today The Grid Phenomenon (or hype) continues infecting the academics, the IT industry and the governments (bureaucrats). Grid computing has become the Holy Grail of distributed computing and a major marketing tool for the IT sector. The next BIG thing promised after the internet: World Wide Web access to information World Wide Grid access to computing capacity and beyond... Meanwhile a lot of research/development has been carried out: Matured middlewares & production grids
7
7 The rational behind the Grids remains the same Network vs. computer performance Computer speed doubles every 18 months Network speed doubles every 9 months Even data storage outperforms CPU Science is pursued in collaborations Teamwork: large experiments, instruments, data sets Need for dynamic, flexible, secure, coordinated resource sharing Need for sharing of geographically distributed resources The world gets super-connected Moore’s Law vs. storage improvements vs. optical improvements. Graph from Scientific American (Jan-2001) by Cleo Vilett, source Vined Khoslan, Kleiner, Caufield and Perkins.
8
8 Standards & Interoperability ??? Grid Computing as of Today: middlewares The middleware is the Operating system of the Grid A collection of software which implements grid functionalities No complete solution Major Middlewares (in alphabetic order): 1)EDG-line: EDG/LCG/Glite 2)Globus Toolkit (incompatible versions) GT v2 (pre-WS Globus) GT v3 (deprecated) Globus Toolkit v4 WS-RF Framework alpha/beta quality, first release in May 3)Grid3/OSG (GT2 + Condor) 4)NorduGrid/ARC 5)Unicore
9
9 How Grid works: overview of a Grid session The GRID middleware: Finds convenient places for the scientists "job" (computing task) to be run Optimizes use of the widely dispersed resources Organizes efficient access to scientific data Deals with authentication to the different sites that the scientists will be using Interfaces to local site authorization and resource allocation policies Runs the jobs Monitors progress Recovers from problems... and... Tells you when the work is complete and transfers the result to the requested location
10
10 Grid session (animated) RSL input output Cluster infosys program input ? ? ? gridjob ?
11
11 Production Grids: EGEE The largest Grid project, funded by the EU FP6 http://eu-egee.org Time span: April 2004 to April 2006, with planned 2-years extension Partners: 70+ institutes worldwide Major activities: Raise Grid awareness Provide resources and operational support via Grid technologies for scientists Maintain and improve Grid software Is a follow-up to the European DataGrid project (EDG), inheriting large parts of its Grid solutions Middleware: gLite, based on Globus and Condor Is based on the resources contributed to the LHC Computing Grid (LCG) The first release of the gLite middleware is to come out this month Is widely expected to become the largest production Grid Sweden is participating with SweGrid
12
12 Production Grids: Grid3 Grid3: originally, provided infrastructure and simple Grid-like solution for High Energy Physics computing in USA http://www.ivdgl.org/grid2003/ Collaboration of several research centers, active: 2003-2004 Uses Globus and Condor, plus few own developments Was proven to be able to provide reliable services to other applications cms dc04 atlas dc2
13
13 Production Grids: Open Science Grid Continuation and extension of Grid3 achievements http://www.opensciencegrid.org/ Consortium, aims at creating a national US Grid infrastructure Focus on general services, operations, end-to-end performance Takes over Grid3 in Spring 2005 (NOW)
14
14 Production Grids: NorduGrid Collaboration of Nordic researchers, developing an own Grid middleware solution (ARC) since 2001 http://www.nordugrid.org A Grid based on ARC- enabled sites Driven (so far) mostly by the needs and resources of the LHC experiments Dozens of other applications Assistance in Grid deployment outside the Nordic area SweGrid is part of this Grid
15
15 Production Grids: GRID.IT National Grid infrastructure in Italy http://www.grid.it/ Funded for 2003-2005 Also triggered by HEP community needs, but expands to many other applications Like EGEE, heavily based on the EDG Grid middleware Does specific developments, most notably, portals and monitoring tools Contributes to EGEE
16
16 Production Grids: SweGrid basic facts SweGrid is a computational Grid consisting of six dedicated clusters A Swedish National Computational resource The hardware is funded by a grant from the Wallenberg foundation Operational costs and personnel for support and maintenance are funded by the Swedish Research Council Official inauguration: March 2004 It was the first dedicated Grid resource in Scandinavia SweGrid operates in a production mode since then Runs on the reliable ARC middleware Offers support and maintenance services www.swegrid.se
17
17 SweGrid:Architecture Six dedicated clusters located at the Swedish academic computing centres. Each of the 6 clusters consists of 100 computing nodes and 2 TB disk storage Homogeneous hardware to simplify initial development and deployment The sites are connected through the 10 Gb/s GigaSunet network The OS installed differs between the clusters: RedHat Linux 7.3, Fedora Core 1, Debian 3.0 The primary Grid middleware is ARC, LCG/Glite is also being deployed
18
18 SweGrid: Grid node 100 compute nodes IA-32, 1 processor/node 2.8 GHz Intel P4 2 Gbyte memory 875P chipset 800 MHz FSB dual memory channel 2 TByte temporary storage FibreChannel for bandwidth 14 x 146 GByte 10000 rpm 1 Gigabit internal interconnect Not full bisectional bandwidth Access server Limited login 1 Gigabit to SUNET, directly attached
19
19 SweGrid middleware: ARC One of the few matured Grid solutions Continous development Good support Attractive for resource owners Non-intrusive Portable (variety of OS, LRMS) Simple installation procedure Scalable, reliable Performs well in ATLAS Data Challenges Scalabe: serves a grid of ~50 sites and 5000 CPUs, 50000 jobs/month Attractive for users Robust, portable Relatively feature rich Client can be installed everywhere by anyone Plenty of documentation IANA registered Grid ports: 2135, 2811
20
20 Swegrid Users: Resource allocation SweGrid is part of the Swedish National Infrastructure for Computing (SNIC) Potential users must apply for resource allocations (CPU quotas) Resource allocations for users are done on a peer-review basis via the Swedish National Allocation Committee (SNAC), upon requests 1/3 is (pre-)allocated for LHC computing The rest is distributed between chemistry, genomics, meteorology etc – whoever applies Allocation is done twice a year 20 applications in the last round requesting 153740 "CPU-hours". 18 was granted to the sum of 82300 "CPU-hours"
21
21 SweGrid services Operation, Support, Maintenance Six assigned Swegrid system administrators, one per site Detect and resolve hardware failures Installation and upgrades of software Monitoring and performance enhancements Tutorials Helpdesk Application support SweGrid Accounting System (SGAS) Developed within the SweGrid project Resource allocation and enforcement system Grid bank Usage tracking Test deployment on SweGrid
22
22 SweGrid Users: Application areas & Research SweGrid is designed mainly for through-put computation, to quickly process large numbers of loosely coupled non-parallel computations Users come from various fields of science: climate research, material science, physics, chemistry and biology www.pdc.kth.se/grid/swegrid-vo/volist.txt The SweGrid also hosts research activities in various fields in IT such as Distributed data bases, Scheduling, brokering Data management, replication SweGrid has also been developing the SweGrid Accounting System (SGAS)
23
23 Swegrid Users: Main User Group High Energy Physics (HEP) Community has been the driving force behind Grid Initially the main customers of SweGrid 1/3 of SweGrid is reserved for HEP Data Challenges ~110.000 jobs in the second half of 2004
24
24 SweGrid: a multidisciplinary grid Date: 04/15/2005 02:38 PM To: nordugrid-discuss@nordugrid.orgnordugrid-discuss@nordugrid.org Subject: [NG-disc] Multidisciplinary grid Right now bluesmoke has jobs running belonging to: - climate simulation - astrophysics - bioinformatics - materials science Actually no HEP at all, at the moment. I think we might be underestimating the importance and coolness of this. -- Leif Nixon - Systems expert ------------------------------------- ----------------------- National Supercomputer Centre - Linkoping University -------- ----------------------------------------------------
25
25 Grid of Grids Eesti Grid Finn activities SweGrid NorGrid DCGC NDGF Germany Switzerland Slovenia Slovakia
26
26 Estonian Grid Technical details: - 122 CPUs in 9 clusters - 3.8 TB storage - NorduGrid ARC middleware - GÉANT connection 622Mbit/s, between the clusters 1Gbit/s Operations: EG CA EG CA is member of EUGridPMA Technical support and coordination group: Steering committee established at the Ministry of Education and Research of Estonia Challenges: Estonian electronic ID-card infrastructure on the Grid (700 000 valid electronic ID-cards issued in Estonia!) Local experiences with E-money and rental software Interoperability of ARC and UNICORE
27
27 DCGC Danish Center for Grid Computing Danish Center for Grid Computing, established in August 2003 http://www.dcgc.dk 3 years project, aiming at Provide Grid-access to test facilities (including HPC) Host development activities Provide user support Aims at getting industrial partners and users involved Uses ARC for middleware
28
28 Finland: Material Sciences Grid (M- grid) First large initiative to put Grid middleware into production use in Finland: http://www.csc.fi/proj/mgrid/ Based on ARC and Linux clusters, currently 443 CPUs, targeted for serial and ”pleasantly parallel” applications, clusters can be accessed both locally and via ARC Joint project between seven Finnish universities, Helsinki Institute of Physics and CSC founded by the universities and the Academy of Sciences Users mainly from the physics and chemistry departments in the partner universities Material physicists, particle physicists, chemists, some bioscientists Typical applications: Gromacs, Gaussian, Dalton
29
29 NORGRID Norwegian project for Grid competence building, January – December 2004 http://norgrid.uio.no ca. 3 FTE Partners : NTNU, UiB, UiO, UiT, UNINETT Funding 50% NFR and 50% partners Objectives: competence building on Grid middleware and related technologies in Norway Prepare a middleware infrastructure for the next HPC project (starting 2005) Emphasis on distributed data management, metascheduling, portals Presently uses ARC Aims to evaluate it against UNICORE, GT4.x and LCG2/gLite
30
30 NDGF Nordic Data Grid Facility pilot project was launched in spring 2003 Initial success of NorduGrid provided grounds for a Nordic Grid facility http://www.ndgf.org Funds for 1 director + 4 postdocs in each country Strong emphasis towards portal development and storage facilities Aimed to evaluate various Grid solutions, uses ARC Will produce recommendations for the Nordic Grid facility Aims to harness all the resources in the Nordic countries Grid of Grids with a large centralized storage facility This facility is expected to become the Nordic Tier1 candidate
31
31 NGN Nordic Grid Neighborhood is a networking project funded by the Nordplus program, started in September 2004 http://www.nicpb.ee/NordicGrid/ Expands to the Baltic states and North-West Russia (St.Petersburg) 20 partners supports and strengthens contacts in the field of Grid technologies activities cover education, reciprocal knowledge transfer and Grid research and development Plans to set up a testbed to deploy and demonstrate ARC and AliEn The scope is to attract and educate users that can benefit from Grid, e.g., medical applications
32
32 Summary With the creation of SweGrid Sweden made the first dedicated Scandinavian investment in a Grid computing Infrastructure. The SweGrid model has been followed by other Scandinavian countries SweGrid offers a production quality grid infrastructure since the beginning of 2004 SweGrid is part of the Swedish National Infrastructure for Computing and being heavily used by a multidisciplinary user base Through the SweGrid resources Sweden joins to major world-class Grids: NorduGrid, EGEE STAC (Swedish Technical Advisory Committee) is determined that the SweGrid project will have a continuation
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.