ANL Royal Society - June 2004 The TeraGyroid Project - Aims and Achievements Richard Blake Computational Science and Engineering Department CCLRC Daresbury.

Slides:



Advertisements
Similar presentations
HPCx Power for the Grid Dr Alan D Simpson HPCx Project Director EPCC Technical Director.
Advertisements

Peter Berrisford RAL – Data Management Group SRB Services.
SC2004 GLIF Panel Peter Clarke GLIF Infrastructure SC2004 Panel Discussion Peter Clarke UK National e-Science Centre.
Rob Allan Daresbury Laboratory NW-GRID Training Event 26 th January 2007 NW-GRID Future Developments R.J. Allan CCLRC Daresbury Laboratory.
Joint CASC/CCI Workshop Report Strategic and Tactical Recommendations EDUCAUSE Campus Cyberinfrastructure Working Group Coalition for Academic Scientific.
Towards a Virtual European Supercomputing Infrastructure Vision & issues Sanzio Bassini
Background Chronopolis Goals Data Grid supporting a Long-term Preservation Service Data Migration Data Migration to next generation technologies Trust.
Stephen Pickles UKLight Town Meeting, NeSC, Edinburgh, 9/9/2004 TeraGyroid HPC Applications.
John Kewley e-Science Centre GIS and Grid Computing Workshop 13 th September 2005, Leeds Grid Middleware and GROWL John Kewley
Computational Steering on the GRID Using a 3D model to Interact with a Large Scale Distributed Simulation in Real-Time Michael.
Sergey Belov, LIT JINR 15 September, NEC’2011, Varna, Bulgaria.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
NGNS Program Managers Richard Carlson Thomas Ndousse ASCAC meeting 11/21/2014 Next Generation Networking for Science Program Update.
Simo Niskala Teemu Pasanen
Results Matter. Trust NAG. Numerical Algorithms Group Mathematics and technology for optimized performance Andrew Jones IDC HPC User Forum, Imperial College.
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
UK HPC Review 05 Sept 2005 Paris, 31 March 2003 Give me a suitable title.... Peter Coveney Centre for Computational Science University College London.
HPCx: Multi-Teraflops in the UK A World-Class Service for World-Class Research Dr Arthur Trew Director.
The MicroGrid: A Scientific Tool for Modeling Grids Andrew A. Chien SAIC Chair Professor Department of Computer Science and Engineering University of California,
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Cyberinfrastructure: Enabling New Research Frontiers Sangtae “Sang” Kim Division Director – Division of Shared Cyberinfrastructure Directorate for Computer.
Patterns for E-Research Dave Berry, Research Manager E-Research within the University of Edinburgh, 2 nd March 2005.
Computational Steering on Grids A survey of RealityGrid Peter V. Coveney Centre for Computational Science, University College London.
Manchester Computing Supercomputing, Visualization & e-Science Realistic modelling of complex problems on Grids John Brooke (University of Manchester)
The Globus Project: A Status Report Ian Foster Carl Kesselman
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
GGF-16 Athens Production Grid Computing in the UK Neil Geddes CCLRC Director, e-Science.
© 2006 Open Grid Forum Enabling Pervasive Grids The OGF GIN Effort Erwin Laure GIN-CG co-chair, EGEE Technical Director
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
Rob Allan Daresbury Laboratory NW-GRID Training Event 25 th January 2007 North West Grid Overview R.J. Allan CCLRC Daresbury Laboratory A world-class Grid.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Introduction to Grid Computing Ed Seidel Max Planck Institute for Gravitational Physics
Robust Grid-based environment for large scale lattice-Boltzmann simulations Maddalena Venturoli Matt Harvey Peter Coveney Giovanni Giupponi Jonathan Chin.
“Grids and eScience” Mark Hayes Technical Director - Cambridge eScience Centre GEFD Summer School 2003.
The UK eScience Grid (and other real Grids) Mark Hayes NIEeS Summer School 2003.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Building the e-Minerals Minigrid Rik Tyer, Lisa Blanshard, Kerstin Kleese (Data Management Group) Rob Allan, Andrew Richards (Grid Technology Group)
P ITTSBURGH S UPERCOMPUTING C ENTER R ESOURCES & S ERVICES Marvel 0.3 TF HP GS 1280 SMP OS: Tru64 Unix 2 nodes (128 processors) Nodes: 64 x 1.15 GHz EV67.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Cactus/TIKSL/KDI/Portal Synch Day. Agenda n Main Goals:  Overview of Cactus, TIKSL, KDI, and Portal efforts  present plans for each project  make sure.
Terena conference, June 2004, Rhodes, Greece Norbert Meyer The effective integration of scientific instruments in the Grid.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Combining the strengths of UMIST and The Victoria University of Manchester “Use cases” Stephen Pickles e-Frameworks meets e-Science workshop Edinburgh,
GridPP 11 th Collaboration Meeting Networking: Current Status Robin Tasker 14 September 2004.
Computational Steering on Grids A survey of RealityGrid Peter V. Coveney Centre for Computational Science, University College London.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
Manchester Computing S. M. Pickles, R. Haines, R. L. Pinning and A. R. Porter UK e-Science All Hands Meeting, nd September 2004 Practical Tools.
Utility Computing: Security & Trust Issues Dr Steven Newhouse Technical Director London e-Science Centre Department of Computing, Imperial College London.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
National Center for Supercomputing Applications University of Illinois at Urbana–Champaign Visualization Support for XSEDE and Blue Waters DOE Graphics.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
John Kewley e-Science Centre All Hands Meeting st September, Nottingham GROWL: A Lightweight Grid Services Toolkit and Applications John Kewley.
Albert-Einstein-Institut Exploring Distributed Computing Techniques with Ccactus and Globus Solving Einstein’s Equations, Black.
The National Grid Service Mike Mineter.
1 Porting applications to the NGS, using the P-GRADE portal and GEMLCA Peter Kacsuk MTA SZTAKI Hungarian Academy of Sciences Centre for.
RobuSTore: Performance Isolation for Distributed Storage and Parallel Disk Arrays Justin Burke, Huaxia Xia, and Andrew A. Chien Department of Computer.
WHAT SURF DOES FOR RESEARCH SURF’s Science Engagement TNC15 June 18, 2015 Sylvia Kuijpers (SURFnet)
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
Building PetaScale Applications and Tools on the TeraGrid Workshop December 11-12, 2007 Scott Lathrop and Sergiu Sanielevici.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Page : 1 SC2004 Pittsburgh, November 12, 2004 DEISA : integrating HPC infrastructures in Europe DEISA : integrating HPC infrastructures in Europe Victor.
Clouds , Grids and Clusters
Electron Ion Collider New aspects of EIC experiment instrumentation and computing, as well as their possible impact on and context in society (B) COMPUTING.
UK Grid: Moving from Research to Production
DEISA : integrating HPC infrastructures in Europe Prof
Grid Portal Services IeSE (the Integrated e-Science Environment)
The National Grid Service
Presentation transcript:

ANL Royal Society - June 2004 The TeraGyroid Project - Aims and Achievements Richard Blake Computational Science and Engineering Department CCLRC Daresbury Laboratory This ambitious project was the result of an international collaboration linking the USA’s TeraGrid and the UK’s e- Science Grid, jointly funded by NSF and EPSRC. Trans- Atlantic optical bandwidth is supported by British Telecommunications.

ANL Royal Society - June 2004 Overview Project Objectives The TeraGyroid scientific experiment Testbed and Partners Applications Porting and RealityGrid Environment Grid Software Infrastructure Visualization Networking What was done Project Objectives - How well did we do? Lesson Learned

ANL Royal Society - June 2004 UK-Teragrid HPC Project Objectives Joint experiment combining high-end computational facilities in the UK e-Science Grid (HPCx and CSAR) and the Teragrid sites: –world class computational science experiment –enhanced expertise/ experience to benefit UK and USA –inform construction/operation of national/ international grids –stimulate long-term strategic technical collaboration –support long-term scientific collaborations –experiments with clear scientific deliverables –choice of applications to be based on community codes –inform future programme of complementary experiments

ANL Royal Society - June 2004 The TeraGyroid Scientific Experiment High-density isosurface of the late-time configuration in a ternary amphiphilic fluid as simulated on a 64 3 lattice by LB3D. Gyroid ordering coexists with defect-rich, sponge-like regions. The dynamical behaviour of such defect-rich systems can only be studied with very large scale simulations, in conjunction with high- performance visualisation and computational steering.

ANL Royal Society - June 2004 The RealityGrid project Mission: “Using Grid technology to closely couple high performance computing, high throughput experiment and visualization, RealityGrid will move the bottleneck out of the hardware and back into the human mind.” to predict the realistic behavior of matter using diverse simulation methods LB3D - highly scalable grid based code to model dynamics and hydrodynamics of complex multiphase fluids mesoscale simulations enables access to larger physical and longer timescales RealityGrid environment enables multiple steered and spawned simulations, the visualised output being streamed to a distributed set of collaborators located at AG nodes across the USA and UK.

ANL Royal Society - June 2004 Testbed and Project Partners Reality Grid partners: –University College London (Application, Visualisation, Networking) –University of Manchester (Application, Visualisation, Networking) –Edinburgh Parallel Computing Centre (Application) –Tufts University (Application) Teragrid sites at: –Argonne National Laboratory (Visualization, Networking) –National Center for Supercomputing Applications (Compute) –Pittsburgh Supercomputing Center (Compute, Visualisation) –San Diego Supercomputer Center (Compute) UK High-End Computing Services - HPCx run by the University of Edinburgh and CCLRC Daresbury Laboratory (Compute, Networking, Coordination) - CSAR run by the University of Manchester and CSC (Compute and Visualisation)

ANL Royal Society - June 2004 Computer Servers ~ 7 TB memory - 5K processors in integrated resource The TeraGyroid project has access to a substantial fraction of the world's largest supercomputing resources, including the whole of the UK's supercomputing facilities and the USA's TeraGrid machines. The largest simulations are in excess of one billion lattice sites.

ANL Royal Society - June 2004 Networking Cambridge Newcastle Edinburgh Oxford Glasgow Manchester Cardiff Southampton London Belfast DL RAL TeraGrid UK Amsterdam BT provision Netherlight

ANL Royal Society - June 2004 Applications Porting LB3D written in Fortran90 Order 128 variables per grid point 1Gpoint = 1TB Various compiler issues to be overcome at different sites Site configuration issues important eg I/O access to high speed global file systems for checkpoint files Connectivity of high-speed file systems to network Multi heading required of several systems to separate control network from data network Port forwarding required for compute nodes on private network

ANL Royal Society - June 2004 Exploring parameter space through computational steering Initial condition: Random water/ surfactant mixture. Self-assembly starts. Rewind and restart from checkpoint. Lamellar phase: surfactant bilayers between water layers. Cubic micellar phase, low surfactant density gradient. Cubic micellar phase, high surfactant density gradient.

ANL Royal Society - June 2004 Reality Grid - Environment Computations run at HPCx, CSAR, SDSC, PSC and NCSA Visualisation run at Manchester, UCL, Argonne, NCSA, Phoenix Scientists steering calculations from UCL and Boston over Access Grid Visualisation output and collaborations multicast to Phoenix and visualised on the show floor in the University of Manchester booth

ANL Royal Society - June 2004 Visualisation servers Amphiphilic fluids produce exotic mesophases with a range of complex morphologies - need visualisation The complexity of these data sets (128 variables) makes visualisation a challenge Using the VTK library, with patches refreshing each time new data available Video stream multicast to Access Grid using FLXmitter library SGI OpenGL Vizserver used to allow remote control of visualisation Visualisation of billion node models requires 64-bit hardware and multiple rendering units Achieved visualisation of lattice using ray- tracing algorithm developed at University of Utah on 100 proc Altix on showroom floor at SC’03

ANL Royal Society - June 2004 Grid Software Infrastructure Various versions of Globus Toolkit 2.2.3, 2.2.4, and 3.1 (including GT 2 compatibility bundles) Used GRAM, GridFTP Globus-I/O - no incompatibilities Not use MDS- robustness/ utility of data 64 bit version of GT2 required for AIX (HPCx) system - some grief due to tendency to require custom-patched versions of third party libraries Lot of system management effort required to work with/ around toolkit Need a more scalable CA system that bypasses every system administrator having to study everyone else’s certificates

ANL Royal Society - June 2004 TeraGyroid Network

ANL Royal Society - June 2004 VizEng2 PHOENIX SimEng1 UK SimEng2 PSC Disk1 UK Networking

ANL Royal Society - June 2004 Networking On-line visualization requires O(1 Gbps) bandwidth for larger problem sizes Steering requires 100% reliable near-real time data transport across the Grid to visualization engines. Reliable transfer is achieved using TCP/IP: handshaking for each single packet that is transferred (to check and repair loss). This slows down transport  limits data transfer rates  limits LB3D steering of larger systems. Point-to-n-point transport for visualization, storage and job migration uses n times more bandwidth since unicast is used.

ANL Royal Society - June 2004 What Was Done? The TeraGyroid experiment represents the first use of collaborative, steerable, spawned and migrated processes based on capability computing. –generated 2TB of data –exploration of the multi-dimensional fluid coupling parameter space with 64 3 simulations accelerated through steering –study of finite size periodic boundary condition effects, exploring the stability of the density of defects in the 64 3 simulations as they are scaled up to 128 3, 256 3, 512 3, –100K to 1,000K time steps –exploring the stability the crystalline phases to perturbations and variations in effective surfactant temperature and simulations - clear of finite size effects Perfect crystal not formed in systems - 600K steps Statistics of number of defects, velocity and lifetimes requires large systems as these have sufficient defects

ANL Royal Society - June 2004 World’s Largest Lattice Boltzmann Simulation? lattice sites scale up simulations with periodic tiling and perturbations for initial state Finite-size effect free dynamics 2048 processors 1.5 TB of memory 1 minute per time step on 2048 processors 3000 time steps 1.2TB of visualisation data Run on LeMieux at Pittsburgh SC

ANL Royal Society - June 2004 Access Grid Screen at SC ‘03 during SC Global Session on Application Steering

ANL Royal Society - June 2004 Measured Transatlantic Bandwidths during SC’03

ANL Royal Society - June 2004 Demonstrations/ Presentations Demonstrations of the TeraGyroid experiment at SC’03: TeraGyroid on the PSC Booth Tue 18, 10:00-11:00 Thu 20, 10:00-11:00 RealityGrid and TeraGyroid on UK e-Science Booth Tue 18, 16:00-16:30 Wed 19, 15:30-16:00 RealityGrid during the SC'03 poster session: Tue 18, 17:00-19:00 HPC-Challenge presentations: Wed 19 10:30-12:00 SC Global session on steering: Thu 20, 10:30-12:00 Demonstrations and real-time output at the University of Manchester and HPCx booths.

ANL Royal Society - June 2004 Most Innovative Data Intensive Application - SC 03

ANL Royal Society - June 2004 Project Objectives - How Well Did We Do? - 1 world class computational science experiment –science analysis is ongoing - leading to new insights into properties of complex fluids at unprecedented scales –SC’03 award - ‘Most Innovative Data Intensive App’ enhanced expertise/ experience to benefit UK and USA –first transatlantic federation of major HEC facilities –applications need to be adaptable to different architectures inform construction/operation of national/ int grids –most insight gained into end to end network integration, performance and dual homed systems –remote visualisation, steering and checkpointing require high bandwidth which is dedicated and reservable –results fed directly into ESLEA proposal to exploit UKLight optical switched network infrastructure stimulate long-term strategic technical collaboration –strengthened relationships between Globus, networking and visualisation groups

ANL Royal Society - June 2004 Project Objectives - How Well Did We Do? - 2 support long-term scientific collaborations –built on strong and fruitful existing scientific collaborations between researchers in UK and USA experiments with clear scientific deliverables - an explicit science plan was published, approved and then executed. Data analysis is ongoing. choice of applications to be based on community codes –experiences will be of benefit to other grid based applications in particular in the computation engineering community inform future programme of complementary experiments –Report to be made available on RG Website –EPSRC Initiating another Call for Proposals - not targetting SC’04.

ANL Royal Society - June 2004 Lessons Learned How to support such projects - full peer review? Timescales were very tight - September - November Resource estimates need to be flexible Need complementary experiments for US and UK to reciprocate benefits HPC centres/ e-science and networking groups can work very effectively together on challenging common goals Site configuration issues very important - network access Visualisation capabilities in UK need upgrading Scalable CA, dual address systems Network QoS very important for checkpointing, remote steering and visualisation Do it again?