Www.consorzio-cometa.it FESR Consorzio COMETA Giuseppe Andronico INFN Sez. CT & Consorzio COMETA Workshop Grids vs. Clouds Beijing, 18.05.2011 Consorzio.

Slides:



Advertisements
Similar presentations
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Advertisements

SCARF Duncan Tooke RAL HPCSG. Overview What is SCARF? Hardware & OS Management Software Users Future.
Beowulf Supercomputer System Lee, Jung won CS843.
High Performance Computing Course Notes Grid Computing.
Consorzio COMETA - Progetto PI2S2 UNIONE EUROPEA HPC Applications on the Sicilian Grid Infrastructure Marcello Iacono-Manno
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 6 2/13/2015.
Information Technology Center Introduction to High Performance Computing at KFUPM.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
INTRODUCTION TO CLOUD COMPUTING Cs 595 Lecture 5 2/11/2015.
1 Challenges Facing Modeling and Simulation in HPC Environments Panel remarks ECMS Multiconference HPCS 2008 Nicosia Cyprus June Geoffrey Fox Community.
1 1 Hybrid Cloud Solutions (Private with Public Burst) Accelerate and Orchestrate Enterprise Applications.
Quantitative Methodologies for the Scientific Computing: An Introductory Sketch Alberto Ciampa, INFN-Pisa Enrico Mazzoni, INFN-Pisa.
Connecting OurGrid & GridSAM A Short Overview. Content Goals OurGrid: architecture overview OurGrid: short overview GridSAM: short overview GridSAM: example.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
M.A.Doman Short video intro Model for enabling the delivery of computing as a SERVICE.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
High Performance Computing on the GRID Infrastructure of COMETA S. Orlando 1,2, G. Peres 3,1,2, F. Reale 3,1,2, F. Bocchino 1,2, G.G. Sacco 2, M. Miceli.
WNoDeS – Worker Nodes on Demand Service on EMI2 WNoDeS – Worker Nodes on Demand Service on EMI2 Local batch jobs can be run on both real and virtual execution.
FESR Consorzio COMETA - Progetto PI2S2 The COMETA consortium and its activities for Grid adoption by Industry in the context of.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
E-science grid facility for Europe and Latin America Porting a program to run on the Grid Marcello Iacono Manno Consorzio COMETA
FP6−2004−Infrastructures−6-SSA EUChinaGrid status report Giuseppe Andronico INFN Sez. Di Catania CERN – March 3° 2006.
Computing Research Testbeds as a Service: Supporting large scale Experiments and Testing SC12 Birds of a Feather November.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) Giuseppe Andronico INFN Sez. CT / Consorzio COMETA Beijing,
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
HPC HPC-5 Systems Integration High Performance Computing 1 Application Resilience: Making Progress in Spite of Failure Nathan A. DeBardeleben and John.
GRID ANATOMY Advanced Computing Concepts – Dr. Emmanuel Pilli.
FESR Consorzio COMETA - Progetto PI2S2 GRID Success Stories in Sicily (The TriGrid VL and PI2S2 Projects) Roberto Barbera University.
FESR Consorzio COMETA - Progetto PI2S2 Porting a program to run on the Grid Marcello Iacono Manno Consorzio COMETA
Università di Perugia Enabling Grids for E-sciencE Status of and requirements for Computational Chemistry NA4 – SA1 Meeting – 6 th April.
OPTIMIZATION OF DIESEL INJECTION USING GRID COMPUTING Miguel Caballer Universidad Politécnica de Valencia.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Spanish National Research Council- CSIC Isabel.
OpenNebula: Experience at SZTAKI Peter Kacsuk, Sandor Acs, Mark Gergely, Jozsef Kovacs MTA SZTAKI EGI CF Helsinki.
The FI-WARE Project – Base Platform for Future Service Infrastructures FI-WARE OCTOBER 2011 Presentation at proposers day.
FESR Consorzio COMETA - Progetto PI2S2 PI2S2 and S-Sicilia towards the industrial world: services and applications Prof. Antonio.
Co-ordination & Harmonisation of Advanced e-Infrastructures for Research and Education Data Sharing Grant.
Tutorial on Science Gateways, Roma, Catania Science Gateway Framework Motivations, architecture, features Riccardo Rotondo.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
Introducing: Chengdu’s Industrial Cloud Huawei & GDS Services Industrial Cloud Solution for SMEs Author/ID: Zhao Zhijuan/ Dept: Industry Solutions.
INTRODUCTION TO GRID & CLOUD COMPUTING U. Jhashuva 1 Asst. Professor Dept. of CSE.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Gang Chen, Institute of High Energy Physics Feb. 27, 2012, CHAIN workshop,Taipei Co-ordination & Harmonisation of Advanced e-Infrastructures Research Infrastructures.
INFN/IGI contributions Federated Clouds Task Force F2F meeting November 24, 2011, Amsterdam.
G. Russo, D. Del Prete, S. Pardi Kick Off Meeting - Isola d'Elba, 2011 May 29th–June 01th A proposal for distributed computing monitoring for SuperB G.
A worldwide e-Infrastructure and Virtual Research Community for NMR and structural biology Alexandre M.J.J. Bonvin Project coordinator Bijvoet Center for.
FESR Trinacria Grid Virtual Laboratory The Trinacria Grid Virtual Laboratory Roberto Barbera University of Catania and INFN Grid Open Day.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) Giuseppe Andronico INFN Sez. CT & Consorzio COMETA Workshop Clouds.
HPC need and potential of ANSYS CFD and mechanical products at CERN A. Rakai EN-CV-PJ2 5/4/2016.
Evolution at CERN E. Da Riva1 CFD team supports CERN development 19 May 2011.
SuperB – Naples Site Dr. Silvio Pardi. Right now the Napoli Group is employed in 3 main tasks relate the computing in SuperB Fast Simulation Electron.
Enabling Grids for E-sciencE University of Perugia Computational Chemistry status report EGAAP Meeting – 21 rst April 2005 Athens, Greece.
Brief introduction about “Grid at LNS”
COMETA Sara Pirrone INFN.
Accessing the VI-SEEM infrastructure
Review of the WLCG experiments compute plans
Current Generation Hypervisor Type 1 Type 2.
Grid Projects in Sicily
SuperComputing 2003 “The Great Academia / Industry Grid Debate” ?
Recap: introduction to e-science
Infrastructure for testing accelerators and new
University of Technology
Cloud Computing Dr. Sharad Saxena.
H2020 EU PROJECT | Topic SC1-DTH | GA:
Presentation transcript:

FESR Consorzio COMETA Giuseppe Andronico INFN Sez. CT & Consorzio COMETA Workshop Grids vs. Clouds Beijing, Consorzio COMETA: MPI use case

Outline Consorzio COMETA Specific solutions CFD & MPI on Consorzio COMETA infrastructure Some results Conclusions

3 The Italian e-Infrastructure (interoperable with common communication protocols ) The Network FESR The Grid

Objectives of an e-Infrastructure in Sicily 4 Create a Virtual Laboratory in Sicily,both for scientific and industrial applications, built on top of a Grid infrastructure Connect the Sicilian e-Infrastructure to those already existing in Italy, in Europe, and in the rest of the world improving the scientific collaboration and increasing the “competitiveness” of e-Science and e-Industry “made in Sicily” Disseminate the “grid paradigm” through the organization of dedicated events and training courses Trigger/foster the creation of spin-offs in the ICT area in order to reduce the “brain drain” of brilliant young people to other parts of Italy and beyond

cores for HPC 250 TB for storage COMETA Consortium and its e-Infrastructure

The COMETA Infrastructure 6 Catania ~15 M € in 3 years >350 people involved! ~2500 CPU ~250 TB

7 Hardware 1. ~2000 cores AMD Opteron 2218 rev. F 2. 2 GB of RAM per core 3.Commercial LRMS (LSF) 4.Infiniband-4X (for MPI applications) 1. ~ 250+ TB of storage 2.Distributed parallel filesystem (GPFS) gLite 3.2 as Grid middleware everywhere. A deliberate investment on a “de facto” standard.

Catania Computing Room 8 Full Area: ~200 m 2 Area # 1 10 racks / 40 kW UPS/PDU Area # 2 80 kW UPS/PDU Area # 2 13 racks Area # 2 80 kW Air Cond. with ~100 kW external chiller 3D Model of Catania Data Center

9 Infiniband-4x Mt 250 Catania University Campus Infiniband-4x net layer allows latency times of a few  s – ideal for HPC Sensitive performance enhancement with >40 cores Currently under project: of a 1000 core-site inside the Catania University Campus linking INFN-CT UNICT-MATH UNICT-ENGIN INFN-LNS

10 Scheduling Policy Usual Policy –Queues and related longest durations  Short (15 min), long (12 hrs), “infinite” (21 days) jobs –Decreasing priority from short to infinite –All resources available before starting –Too restrictive for HPC jobs: long waiting times HPC: RESERVATION (resources are assigned to the incoming HPC job as soon as they become available) and COALLOCATION (the reserved resources are assigned to SHORT jobs while the HPC is collecting its resources) EMERGENCY: PRE-EMPTION (current job is interrupted and then restored after the completion of the incoming job)

11 GridFlex

Watchdog Tool to monitor job status at runtime. It is made from scripts. Some have to be sent with the job, other will be used from user to retrieve information WMS CEWN AMGA Server JDL: job + watchdog (watchdog.conf, watchdog.ctrl, watchdog.sh) Querying job status using getinfo.ctrl getcontent.sh watchdog

Application distribution & Fluent - Many fields impacted in both academic and business areas - More than 300 users from Italy and abroad

COMETA workflow 1.COMETA people, application developer(s) and application user(s) sitting together to verify application portability 2.If point 1 was OK, then a JDL was developed to run application on grid in the simplest case 3.A more general JDL was then developed to let application usage 4.In some cases a portal was developed to simplify interaction between users and grid environment At this point a user interested in the application connected to the portal, filled the required data to run application, pushed a button to have application running, monitored application status and retrieved results at the end. 14 SaaS

15 Introduction: FLUENT & OpenFOAM FLUENT & OpenFOAM are SW packages for CFD simulations Two different approaches for the same computational field (mainly used for flow modelling, heat and mass transfer simulations)  Fluent is a Commercial Product several libraries included, ready for many architectures  OpenFOAM is an open-source SW easily adaptable by the user

Fluent -Fluent (Computer Fluid Dynamics) -ANSYS software for parallel computations on fluid, heat and mass flows -COMETA has been a leading customer in Italy with >150 licenses -Applications ranged from car design to the Marmore Falls simulation going through the study of heat dispersion in engines and refrigerators 16

17 Fluent & OpenFOAM PDE solvers require long CPU times, large memories A commonly used architecture is MPI/MPI2 with low-latency net layers (InfiniBand) and dedicated compilers (PGI C++/Fortran, Intel, GCC) Both packages run on the COMETA Infrastructure

Fluent use case Users of COMETA interested in running Fluent have problems that require the use of MPI clusters with 128 or 256 nodes for weeks or months. Big enough to switch to dedicate hardware, not enough to move on supercomputers. It is not possible to use virtualization: VMs do NOT support Infiniband nor any other low latency connection Currently it is not possible to use cloud computing, as it is intended nowadays, to provide service to such users Also, 5% loss in performance in a run a month long means 1.5 days more. Not a lot, but cloud computing is paid on a hourly basis 18

Fluent use case In this case using grid was possible to integrate MPI support in the infrastructure Compilers, software and licensing were nicely integrated too Required expert users: Fluent application is obtained writing C/C++ code and compiling against Fluent library With the COMETA solution users can ask for a MPI pool till to about 600 cores. 19 IaaS

20 Marmore Falls Simulation

21 Fluent Video (1/2)

Fluent Video (2/2)

23 Other Applications COMETA hosts several MPI codes:  FLASH is a 3D astrophysical hydrodynamic code for supercomputers used in current astrophysical research  ABINIT is a Molecular PhysicsAb-initio parallel code for molecular clustering  CLUSTAL-W is a Bio-informatics code for molecular affinity  TEPHRA is a Civil Defense code for the forecast of ash-cloud evolution (Etna/Ejafiallajokull)

24 Performance

25 Support

Conclusions We showed as SaaS and IaaS were implemented on Consorzio COMETA infrastructure Also we showed that exist yet cases in which is better to rely on grid computing grid computing can be used to provide a service simple to use and with good QoS 26

27 MPI on the web

28 Questions…