Enabling Grids for E-sciencE CRESCO COMPUTATIONAL RESOURCES AND ITS INTEGRATION IN ENEA-GRID.

Slides:



Advertisements
Similar presentations
ITALIAN NATIONAL AGENCY FOR NEW TECNOLOGY, ENERGY AND THE ENVIRONMENT DIESIS Meeting – Portici
Advertisements

ENEA-GRID, Incontro GARR, Roma 15/7/2004 Incontro GARR 15 Luglio 2004, CRUI, ROMA Armonizzazione delle strutture di rete e delle griglie computazionali.
ITALIAN NATIONAL AGENCY FOR NEW TECNOLOGY, ENERGY AND THE ENVIRONMENT CRESCO Seminario Portici 05 Febbraio 2009.
MAGO Monitoring All Grid Objects Anna Jannace 1, Carmine Spizuoco 1, Francesco Adinolfi 1, Giovanni Bracco 2 1- Consorzio Campano per lInformatica e lAutomazione.
Computing Infrastructure
G. Bracco Evolution ENEAGRID/CRESCO - EERA-SP4 Bologna 15/5/2013 Incontro EERA-SP4 Bologna 15 maggio 2013 The evolution of ENEAGRID/CRESCO HPC infrastructure.
HPC in Poland Marek Niezgódka ICM, University of Warsaw
CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
Page 1 Dorado 400 Series Server Club Page 2 First member of the Dorado family based on the Next Generation architecture Employs Intel 64 Xeon Dual.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Information Technology Center Introduction to High Performance Computing at KFUPM.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
ENEA-GRID and gLite Interoperability: robustness of SPAGO approach Catania, Italy, February ENEA-GRID and gLite Interoperability: robustness.
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
HIGH PERFORMANCE COMPUTING ENVIRONMENT The High Performance Computing environment consists of high-end systems used for executing complex number crunching.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
Silicon Graphics, Inc. Poster Presented by: SGI Proprietary Technologies for Breakthrough Research Rosario Caltabiano North East Higher Education & Research.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
An Introduction to Princeton’s New Computing Resources: IBM Blue Gene, SGI Altix, and Dell Beowulf Cluster PICASso Mini-Course October 18, 2006 Curt Hillegas.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
5.3 HS23 Blade Server. The HS23 blade server is a dual CPU socket blade running Intel´s new Xeon® processor, the E5-2600, and is the first IBM BladeCenter.
F1031 COMPUTER HARDWARE CLASSES OF COMPUTER. Classes of computer Mainframe Minicomputer Microcomputer Portable is a high-performance computer used for.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Cloud Computing in ENEA-GRID: Virtual Machines, Roaming Profile, and Online Storage Ing. Giovanni Ponti, Ph.D. ENEA – UTICT-HPC
HPC at IISER Pune Neet Deo System Administrator
Corporate Partner Overview and Update September 27, 2007 Gary Crane SURA Director IT Initiatives.
Copyright 2009 Fujitsu America, Inc. 0 Fujitsu PRIMERGY Servers “Next Generation HPC and Cloud Architecture” PRIMERGY CX1000 Tom Donnelly April
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
SA1 / Operation & support Enabling Grids for E-sciencE Integration of heterogeneous computational resources in.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
Enabling Grids for E-sciencE CRESCO HPC SYSTEM INTEGRATED INTO ENEA GRID ENVIRONMENT G. Bracco, S.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Rensselaer Why not change the world? Rensselaer Why not change the world? 1.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
HPCVL High Performance Computing Virtual Laboratory Founded 1998 as a joint HPC lab between –Carleton U. (Comp. Sci.) –Queen’s U. (Engineering) –U. of.
Looking Ahead: A New PSU Research Cloud Architecture Chuck Gilbert - Systems Architect and Systems Team Lead Research CI Coordinating Committee Meeting.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
Rob Allan Daresbury Laboratory NW-GRID Training Event 25 th January 2007 Introduction to NW-GRID R.J. Allan CCLRC Daresbury Laboratory.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
Integration of heterogeneous computational resources in EGEE: a live demo Istanbul, Turkey, September Integration of heterogeneous computational.
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Background Computer System Architectures Computer System Software.
Hardware Architecture
ENEA GRID & JPNM WEB PORTAL to create a collaborative development environment Dr. Simonetta Pagnutti JPNM – SP4 Meeting Edinburgh – June 3rd, 2013 Italian.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
Jefferson Lab Site Report Sandy Philpott HEPiX Fall 07 Genome Sequencing Center Washington University at St. Louis.
CIT 140: Introduction to ITSlide #1 CSC 140: Introduction to IT Operating Systems.
NIIF HPC services for research and education
Accessing the VI-SEEM infrastructure
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
Buying into “Summit” under the “Condo” model
Clouds , Grids and Clusters
UK GridPP Tier-1/A Centre at CLRC
CRESCO Project: Salvatore Raia
Computer Basics Section 2.1 YOU WILL LEARN TO… Identify hardware
Support for ”interactive batch”
MAGO Monitoring All Grid Objects
EFDA Meeting – Portici ITALIAN NATIONAL AGENCY
Presentation transcript:

Enabling Grids for E-sciencE CRESCO COMPUTATIONAL RESOURCES AND ITS INTEGRATION IN ENEA-GRID ENVIRONMENT G. Bracco, S. Podda, S. Migliori, P. D'angelo, A. Quintiliani, D. Giammattei, M. De Rosa, S. Pierattini, G. Furini, R. Guadagni, F. Simoni, A. Perrozziello, A. De Gaetano, S. Pecoraro, A. Santoro, C. Sciò*, A. Rocchi**, A. Funel, S. Raia, G. Aprea, U. Ferrara, F. Prota, D. Novi, G. Guarnieri ENEA-FIM, L.Tevere Thaon di Revel (Roma) Italy, (*) Esse3Esse (**) CASPUR Summary ENEA, the Italian agency for the energy, environment and new technologies, has a substantial experience in GRID technologies and its multi-platform HPC resources are integrated in the ENEA-GRID infrastructure. This poster describes the architecture of the high performance computing (HPC) system that has been installed to provide the required computing power to the CRESCO project applications and the dedicated activity required to integrate CRESCO HPC system into ENEA-GRID infrastructure. CRESCO (Computational Research Center for Complex Systems) is an ENEA Project, co-funded by the Italian Ministry of University and Research (MUR). The project is functionally built around a HPC platform and 3 scientific thematic laboratories: the Computing Science Laboratory, hosting activities on HW and SW design, GRID technology and HPC platform management the Computational Systems Biology Laboratory, with activities in the Life Science domain, ranging from the “post-omic” sciences (genomics, interactomics, metabolomics) to Systems Biology; the Complex Networks Systems Laboratory, hosting activities on complex technological infrastructures, for the analysis of Large National Critical Infrastructures CRESCO HPC system consists of more then 2700 computing cores, divided into three main sections. A section is dedicated to applications with high memory requirements ( 42 nodes with 16 cores and 32 or 64 GB memory for a total of 672 cores), a section dedicated to high scalable applications (256 nodes with 8 cores and 16 GB memory, for a total of 2048 cores) and a third experimental section providing systems with Cell processors (4 blades), FPGA (6 VIRTEX systems) and high performance video adapters (4 NVIDIA FX 4500 X2 systems) dedicated to computational applications. High bandwidth and low latency connections are provided by an InfiniBand 4xDDR network. The main storage consists of an IBM/DDN 9550 system with 160 TB raw data, organized in a GPFS file system. CRESCO HPC system has been integrated into ENEA-GRID infrastructure which has been developed to provide a unified environment for all the main ENEA HPC resources. The main software components of ENEA-GRID are the multi-site resource manager LSF Multicluster, the OpenAFS distributed file system, the integrated Kerberos 5 authentication and a Java and Web based Graphical User interface making use of CITRIX technologies. The choice of mature, reliable and multi-platform software components has permitted along the years to integrate in a GRID oriented infrastructure HPC resources at the state of the art performances, with minimal changes in the user environment. ENEA ENEA-GRID computational resources Hardware (before CRESCO HPC system!!): ~100 hosts and ~650 cpu : IBM SP; SGI Altix & Onyx; Linux clusters 32/ia64/x86_64; Apple cluster; Windows servers. Most relevant resources: IBM SP5 258 cpu; 3 frames of IBM SP4 96 cpu Software: Commercial codes (fluent, ansys, abaqus..) Research codes. (mcpn/x, eranos, fluka... Elaboration environments (Matlab, IDL, Scilab...) Windows Applications 12 Research sites and a Central Computer and Network Service with 6 computer centres managing multi-platform resources for serial & parallel computation and graphical post processing. ENEA GRID GRID functionalities (unique authentication, authorization, resource access and resource discovery) are provided using “mature”, multi-platform components: Distributed File System: OpenAFS Resource Manager: LSF Multicluster [ Unified user interface: Java & Citrix Technologies These components constitute the ENEA-GRID Middleware. OpenAFS user homes, software and data distribution integration with LSF user authentication/authorization, Kerberos V ENEA GRID architecture CRESCO HPC system CRESCO HPC system has been designed with the aim of offering a general purpose facility based on the leading multi-core x86_64 technology. The performance for the CRESCO HPC plant set-up has ranked #180 in the Nov top500 list with Rmax=9.3 TeraFlops (rank #3 between the Italian HPC systems in the list). In order to provide the best environment for different types of applications the system consists of two main sections respectively oriented (1) for high memory request and moderate parallel scalability and (2) for limited memory and high scalability cases. Both sections are interconnected by a common Infiniband 4X DDR network (IB) and can operate as a single large integrated system. The first main section is composed by 42 fat nodes IBM x3850-M2 with 4 Xeon Quad-Core Tigerton E7330 processors (2.4GHz/1066MHz/6MB L2), 32 MB RAM (4 extra-fat nodes with 64 GB RAM). The total number of cores in the first section is then equal to 672. The second main section is composed by 256 blades IBM HS21 each supporting dual Xeon Quad-Core Clovertown E5345 processors (2.33GHz/1333MHz/8MB L2), 8 GB RAM (16 blades with 16 GB RAM) for total of 2048 cores. The blades are hosted by the14 slots blades chassis for a total of 19 chassis and each blade has a dedicated IB connection. The larger system created by joining the two main sections is has 2720 cores. A third experimental section consists of 3 subsections dedicated to special processor architectures: 4 blades IBM QS21 with 2 Cell BE Processors 3.2 Ghz each. 6 nodes IBM x3755, 4 sockets AMD Dualcore 8222 equipped with a FPGA VIRTEX5 LX330 card 4 node IBM x 3755, 4 sockets AMD Dualcore 8222 with a NVIDIA Quadro FX 4500 X2 video card The IB network is based on a CISCO SFS 7024 (288 ports), a CISCO SFS 7012 (144 ports) and 5 CISCO SFS 7000 (120 ports). The Ethernet network consists of one CISCO 4506 (240 ports), 3 CISCO 4948 (144 ports) and 3 CISCO 3750G (144 ports). The storage of CRESCO HPC system is provided by an IBM DCS9550 system, 160 TB raw space based on 500 GB SATA Hard Disk. An IBM Tape Library IBM TS3500 provides the backup facility. The power required to run the system has been estimated to 150 kw and proper cooling systems have been provided. The operating system is RedHat EL 5.1 and the usual set of Portland and Intel Compilers are available. A GPFS parallel file system is shared via Infiniband between all the computing nodes of all the main section of the system. User homes are located in an OpenAFS file system, one of the base elements of the ENEA-GRID infrastructure. The three sections together with other 35 service machines (front-end, controls, file-servers, installation servers) and storage and network components make use of a total of 18 standard racks (19”, 42 U). SFS 7000D (D) SFS 7024D 256 HS21 XM + 4 QS21 Section 2 & 3 SFS 7000D (E) (3) (4) DDN with FC controllers (8 x FC 4Gbps) (1) SFS 7000D (1) IBM 3650 I/O node IBM 3650 I/O node GPFS SM (active)SM (standby) SFS 7000D (1) SFS7012D SFS 7000D (3) (4) (1) IBM 3650 I/O node IBM 3650 I/O node Section 1 42 x3850M2 +FPGA 6 +NVIDIA 4 +Backup 3 +Front end 8 +Graph FE 8 Total 71 ENEA GRID mission [started 1999]: provide an unified user environment and an homogeneous access method for all ENEA researchers and their collaborators, irrespective of their location. optimize the utilization of the available resources Architecture of the InfiniBand network including the IBM/DDN 9550 storage system. The 4 I/O Nodes, directly FC attached to the storage, are the GPFS NSD servers. ENEA computational resources are distributed over WAN, connected by GARR, the Italian Academic & Research Network ENEA Network connection ENEA-GARR 9 PoP, Mbps Brindisi 150 Mb/s Bologna 30 Mb/s Casaccia 100 Mb/s Frascati 155 Mb/s Portici 400 Mb/s Trisaia 18 Mb/s Palermo Pisa Roma Sede ENEA-GRID USER INTERFACE ENEA GRID makes use of Citrix Metaframe to publish an application providing all the available resources and monitoring facilities with a unified GUI interface GUI Application components: Java (GUI) shell scripts Windows services Unix/Linux Services ENEA GRID Web Access