Distributed storage system with dCache E.Y.Kuklin, A.V.Sozykin, A.Y.Bersenev Institute of Mathematics and Mechanics UB RAS, Yekaterinburg G.F.Masich Institute.

Slides:



Advertisements
Similar presentations
Jens G Jensen Atlas Petabyte store Supporting Multiple Interfaces to Mass Storage Providing Tape and Mass Storage to Diverse Scientific Communities.
Advertisements

SALSA HPC Group School of Informatics and Computing Indiana University.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Infrastructure overview Arnold Meijster &
Institute for High Energy Physics ( ) NEC’2007 Varna, Bulgaria, September Activities of IHEP in LCG/EGEE.
Sergey Belov, LIT JINR 15 September, NEC’2011, Varna, Bulgaria.
Bondyakov A.S. Institute of Physics of ANAS, Azerbaijan JINR, Dubna.
Peter Stefan, NIIF 29 June, 2007, Amsterdam, The Netherlands NIIF Storage Services Collaboration on Storage Services.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Sergey Belov, Tatiana Goloskokova, Vladimir Korenkov, Nikolay Kutovskiy, Danila Oleynik, Artem Petrosyan, Roman Semenov, Alexander Uzhinskiy LIT JINR The.
Report : Zhen Ming Wu 2008 IEEE 9th Grid Computing Conference.
Distributed High Performance Computing Environment of Ural Branch of RAS M.L.Goldshtein, A.V.Sozykin, Institute of Mathematics and Mechanics UrB RAS, Yekaterinburg.
E-Infrastructure hierarchy Networking and Computational facilities in Armenia ASNET AM Network Armenian National Grid Initiative Armenian ATLAS site (AM-04-YERPHI)
A comparison of distributed data storage middleware for HPC, GRID and Cloud Mikhail Goldshtein 1, Andrey Sozykin 1, Grigory Masich 2 and Valeria Gribova.
Rensselaer Why not change the world? Rensselaer Why not change the world? 1.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
SoCal Infrastructure OptIPuter Southern California Network Infrastructure Philip Papadopoulos OptIPuter Co-PI University of California, San Diego Program.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Large Scale Parallel File System and Cluster Management ICT, CAS.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
HEPix April 2006 NIKHEF site report What’s new at NIKHEF’s infrastructure and Ramping up the LCG tier-1 Wim Heubers / NIKHEF (+SARA)
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
Production Grid Challenges in Hungary Péter Stefán Ferenc Szalai Gábor Vitéz NIIF/HUNGARNET.
GridFTP GUI: An Easy and Efficient Way to Transfer Data in Grid
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
Lofar Information System on GRID A.N.Belikov. Lofar Long Term Archive Prototypes: EGEE Astro-WISE Requirements to data storage Tiers Astro-WISE adaptation.
Padova, 5 October StoRM Service view Riccardo Zappi INFN-CNAF Bologna.
Configuring, Managing and Maintaining Windows Server® 2008 Servers Course 6419A.
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Ole’ Miss DOSAR Grid Michael D. Joy Institutional Analysis Center.
Evangelos Markatos and Charalampos Gkikas FORTH-ICS Athens, th Mar Institute of Computer Science - FORTH Christos.
The VOMS and the SE in Tier2 Presenter: Sergey Dolgobrodov HEP Meeting Manchester, January 2009.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
the project of the voluntary distributed computing ver.4.06 Ilya Kurochkin Institute for information transmission problem Russian academy of.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
KIT – Universität des Landes Baden-Württemberg und nationales Forschungszentrum in der Helmholtz-Gemeinschaft Steinbuch Centre for Computing
Report on availability of the JINR networking, computing and information infrastructure for real data taking and processing in LHC experiments Ivanov V.V.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Large data storage (examples, volumes, challenges) Cluster, Grid, Clouds – Julien Dhallenne.
Brief introduction about “Grid at LNS”
CASTOR: possible evolution into the LHC era
Enterprise Architectures
GridOS: Operating System Services for Grid Architectures
Cumulus - dynamic cluster available under Clusterix
Overview of the Training Agenda
Data services on the NGS
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
Distributed Network Traffic Feature Extraction for a Real-time IDS
Heterogeneous Computation Team HybriLIT
Experiences with http/WebDAV protocols for data access in high throughput computing
Data services on the NGS
Introduction to Data Management in EGI
5th DOSAR Workshop Louisiana Tech University Sept. 27 – 28, 2007
Clouds of JINR, University of Sofia and INRNE Join Together
Study course: “Computing clusters, grids and clouds” Andrey Y. Shevel
Southwest Tier 2.
Christof Hanke, HEPIX Spring Meeting 2008, CERN
Introduction to Networks
Enabling High Speed Data Transfer in High Energy Physics
Introduction to Networks
Large Scale Test of a storage solution based on an Industry Standard
Storage Networks and Storage Devices
Hadoop Technopoints.
The Anatomy and The Physiology of the Grid
Cost Effective Network Storage Solutions
GridTorrent Framework: A High-performance Data Transfer and Data Sharing Framework for Scientific Computing.
A Virtual Machine Monitor for Utilizing Non-dedicated Clusters
Presentation transcript:

Distributed storage system with dCache E.Y.Kuklin, A.V.Sozykin, A.Y.Bersenev Institute of Mathematics and Mechanics UB RAS, Yekaterinburg G.F.Masich Institute of Сontinuous Мedia Мechanics UB RAS, Perm Distributed Computing and Grid-technologies in Science and Education, Dubna, 2014

2 Introduction  High-performance distributed computing environment of UB RAS: 40 Research Institutes of UB RAS in 7 regions (Yekaterinburg, Arkhangelsk, Syktyvkar, Orenburg, Perm, Izhevsk, Chelyabinsk). Universities. Industrial enterprises.  Distributed data storage system.

3 Introduction  Project participants: Institute of Mathematics and Mechanics UB RAS (IMM UB RAS) – computational recourses and storage system. Institute of Сontinuous Мedia Мechanics UB RAS (ICMM UB RAS) – backbone networks.  Supported by grant of UB RAS 12-P and by the RFBR grant r_ural_a.

4 Overview  About dCache. Middleware selection.  Current implementation. Some problems.  Future plans.

5 About dCache  Distributed storage system. Experimental data.  Commodity hardware. Facilities in hundreds of terabytes.  Single file system tree.  Simple storage extension.

6 About dCache  Access protocols: FTP WebDAV NFS 4.1 SRM gridFTP dCap

7 Middleware selection dCacheDPM StoRM European Middleware Initiative (EMI)

8 Middleware selection  Support for both types of protocols.  High quality documentation.  Wide community.  Easy to install and administer.

9 Middleware selection JINR NDGF

10 Current implementation

Current implementation Plan 456 km 10 GE ICMM UB RAS, Perm IMM UB RAS, Yekaterinburg «URAN» supercomputer dCache broker host, pool1a, pool1b dCache pool2a, pool2b dCache pool3a, pool3b dCache pool4 NFS v4.1 DSS UB RAS «Triton» cluster NFS v4.1 Experimental setups

12 Current implementation Hardware  Supermicro servers: 4 storage nodes. Scientific Linux 6.5. Intel Xeon CPU E HDD Seagate SATA3 3Tb. 210 Tb usable capacity.

13 Current implementation File system XFSEXT4  Choice between XFS and EXT4. We need good I/O performance. The file system should be natively supported by Linux kernel.

Current implementation File system XFSEXT4 14

15  456 km distance.  DWDM optical fiber. Operated by ECI-Telecom.  10GE encapsulated in 2x λ-channels.  NFS v4.1. Current implementation Backbone network Laboratory of Telecommunication and Information Systems ICMM

16 Current implementation Problems  Long distance.  Network stack configuration.  Intel network cards drawbacks.  Custom-built kernel.

17 Current implementation Computational resources  IMM UB RAS: the “URAN” supercomputer. 7 th position in Top50 (CIS).  ICMM UB RAS: the “Triton” cluster. “URAN”

Current implementation Network benchmark

Current implementation Network benchmark

20 Current implementation Possible replication Data2 Data1 IMM UB RAS ICMM UB RAS Data1 Data2 Replication [DWDM channel] “URAN” Experimental setups NFS 4.1

21 Future plans  Near future: Performance and fault tolerance benchmarks. More concern on security and monitoring.  Long-term plans: Increasing of the storage system volume. Integration with experimental setups in ICMM UB RAS and UrFU.

22 Future plans  Joining GRID projects: National Nanotechnology Network (GridNNN). Worldwide Large Hadron Collider Computing Grid (WLCG).

Thank you for your attention. Evgheniy Kuklin IMM UB RAS