A Scalable Distributed Datastore for BioImaging R. Cai, J. Curnutt, E. Gomez, G. Kaymaz, T. Kleffel, K. Schubert, J. Tafas {jcurnutt, egomez, keith,

Slides:



Advertisements
Similar presentations
Distributed Processing, Client/Server and Clusters
Advertisements

By the end of this section, you will know and understand the hardware and software involved in making a LAN!
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Crossing the Chasm: Sneaking a parallel file system into Hadoop Wittawat Tantisiriroj Swapnil Patil, Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon.
1 GridTorrent Framework: A High-performance Data Transfer and Data Sharing Framework for Scientific Computing.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Emery Berger University of Massachusetts Amherst Operating Systems CMPSCI 377 Lecture.
Multithreaded FPGA Acceleration of DNA Sequence Mapping Edward Fernandez, Walid Najjar, Stefano Lonardi, Jason Villarreal UC Riverside, Department of Computer.
HIGH PERFORMANCE COMPUTING ENVIRONMENT The High Performance Computing environment consists of high-end systems used for executing complex number crunching.
Teraserver Darrel Sharpe Matt Todd Rob Neff Mentor: Dr. Palaniappan.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Presented by Jacob Wilson SharePoint Practice Lead Bross Group 1.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
VMware vCenter Server Module 4.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
UNIVERSITY of MARYLAND GLOBAL LAND COVER FACILITY High Performance Computing in Support of Geospatial Information Discovery and Mining Joseph JaJa Institute.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.

Small File File Systems USC Jim Pepin. Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
Workload-driven Analysis of File Systems in Shared Multi-Tier Data-Centers over InfiniBand K. Vaidyanathan P. Balaji H. –W. Jin D.K. Panda Network-Based.
Data Warehousing 1 Lecture-24 Need for Speed: Parallelism Virtual University of Pakistan Ahsan Abdullah Assoc. Prof. & Head Center for Agro-Informatics.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Cluster Workstations. Recently the distinction between parallel and distributed computers has become blurred with the advent of the network of workstations.
1 Selecting LAN server (Week 3, Monday 9/8/2003) © Abdou Illia, Fall 2003.
Building a Parallel File System Simulator E Molina-Estolano, C Maltzahn, etc. UCSC Lab, UC Santa Cruz. Published in Journal of Physics, 2009.
Diamond Computing Status Update Nick Rees et al..
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
Types of Operating Systems
Computer Systems Lab The University of Wisconsin - Madison Department of Computer Sciences Linux Clusters David Thompson
Architecture for Caching Responses with Multiple Dynamic Dependencies in Multi-Tier Data- Centers over InfiniBand S. Narravula, P. Balaji, K. Vaidyanathan,
Presented by Leadership Computing Facility (LCF) Roadmap Buddy Bland Center for Computational Sciences Leadership Computing Facility Project.
Networks.
1/30/2003 BARC1 Profile-Guided I/O Partitioning Yijian Wang David Kaeli Electrical and Computer Engineering Department Northeastern University {yiwang,
Bioimage database architecture and infrastructure 2005, Bio-ITR, UCSB.
Large Scale Parallel File System and Cluster Management ICT, CAS.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
CLUSTER COMPUTING TECHNOLOGY BY-1.SACHIN YADAV 2.MADHAV SHINDE SECTION-3.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
Queensland University of Technology CRICOS No J VMware as implemented by the ITS department, QUT Scott Brewster 7 December 2006.
Types of Operating Systems 1 Computer Engineering Department Distributed Systems Course Assoc. Prof. Dr. Ahmet Sayar Kocaeli University - Fall 2015.
Communications & Networks National 4 & 5 Computing Science.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Building and managing production bioclusters Chris Dagdigian BIOSILICO Vol2, No. 5 September 2004 Ankur Dhanik.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Juraj Sucik, Michal Kwiatek, Rafal.
Background Computer System Architectures Computer System Software.
AMS02 Software and Hardware Evaluation A.Eline. Outline  AMS SOC  AMS POC  AMS Gateway Computer  AMS Servers  AMS ProductionNodes  AMS Backup Solution.
Intro to Distributed Systems Hank Levy. 23/20/2016 Distributed Systems Nearly all systems today are distributed in some way, e.g.: –they use –they.
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
Configuring SQL Server for a successful SharePoint Server Deployment Haaron Gonzalez Solution Architect & Consultant Microsoft MVP SharePoint Server
An Introduction to GPFS
29/04/2008ALICE-FAIR Computing Meeting1 Resulting Figures of Performance Tests on I/O Intensive ALICE Analysis Jobs.
iSCSI Storage Area Network
CLUSTER COMPUTING Presented By, Navaneeth.C.Mouly 1AY05IS037
Network+ Guide to Networks, Fourth Edition
GridTorrent Framework: A High-performance Data Transfer and Data Sharing Framework for Scientific Computing.
Module 1: Overview of Systems Management Server 2003
Presentation transcript:

A Scalable Distributed Datastore for BioImaging R. Cai, J. Curnutt, E. Gomez, G. Kaymaz, T. Kleffel, K. Schubert, J. Tafas {jcurnutt, egomez, keith, Renaissance Research Labs Department of Computer Science California State University San Bernardino, CA Supported by NSF ITR #

Background CSUSB Institute for Applied Supercomputing Low Latency Communications UCSB Center for BioImage Informatics Retinal images Texture map searches Distributed consortium (UCB, CMU)

Retina Images Normal (n) 3 month detachment (3m) 1 day detachment followed by 6 day reattached with increased oxygen (1d+6dO2) 3 day detachment (3d) Laser scanning confocal microscope images of the retina

Environment UCSB Raven Cluster Image and metadata server search external internal BISQUE features analysis Hammer/Nail Cluster Local LAN CSUSB Image and metadata server Image and metadata server WAN Local Lustre

Software Open source OME Postgresql 7 Bisque Distributed datastore Clustering NFS Lustre Benchmark: OSDB

Hardware - Raven 5 year old dual processor 1.4 GHz Pentium 3 256MB RAM 60GB SCSI Compaq Proliant DL-360 servers. Raven has been latency tuned.

Hardware – Hammer/Nail UCSB CSUSB Hammer headnode 5 Nail nodes quad CPUs 3.2 Ghz Xeon 4GB RAM 140GB SCSI Dell servers Bandwidth tuned (default)

Outline Effect of node configuration Comparison of network file systems Effects of a wide area network (WAN)

Relative LAN Performance

NFS LAN/WAN Performance

Design Effects? A few expert users Metadata searches Small results to user Texture searches Heavy calculation on cluster Small results to user Latency tuning

Outline Effect of node configuration Comparison of network file systems Effects of a wide area network (WAN)

NFS / Luster Performance. NFS well known standard Configuration problems with OME performance comparison of the Lustre file system Lustre Journaling Stripe across multiple computers Data redundancy and failover

Relative Performance on LAN NSF/Lustre Compared to local DB 1GB LAN two significant differences

Significant Differences NSF caching bulk deletes and bulk modifies Lustre stripes across computers increase the bandwidth

Outline Effect of node configuration Comparison of network file systems Effects of a wide area network (WAN)

Effect on Wide Area Network WAN Compared three connections Local Switched, high speed LAN (1 Gb/s) WAN between UCSB and CSUSB (~50 Mb/s) NFS only UCSB didn’t have Lustre installed Active research prevented reinstalling OS

Local/LAN/WAN Performance

Effect on Wide Area Network WAN Most significant effect Not bandwidth intensive operations Latency intensive operation Next generation WAN will not solve the problem. Frequently used data must be kept locally Database cluster Daily sync of remote databases

Conclusions Scientific researchers Latency tune network Don’t bandwidth tune Latency of WAN is too large replicate data and update. Bisque/OME NFS issues Lustre High bandwidth operations Stripe Lustre across systems

Future directions: Agent based texture search engine Loosely coupled cluster WAN connection Unreliable connection Fault tollerant Parallelize Jobs Open source components Scilab Convert NSF funded algorithms in Matlab Simple interface Superior caching scheme for Lustre

Questions…