HPC USER FORUM I/O PANEL April 2009 Roanoke, VA Panel questions: 1 response per question Limit length to 1 slide.

Slides:



Advertisements
Similar presentations
Data Storage Solutions Module 1.2. Data Storage Solutions Upon completion of this module, you will be able to: List the common storage media and solutions.
Advertisements

HPC USER FORUM I/O PANEL April 2009 Roanoke, VA Panel questions: 1 response per question Limit length to 1 slide.
IB in the Wide Area How can IB help solve large data problems in the transport arena.
HEP Data Sharing … … and Web Storage services Alberto Pace Information Technology Division.
11 BACKING UP AND RESTORING DATA Chapter 4. Chapter 4: BACKING UP AND RESTORING DATA2 CHAPTER OVERVIEW Describe the various types of hardware used to.
IDC HPC User Forum Conference Appro Product Update Anthony Kenisky, VP of Sales.
Inclusive Publishing – Reading for Everyone Neil Jarvis 24 November 2014.
An End-to-End Approach to Globally Scalable Network Storage Presented in cs294-4 P2P Systems by Sailesh Krishnamurthy 15 October 2003.
D. Düllmann - IT/DB LCG - POOL Project1 POOL Release Plan for 2003 Dirk Düllmann LCG Application Area Meeting, 5 th March 2003.
File Management Systems
Storage Networking Technologies and Virtualization Section 2 DAS and Introduction to SCSI1.
 The Open Systems Interconnection model (OSI model) is a product of the Open Systems Interconnection effort at the International Organization for Standardization.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
BACKUP/MASTER: Immediate Relief with Disk Backup Presented by W. Curtis Preston VP, Service Development GlassHouse Technologies, Inc.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
By N.Gopinath AP/CSE. Why a Data Warehouse Application – Business Perspectives  There are several reasons why organizations consider Data Warehousing.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
1 Advanced Storage Technologies for High Performance Computing Sorin, Faibish EMC NAS Senior Technologist IDC HPC User Forum, April 14-16, Norfolk, VA.
Operating Systems (CS 340 D) Dr. Abeer Mahmoud Princess Nora University Faculty of Computer & Information Systems Computer science Department.
Silicon Building Blocks for Blade Server Designs accelerate your Innovation.
Small File File Systems USC Jim Pepin. Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Click to Edit Master Subtitle Style HPC USER FORUM I/O PANEL April 2009 Roanoke, VA Panel questions: 1 response per question Limit length to 1 slide.
DOE PI Meeting at BNL 1 Lightweight High-performance I/O for Data-intensive Computing Jun Wang Computer Architecture and Storage System Laboratory (CASS)
The Functions of Operating Systems Desktop PC Operating Systems.
Large Scale Parallel File System and Cluster Management ICT, CAS.
SAN DIEGO SUPERCOMPUTER CENTER SDSC's Data Oasis Balanced performance and cost-effective Lustre file systems. Lustre User Group 2013 (LUG13) Rick Wagner.
Actualog Social PIM Helps Companies to Manage and Share Product Information Using Secure, Scalable Ease of Microsoft Azure MICROSOFT AZURE ISV PROFILE:
Creating a Qt Widget Based Application From Qt Creator Manual.
Built on Azure, Moodle Helps Educators Create Proprietary Private Web Sites Filled with Dynamic Courses that Extend Learning Anytime, Anywhere MICROSOFT.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Welcome to the PVFS BOF! Rob Ross, Rob Latham, Neill Miller Argonne National Laboratory Walt Ligon, Phil Carns Clemson University.
DOCUMENT #:GSC15-PLEN-82r2 FOR:Presentation SOURCE:ATIS AGENDA ITEM: PLEN 6.14 CONTACT(S): Andrew White ATIS’
7 Strategies for Extracting, Transforming, and Loading.
Powered by Microsoft Azure, PointMatter Is a Flexible Solution to Move and Share Data between Business Groups and IT MICROSOFT AZURE ISV PROFILE: LOGICMATTER.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
Evolution of storage and data management Ian Bird GDB: 12 th May 2010.
Highest performance parallel storage for HPC environments Garth Gibson CTO & Founder IDC HPC User Forum, I/O and Storage Panel April 21, 2009.
4P13 Week 12 Talking Points Device Drivers 1.Auto-configuration and initialization routines 2.Routines for servicing I/O requests (the top half)
By Alex Norman and Ricky Brentano.  Contacted us to consult on web hosting / web content management  OS provides a wide variety of athletic & training.
Accelerating High Performance Cluster Computing Through the Reduction of File System Latency David Fellinger Chief Scientist, DDN Storage ©2015 Dartadirect.
Power LogOn® Adds Card-Based, Multi- Factor Authentication to Microsoft Azure Logon, Plus Password Management for All Other Logons MICROSOFT AZURE ISV.
Interconnect Trends in High Productivity Computing Actionable Market Intelligence for High Productivity Computing Addison Snell, VP/GM,
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Flight is a SaaS Solution that Accelerates the Secure Transfer of Large Files and Data Sets Into and Out of Microsoft Azure Blob Storage MICROSOFT AZURE.
Axis AI Solves Challenges of Complex Data Extraction and Document Classification through Advanced Natural Language Processing and Machine Learning MICROSOFT.
Tackling I/O Issues 1 David Race 16 March 2010.
Site Services and Policies Summary Dirk Düllmann, CERN IT More details at
Maximizing Performance – Why is the disk subsystem crucial to console performance and what’s the best disk configuration. Extending Performance – How.
INDIANAUNIVERSITYINDIANAUNIVERSITY Tsunami File Transfer Protocol Presentation by ANML January 2003.
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
Packet processed storage in a software defined world Ash Young fd.io Foundation1.
Mini-Workshop on multi-core joint project Peter van Gemmeren (ANL) I/O challenges for HEP applications on multi-core processors An ATLAS Perspective.
HPC USER FORUM STORAGE PANEL April 15, 2008 Norfolk, VA.
Evolution of storage and data management
Integrating Disk into Backup for Faster Restores
Introduction to Data Management in EGI
Chapter 11: File System Implementation
Introduction.
2018 Huawei H Real Questions Killtest
Chapter 11: File System Implementation
Chapter 11: File System Implementation
Hadoop Technopoints.
File Management – Implementation
HPC USER FORUM I/O PANEL April 2009 Roanoke, VA
Chapter 11: File System Implementation
Performance-Robust Parallel I/O
Factors Driving Enterprise NVMeTM Growth
Presentation transcript:

HPC USER FORUM I/O PANEL April 2009 Roanoke, VA Panel questions: 1 response per question Limit length to 1 slide

Panel Format

Panel Members

Q1. Parallel NFS finally is here! With the formalization of Parallel NFS as a standard - what steps are being provided to enable this to be hosted on current (and future) platform choices? This is really a development question for Garth and Peter Consumers should require this as part of the RFP process Remember that pNFS is a protocol and not a file system POSIX semantics for writing to the same file are not maintained Though pNFS is a very good thing is does not solve world peace and global warming The underlying file system might need to be modified to support the new high speed I/O

Q2. Parallel NFS – implementation details… What tools are available to help optimize this (from application layer level all the way to archival stage)? What is missing and who should provide it? Actually there are no tools for end-to-end integrated I/O optimization and it would require a major effort to provide an integrated tool kit There are many obstacles to these types of tools such as understanding the configuration and contention from other processes Example: A user could make a 30% performance improvement I/O of a job and it might not be recognized given contention from other work How do you measure the I/O of a job without measurement of the system

Q3. Physical media interconnects … We all are facing complexity and cost issues. With IB or 10 GbE (40/100 GbE) : where should the HPC community should focus its resources on - for all I/O? The community should be looking to the future at what will be the commodity technology will be the winner That in my opinion will be FCoE and GbE In my opinion sites should begin to plan for this future by planning for this by looking at NICs and cabling plans The process of change will being in earnest in 2010 and given legacy environments will likely take at least 5 years to complete Things take longer than expected with storage Take the 3.5 to 2.5 disk drive change now in progress for an example

Q4. Layer protocols above the interconnects Too many standards. interconnects, media layers are issues today. iSCSI/FCOE/FcoCEE/FCoIB have all been touted as the solution(s). Is it even relevant in the HPC arena? Is fragmentation the only choice? In my opinion ISCSI will not make it as an HPC solution The issue is encapsulation of metadata and small blocks and wasting channel bandwidth HPC is not all about streaming I/O Small block random happens more than you think and is the bounding factor for many HPC Commodity technology will eventually overtake IB was successful when PCIe became available And it will limit us again 40 GbE will win in cost over QDR Maybe not at first but eventually given volume

Q5. I/O issues not yet addressed? What do you consider to be the top 3 main (technical or human) issues in HPC I/O? 1 HPC file systems have historically had a problem opening a single file for all nodes, which means the file counts are growing significantly 2 There are no really I/O tools for users to tune applications on operational systems especially 3 User data management (ILM) frameworks are not part of the standard process and are implemented in user space Part of the complexity of this is a function of #1