CD central data storage and movement

Slides:



Advertisements
Similar presentations
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Advertisements

MCDST : Supporting Users and Troubleshooting a Microsoft Windows XP Operating System Chapter 8: Troubleshooting Storage Devices and Display Devices.
Exploring the UNIX File System and File Security
The Google File System. Why? Google has lots of data –Cannot fit in traditional file system –Spans hundreds (thousands) of servers connected to (tens.
Hands-On Microsoft Windows Server 2003 Administration Chapter 5 Administering File Resources.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
5.1 © 2004 Pearson Education, Inc. Exam Managing and Maintaining a Microsoft® Windows® Server 2003 Environment Lesson 5: Working with File Systems.
Hands-On Microsoft Windows Server 2003 Administration Chapter 6 Managing Printers, Publishing, Auditing, and Desk Resources.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
CD central data storage and movement. Facilities Central Mass Store Enstore Network connectivity.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
BACKUP/MASTER: Immediate Relief with Disk Backup Presented by W. Curtis Preston VP, Service Development GlassHouse Technologies, Inc.
Chapter 7 Configuring & Managing Distributed File System
CT NIKHEF June File server CT system support.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 12: Managing and Implementing Backups and Disaster Recovery.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
1 Microsoft Distributed File System (Dfs) Brett O’Neill CSE 8343 – Group A6.
Pooja Shetty Usha B Gowda.  Network File Systems (NFS)  Drawbacks of NFS  Parallel Virtual File Systems (PVFS)  PVFS components  PVFS application.
Microsoft Active Directory(AD) A presentation by Robert, Jasmine, Val and Scott IMT546 December 11, 2004.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
Building Advanced Storage Environment Cheng Yaodong Computing Center, IHEP December 2002.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Chapter Two Exploring the UNIX File System and File Security.
Lower Storage projects Alexander Moibenko 02/19/2003.
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
08/30/05GDM Project Presentation Lower Storage Summary of activity on 8/30/2005.
Week #3 Objectives Partition Disks in Windows® 7 Manage Disk Volumes Maintain Disks in Windows 7 Install and Configure Device Drivers.
ORBMeeting July 11, Outline SAM Overview and Station description Resource Management Station Cache Station Prioritized Fair Share Job Control File.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Chapter Two Exploring the UNIX File System and File Security.
 CASTORFS web page - CASTOR web site - FUSE web site -
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
| nectar.org.au NECTAR TRAINING Module 9 Backing up & Packing up.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
Eric Liu – Remote Proactive
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
Distributed File System. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
© 2012 IBM Corporation IBM Linear Tape File System (LTFS) Overview and Demo.
An Introduction to GPFS
Patrick Gartung 1 CMS 101 Mar 2007 Introduction to the User Analysis Facility (UAF) Patrick Gartung - Fermilab.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 7: Advanced File System Management.
Compute and Storage For the Farm at Jlab
CASTOR: possible evolution into the LHC era
DIT314 ~ Client Operating System & Administration
Integrating Disk into Backup for Faster Restores
Network Attached Storage Overview
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Quick Look on dCache Monitoring at FNAL
Introduction to Operating Systems
Introduction to Data Management in EGI
SAM at CCIN2P3 configuration issues
LQCD Computing Operations
Research Data Archive - technology
Storage Virtualization
Exploring the UNIX File System and File Security
Hadoop Technopoints.
CASTOR: CERN’s data management system
Lee Lueking D0RACE January 17, 2002
Introducing NTFS Reliability Security Long file names Efficiency
IBM Tivoli Storage Manager
Presentation transcript:

CD central data storage and movement

Facilities Central Mass Store Enstore Network connectivity

Central Mass Store Disk cache Tape library Server Software Network Client Software FNALU integration Exabyte Import and Export Policies

Hardware IBM 3494 Library 8 IBM 3590 tape drives 1 TB of staging disk internal to system Three IBM TBD mover node FDDI network, 10 MB/sec to outside world Servers

A cache Conceptually a cache, not a primary data repository. Implemented as a hierarchical store, with tape at the lowest level. The data are subject to loss should the tape fail. Quotas are refunded as tapes are squeezed. For “large files”

Allocation The CD Division office gives an allocation in terms of 10 GB volumes Experiments are to use system

Interface

Enstore

Service Envisioned Primary data store for experiments large data sets. Stage files to/from tape via LAN High fault tolerance - ensemble reliability of a large tape drive plant, availability sufficient for DAQ. Allow for automated tape libraries and manual tapes. Put names of files in distributed catalog (name space). CD will operate all the tape equipment Do not hide too much that it is really tape. Easy administration and monitoring. Work with commodity and “data center” tape drives.

Hardware for Early Use 1 each - STK 9310 “powderhorn” silo 5 each - STK 9840 “eagle” tape drives 10 MB/second used at BaBar, Cern, Rhic 1500 - STK 9840 tape cartridges 20 GB/ cartridge LINUX Server and Mover computers FNAL standard network

Service for First Users Software in production (4 TB) for D0 Run II AML/2 tape library: 8MM, DLT drives. STK system: Only working days, working hours. Small data volumes ~1 TB for trial use. Willing to upgrade lan, network interfaces. Willing to point out bugs and problems. New hardware => small chance of data loss.

Vision of ease of use Experiment can access tape as easily as a native file system. Namespace viewable with UNIX commands Transfer mechanism is similar to the unix cp command Syntax: encp infile outfile encp myfile.dat /pnfs/theory/project1/myfile.dat encp * /pnfs/theory/project1/ encp /pnfs/theory/project1/myfile.dat myfile.dat

Basic Structure “PNFS” to name tape files using UNIX like paths, served with NFS 2 transport Servers to schedule, configure, manage. Movers to bridge between network and tape drives

Software for Experiments (Clients) Use the Unix Mount command to view the PNFS namespace. Obtain the “encp” product from kits “encp command” miscellaneous “enstore <command>” enstore file --restore enstore volume [--add | --delete | --restore] enstore library [--delete_work --get_queue -priority]

Volume Principles Do Support clustering related files on the same tapes. Enstore provides grouping primitives. Do not assume we can buy a tape robot slot for every tape. Enstore provides quota in tapes and quotas in “slots” Experiment may have more tapes than slots Allow users to generate tapes outside our system Enstore provides tools to do this. Allow tapes to leave our system and be readable with simple tools Enstore can make tapes dumpable with cpio

Grouping on tapes Grouping by Category Grouping by time “File families” Only files of the same family are on the same tape. A family is just an ascii name names are administered by the experiment. Grouping by time Enstore closes volume for write when the next file does not fit. Constrained parallelism “width” associated with a “file family” limits the number of volumes open for writing, concentrates files on fewer volumes. Allows bandwidth into a file family to exceed the bandwidth of a tape drive.

File family, width=1 over time

Tape Details In production, implementation details are hidden. Files do not stripe or span volumes. Implementation details: Tapes have ANSI VOL1 headers. Tapes are file structured as CPIO archives. one file to an archive, one filemark per archive. You can remove tapes from Enstore and just read them with GNU CPIO (gives a 4GB limit right now). ANSI tapes planned, promised for D0.

Enstore “Libraries” A set of tapes which are uniform with respect to media characteristics low level treatment by the drive One mechanism to mount/unmount tapes An Enstore system can consist of many “Libraries’ : D0 (ait, mam-1, dlt. Mam-2. Ait-2) An Enstore system may have diverse robots (STKEN has STK 9310, and ADIC AML/J)

Namespace: functions Provide a tree to name files as you wish. Provide a tree named as “volume map” /pnfs/<mountpoint>/<ffname>/<volume>/<P-_B_fm> Provide information on how new files should be created, which the experiment can administer. Provide additional information about each file.

Namespace:UNIX features Implemented using PNFS from DESY. NFS v2 “Transport” “Almost all” UNIX Utilities work, ls, find Standard utility reads/writes fail by design Many files in a directory is a poor choice “by design” $ pwd /pnfs/sam/mammoth/mcc99_2/in2p3 [$ du -sk 267171544 . $ ls -al sim.pmc02_in2p3.pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138 -rw-r--r-- 1 sam root 250184748 Nov 30 17:25 sim.pmc02_in2p3.pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138 rm sim.pmc02_in2p3. pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138 rm: sim.pmc02_in2p3 .pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138: Permission denied $ cat sim.pmc02_in2p3.pythia.qcd_pt20.0skip5800_mb1.1av_200evts.299_1138 > /dev/null cat: sim.pmc02_in2p3.pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138: Input/output error

Namespace:defaults for new files Metadata “tags” are associated with directories Accessed by the “enstore pnfs” command. Inherited on “mkdir”. Initial tag on initial directory given by ISD dept Administered by the experiment [petravic@d0enmvr17a in2p3]$ enstore pnfs --tags . .(tag)(library) = sammam .(tag)(file_family) = in2p3 .(tag)(file_family_width) = 1 .(tag)(file_family_wrapper) = cpio_odc

Namespace: File Metadata Describes an existing file Accessed by the “enstore pnfs” command Set by encp when the file is created $ enstore pnfs --info sim.pmc02_in2p3.pythia.qcd_pt20.0_10000evts_skip5800_mb1.1av_200evts.299_1138 bfid="94400431100000L"; volume="PRF020"; location_cookie="0000_000005442_0000004"; size="250184748L"; file_family="in2p3"; map_file="/pnfs/sam/mammoth/volmap/in2p3/PRF020/0000_000005442_0000004";

Some encp command options --crc : data integrity --data_access_layer : structured error msgs --ephemeral : make a tape for export --file_family : override default ff --priority : get first claim on resources --del_pri : get a greater claim if waiting --verbose : be chatty

Removing Files Files may be removed using “rm”. User can scratch tape when all files on it are rm’ed. [enstore volume --delete] User can use a recovery utility to restore files up until the time the volume is scratched. [enstore file --restore] Files are recovered to pathname they were created with.

Sharing the Central Enstore System We make mount point(s) for your experiment Host-based authentication on the server side for mounts. Your meta data is in its own database files. Under the mount point, UNIX file permission apply. Make your uids/gid’s uniform! (FNAL uniform UID/GIDS). file permissions apply to the tag files as well. “Fair Share” envisioned, for tape drive resources. Control over experiment resources by the experiment Priorities implemented for Data Acquisition. Quick use of resources for the most urgent need

System Integration Hardware/system: Software configuration Consideration of upstream network. Consideration of your NIC cards. Good scheduling of the staging program. Good throughput to your file systems. Software configuration Software built for FUE platforms Linux, IRIX, SunOS, OSF1

Elements of Good Throughput

http://stkensrv2.fnal.gov/enstore/ Source of interesting monitoring info Most updates are batched. Can see recent transfers the system is up or down what transfers are queued more

http://stkensrv2/enstore/

System Status: Green == Good

History of Recent Transfers

Status Plots

Checklist to use Enstore Be authorized by the computing division. Identify performant disks and computers. Use “bonnie” and “streams” Provide suitable network connectivity. Use “enstore monitor” to measure. Plan use of namespace, file families. Regularize UIDs and GIDs if required. Mount the namespace. Use encp to access your files.