CD central data storage and movement. Facilities Central Mass Store Enstore Network connectivity.

Slides:



Advertisements
Similar presentations
Fermilab Mass Storage System Gene Oleynik Integrated Administration, Fermilab.
Advertisements

CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
MCDST : Supporting Users and Troubleshooting a Microsoft Windows XP Operating System Chapter 8: Troubleshooting Storage Devices and Display Devices.
Exploring the UNIX File System and File Security
The Google File System. Why? Google has lots of data –Cannot fit in traditional file system –Spans hundreds (thousands) of servers connected to (tens.
Phones OFF Please Operating System Introduction Parminder Singh Kang Home:
Chapter 7: Configuring Disks. 2/24 Objectives Learn about disk and file system configuration in Vista Learn how to manage storage Learn about the additional.
Hands-On Microsoft Windows Server 2003 Administration Chapter 5 Administering File Resources.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
5.1 © 2004 Pearson Education, Inc. Exam Managing and Maintaining a Microsoft® Windows® Server 2003 Environment Lesson 5: Working with File Systems.
Hands-On Microsoft Windows Server 2003 Administration Chapter 6 Managing Printers, Publishing, Auditing, and Desk Resources.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
BACKUP/MASTER: Immediate Relief with Disk Backup Presented by W. Curtis Preston VP, Service Development GlassHouse Technologies, Inc.
Chapter 7 Configuring & Managing Distributed File System
CT NIKHEF June File server CT system support.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 12: Managing and Implementing Backups and Disaster Recovery.
1 Objectives Discuss the Windows Printer Model and how it is implemented in Windows Server 2008 Install the Print Services components of Windows Server.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
1 Microsoft Distributed File System (Dfs) Brett O’Neill CSE 8343 – Group A6.
Pooja Shetty Usha B Gowda.  Network File Systems (NFS)  Drawbacks of NFS  Parallel Virtual File Systems (PVFS)  PVFS components  PVFS application.
Module 9: Configuring Storage
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
Building Advanced Storage Environment Cheng Yaodong Computing Center, IHEP December 2002.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Chapter Two Exploring the UNIX File System and File Security.
Lower Storage projects Alexander Moibenko 02/19/2003.
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
Write-through Cache System Policies discussion and A introduction to the system.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
11 DISASTER RECOVERY Chapter 13. Chapter 13: DISASTER RECOVERY2 OVERVIEW  Back up server data using the Backup utility and the Ntbackup command  Restore.
08/30/05GDM Project Presentation Lower Storage Summary of activity on 8/30/2005.
Week #3 Objectives Partition Disks in Windows® 7 Manage Disk Volumes Maintain Disks in Windows 7 Install and Configure Device Drivers.
ORBMeeting July 11, Outline SAM Overview and Station description Resource Management Station Cache Station Prioritized Fair Share Job Control File.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
 CASTORFS web page - CASTOR web site - FUSE web site -
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
Introduction to HDFS Prasanth Kothuri, CERN 2 What’s HDFS HDFS is a distributed file system that is fault tolerant, scalable and extremely easy to expand.
Microsoft Windows XP Professional MCSE Exam
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
| nectar.org.au NECTAR TRAINING Module 9 Backing up & Packing up.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
© 2012 IBM Corporation IBM Linear Tape File System (LTFS) Overview and Demo.
An Introduction to GPFS
Patrick Gartung 1 CMS 101 Mar 2007 Introduction to the User Analysis Facility (UAF) Patrick Gartung - Fermilab.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Compute and Storage For the Farm at Jlab
DIT314 ~ Client Operating System & Administration
Integrating Disk into Backup for Faster Restores
SAM at CCIN2P3 configuration issues
UK GridPP Tier-1/A Centre at CLRC
LQCD Computing Operations
Research Data Archive - technology
Storage Virtualization
Exploring the UNIX File System and File Security
CD central data storage and movement
Hadoop Technopoints.
CASTOR: CERN’s data management system
Lee Lueking D0RACE January 17, 2002
Creating and Managing Folders
Presentation transcript:

CD central data storage and movement

Facilities Central Mass Store Enstore Network connectivity

Central Mass Store Disk cache Tape library Server Software Network Client Software FNALU integration Exabyte Import and Export Policies

Hardware IBM 3494 Library 8 IBM 3590 tape drives 1 TB of staging disk internal to system Three IBM TBD mover node FDDI network, 10 MB/sec to outside world Servers

A cache Conceptually a cache, not a primary data repository. Implemented as a hierarchical store, with tape at the lowest level. The data are subject to loss should the tape fail. Quotas are refunded as tapes are squeezed. For “large files”

Allocation The CD Division office gives an allocation in terms of 10 GB volumes Experiments are to use system

Interface

Enstore

Service Envisioned Primary data store for experiments large data sets. Stage files to/from tape via LAN High fault tolerance - ensemble reliability of a large tape drive plant, availability sufficient for DAQ. Allow for automated tape libraries and manual tapes. Put names of files in distributed catalog (name space). CD will operate all the tape equipment Do not hide too much that it is really tape. Easy administration and monitoring. Work with commodity and “data center” tape drives.

Hardware for Early Use 1 each - STK 9310 “powderhorn” silo 5 each - STK 9840 “eagle” tape drives –10 MB/second –used at BaBar, Cern, Rhic STK 9840 tape cartridges –20 GB/ cartridge LINUX Server and Mover computers FNAL standard network

Service for First Users Software in production (4 TB) for D0 Run II AML/2 tape library: 8MM, DLT drives. STK system: –Only working days, working hours. –Small data volumes ~1 TB for trial use. –Willing to upgrade lan, network interfaces. –Willing to point out bugs and problems. – New hardware => small chance of data loss.

Vision of ease of use Experiment can access tape as easily as a native file system. Namespace viewable with UNIX commands Transfer mechanism is similar to the unix cp command Syntax: encp infile outfile encp myfile.dat /pnfs/theory/project1/myfile.dat encp * /pnfs/theory/project1/ encp /pnfs/theory/project1/myfile.dat myfile.dat

Basic Structure “PNFS” to name tape files using UNIX like paths, served with NFS 2 transport Servers to schedule, configure, manage. Movers to bridge between network and tape drives

Software for Experiments (Clients) Use the Unix Mount command to view the PNFS namespace. Obtain the “encp” product from kits –“encp command” –miscellaneous “enstore ” enstore file --restore enstore volume [--add | --delete | --restore] enstore library [--delete_work --get_queue -priority]

Volume Principles Do Support clustering related files on the same tapes. –Enstore provides grouping primitives. Do not assume we can buy a tape robot slot for every tape. –Enstore provides quota in tapes and quotas in “slots” –Experiment may have more tapes than slots Allow users to generate tapes outside our system –Enstore provides tools to do this. Allow tapes to leave our system and be readable with simple tools –Enstore can make tapes dumpable with cpio

Grouping on tapes Grouping by Category –“File families” Only files of the same family are on the same tape. –A family is just an ascii name –names are administered by the experiment. Grouping by time –Enstore closes volume for write when the next file does not fit. Constrained parallelism –“width” associated with a “file family” limits the number of volumes open for writing, concentrates files on fewer volumes. –Allows bandwidth into a file family to exceed the bandwidth of a tape drive.

File family, width=1 over time

Tape Details In production, implementation details are hidden. Files do not stripe or span volumes. Implementation details: –Tapes have ANSI VOL1 headers. –Tapes are file structured as CPIO archives. one file to an archive, one filemark per archive. You can remove tapes from Enstore and just read them with GNU CPIO (gives a 4GB limit right now). ANSI tapes planned, promised for D0.

Enstore “Libraries” A set of tapes which are uniform with respect to –media characteristics –low level treatment by the drive One mechanism to mount/unmount tapes An Enstore system can consist of many “Libraries’ : D0 (ait, mam-1, dlt. Mam-2. Ait-2) An Enstore system may have diverse robots (STKEN has STK 9310, and ADIC AML/J)

Namespace: functions Provide a tree to name files as you wish. Provide a tree named as “volume map” –/pnfs/ / / / Provide information on how new files should be created, which the experiment can administer. Provide additional information about each file.

Namespace:UNIX features Implemented using PNFS from DESY. NFS v2 “Transport ” “Almost all” UNIX Utilities work, ls, find Standard utility reads/writes fail by design Many files in a directory is a poor choice “by design” $ pwd /pnfs/sam/mammoth/mcc99_2/in2p3 [$ du -sk $ ls -al sim.pmc02_in2p3.pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138 -rw-r--r-- 1 sam root Nov 30 17:25 sim.pmc02_in2p3.pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138 rm sim.pmc02_in2p3. pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138 rm: sim.pmc02_in2p3.pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138: Permission denied $ cat sim.pmc02_in2p3.pythia.qcd_pt20.0skip5800_mb1.1av_200evts.299_1138 > /dev/null cat: sim.pmc02_in2p3.pythia.qcd_pt20.0_skip5800_mb1.1av_200evts.299_1138: Input/output error

Namespace:defaults for new files Metadata “tags” are associated with directories Accessed by the “enstore pnfs” command. Inherited on “mkdir”. Initial tag on initial directory given by ISD dept Administered by the experiment in2p3]$ enstore pnfs --tags..(tag)(library) = sammam.(tag)(file_family) = in2p3.(tag)(file_family_width) = 1.(tag)(file_family_wrapper) = cpio_odc

Namespace: File Metadata Describes an existing file Accessed by the “enstore pnfs” command Set by encp when the file is created $ enstore pnfs --info sim.pmc02_in2p3.pythia.qcd_pt20.0_10000evts_skip5800_mb1.1av_200evts.299_1138 bfid=" L"; volume="PRF020"; location_cookie="0000_ _ "; size=" L"; file_family="in2p3"; map_file="/pnfs/sam/mammoth/volmap/in2p3/PRF020/0000_ _ ";

Some encp command options --crc : data integrity --data_access_layer: structured error msgs --ephemeral : make a tape for export --file_family : override default ff --priority : get first claim on resources --del_pri : get a greater claim if waiting --verbose : be chatty

Removing Files Files may be removed using “rm”. User can scratch tape when all files on it are rm’ed. [enstore volume --delete] User can use a recovery utility to restore files up until the time the volume is scratched. [enstore file --restore] Files are recovered to pathname they were created with.

Sharing the Central Enstore System We make mount point(s) for your experiment –Host-based authentication on the server side for mounts. –Your meta data is in its own database files. –Under the mount point, UNIX file permission apply. –Make your uids/gid’s uniform! (FNAL uniform UID/GIDS). –file permissions apply to the tag files as well. “Fair Share” envisioned, for tape drive resources. –Control over experiment resources by the experiment Priorities implemented for Data Acquisition. –Quick use of resources for the most urgent need

System Integration Hardware/system: –Consideration of upstream network. –Consideration of your NIC cards. –Good scheduling of the staging program. –Good throughput to your file systems. Software configuration –Software built for FUE platforms Linux, IRIX, SunOS, OSF1

Elements of Good Throughput

Source of interesting monitoring info Most updates are batched. Can see –recent transfers –the system is up or down –what transfers are queued –more

System Status: Green == Good

History of Recent Transfers

Status Plots

Checklist to use Enstore Be authorized by the computing division. Identify performant disks and computers. –Use “bonnie” and “streams” Provide suitable network connectivity. –Use “enstore monitor” to measure. Plan use of namespace, file families. Regularize UIDs and GIDs if required. Mount the namespace. Use encp to access your files.