Fermi National Accelerator Laboratory SC2006 Fermilab Data Movement & Storage Multi-Petabyte tertiary automated tape store for world- wide HEP and other.

Slides:



Advertisements
Similar presentations
Jens G Jensen Atlas Petabyte store Supporting Multiple Interfaces to Mass Storage Providing Tape and Mass Storage to Diverse Scientific Communities.
Advertisements

Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Fermilab Mass Storage System Gene Oleynik Integrated Administration, Fermilab.
XenData SXL-5000 LTO Archive System Turnkey video archive system with near-line LTO capacities scaling from 210 TB to 1.18 PB, designed for the demanding.
XenData SX-520 LTO Archive Servers A series of archive servers based on IT standards, designed for the demanding requirements of the media and entertainment.
XenData SX-10 LTO Archive Appliance An Archive Appliance based on IT standards, designed for the demanding requirements of the media and entertainment.
XenData SXL-3000 LTO Archive System Turnkey video archive system with near-line LTO capacities scaling from 150 TB to 750 TB, designed for the demanding.
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Belle computing upgrade Ichiro Adachi 22 April 2005 Super B workshop in Hawaii.
5 Nov 2001CGW'01 CrossGrid Testbed Node at ACC CYFRONET AGH Andrzej Ozieblo, Krzysztof Gawel, Marek Pogoda 5 Nov 2001.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Lecture 1: Introduction CS170 Spring 2015 Chapter 1, the text book. T. Yang.
Operational Dataset Update Functionality Included in the NCAR Research Data Archive Management System 1 Zaihua Ji Doug Schuster Steven Worley Computational.
Designing Storage Architectures for Preservation Collections Library of Congress, September 17-18, 2007 Preservation and Access Repository Storage Architecture.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Memory. Random Access Memory Defined What is memory? operating system and other system software that control the usage of the computer equipment application.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
Terabyte IDE RAID-5 Disk Arrays David A. Sanders, Lucien M. Cremaldi, Vance Eschenburg, Romulus Godang, Christopher N. Lawrence, Chris Riley, and Donald.
Physical Storage and File Organization COMSATS INSTITUTE OF INFORMATION TECHNOLOGY, VEHARI.
BESIII computing 王贻芳. Peak Data volume/year Peak data rate at 3000 Hz Events/year: 1*10 10 Total data of BESIII is about 2*640 TB Event size(KB)Data volume(TB)
 Design model for a computer  Named after John von Neuman  Instructions that tell the computer what to do are stored in memory  Stored program Memory.
SXL-8 LTO Archive System. SXL-8 Components: HP 1/8 Autoloader XenData SX-10 1RU.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
Chapter 9 Section 2 : Storage Networking Technologies and Virtualization.
Course ILT Basics of information technology Unit objectives Define “information technology” (IT), distinguish between hardware and software, and identify.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Fermilab June 29, 2001Data collection and handling for HEP1 Matthias Kasemann Fermilab Overview of Data collection and handling for High Energy Physics.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
22nd March 2000HEPSYSMAN Oxford Particle Physics Site Report Pete Gronbech Systems Manager.
HPSS for Archival Storage Tom Sherwin Storage Group Leader, SDSC
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
Computer Organization. The Five Hardware Units General purpose computers use the "Von Neumann" architecture Also referred to as "stored program" architecture.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
Operation of the CERN Managed Storage environment; current status and future directions CHEP 2004 / Interlaken Data Services team: Vladimír Bahyl, Hugo.
Niko Neufeld, CERN. Trigger-free read-out – every bunch-crossing! 40 MHz of events to be acquired, built and processed in software 40 Tbit/s aggregated.
August 28, 2003APAN, Logistical Networking WS DiDaS Distributed Data Storage Ludek Matyska Masaryk University, Institute of Comp. Sci. and CESNET, z.s.p.o.
Tackling I/O Issues 1 David Race 16 March 2010.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
CERN IT Department CH-1211 Genève 23 Switzerland t The Tape Service at CERN Vladimír Bahyl IT-FIO-TSI June 2009.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
TEVATRON DATA CURATION UPDATE Gene Oleynik, Fermilab Department Head, Data Movement and Storage 1.
BEHIND THE SCENES LOOK AT INTEGRITY IN A PERMANENT STORAGE SYSTEM Gene Oleynik, Jon Bakken, Wayne Baisley, David Berg, Eileen Berman, Chih-Hao Huang, Terry.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Compute and Storage For the Farm at Jlab
XenData SX-10 LTO Archive Appliance
Video Security Design Workshop:
Thoughts on Computing Upgrade Activities
Chapter 7.
Distinguish between primary and secondary storage.
XenData SX-550 LTO Archive Servers
Primary Storage 1. Registers Part of the CPU
Presentation transcript:

Fermi National Accelerator Laboratory SC2006 Fermilab Data Movement & Storage Multi-Petabyte tertiary automated tape store for world- wide HEP and other scientific endeavors. Multi-Petabyte tertiary automated tape store for world- wide HEP and other scientific endeavors. High Availability (24x7) High Availability (24x7) Local and Grid access Local and Grid access Scalable Hardware and Software Architecture Scalable Hardware and Software Architecture Front-end disk caching Front-end disk caching Evolves to meet evolving requirements Evolves to meet evolving requirements 15 TB/day Peaks of > 25 TB to & from tape/day 4.5 PB on tape

Fermi National Accelerator Laboratory SC2006 The DZero Experiment The CDF Experiment And Many Others DES, KTeV, MINOS, LQCD, MiniBooNE, … The CMS Experiment Sloan Digital Sky Survey Local Sources Remote Sources

Fermi National Accelerator Laboratory SC2006 Users write RAW data to Mass Storage, analyze/reanalyze it in real-time on PC “farms”, then write results back into the Mass Storage Users write RAW data to Mass Storage, analyze/reanalyze it in real-time on PC “farms”, then write results back into the Mass Storage ~3 bytes read for every byte written to tape ~3 bytes read for every byte written to tape Lifetime of data exceeds 5 years.Lifetime of data exceeds 5 years.

Fermi National Accelerator Laboratory SC2006 Front-end disk caching Rate adapting, Fast access to frequently requested files + load balancing > 800 TB and growing in volatile & tape backed raid disk CMS 250 TB read from cache in one day 50 TB

Fermi National Accelerator Laboratory SC2006 Tape Backed Disk Cache CMS CSA06 1 from CERN to Fermilab Data backed from disk to tape (Pink) Disk Cache to tape ~ 250 MB/s Network to disk cache ~ 250 MB/s 1. Computing, Software, Analysis Challenge, October 2006

Fermi National Accelerator Laboratory SC2006 Future capacity through 2010 Currently ~ 4.5 PB on ~35000 tapes, accessed by 120 tape drives from 9 libraries Currently ~ 4.5 PB on ~35000 tapes, accessed by 120 tape drives from 9 libraries Currently more than 500TB in front-end disk cache Currently more than 500TB in front-end disk cache Expecting additional ~30 PB or more on tape by 2010 Expecting additional ~30 PB or more on tape by 2010 Will have > 1PB in disk cache serving US CMS community Will have > 1PB in disk cache serving US CMS community Will need to acquire a tape library and dozens of tape drives per year to accommodate Will need to acquire a tape library and dozens of tape drives per year to accommodate

Fermi National Accelerator Laboratory SC2006 Long-term Retention & Data Integrity capacity recycle NOACCESS Tape Transport problem “3 strikes” Selective CRC error NOACCESS Tape Transport problem “3 strikes” Selective CRC error Automated safeguards and audits detect problem tapes and drives (3 tries) and denies access to them Tickets generated automatically to physically write protect filled tapes Newly filled tapes needing protection Protected tapes Manage tape Life Cycle. “Clone” tapes with too many mounts, Write protect full tapes to prevent accidental erasure. Drop from tape aide write protecting tapes on a ticket Randomly select files, read them and check their integrity (calculated CRC)