OffLine Physics Computing

Slides:



Advertisements
Similar presentations
Introduction CSCI 444/544 Operating Systems Fall 2008.
Advertisements

Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Smart Storage and Linux An EMC Perspective Ric Wheeler
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
Lecture 1: History of Operating System
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
An Introduction to Cloud Computing. The challenge Add new services for your users quickly and cost effectively.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Comp 335 – File Structures Why File Structures?. Goal of the Class To develop an understanding of the file I/O process. Software must be able to interact.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
February 02Page 1 Operating Systems Introduction.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
HepNT - January 15, 1997 : PCSF Frederic Hemmer IT/PDP 1 PCSF - A Pentium ® /Windows NT ® Based simulation farm Frederic Hemmer CERN IT/PDP.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
A.Abhari CPS1251 Topic 1: Introduction to Computers Computer Hardware Computer components Connecting Computers Computer Software Operating System (OS)
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
F. HemmerUltraNet® Experiences SHIFT Model CPU Server CPU Server CPU Server CPU Server CPU Server CPU Server Disk Server Disk Server Tape Server Tape Server.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Background Computer System Architectures Computer System Software.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Virtual Server Server Self Service Center (S3C) JI July.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
15.June 2004Bernd Panzer-Steindel, CERN/IT1 CERN Mass Storage Issues.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
1 OPERATING SYSTEMS. 2 CONTENTS 1.What is an Operating System? 2.OS Functions 3.OS Services 4.Structure of OS 5.Evolution of OS.
CASTOR: possible evolution into the LHC era
Getting the Most out of Scientific Computing Resources
Introduction to Operating Systems
Getting the Most out of Scientific Computing Resources
PROOF – Parallel ROOT Facility
Introduction to Data Management in EGI
Bernd Panzer-Steindel, CERN/IT
Luca dell’Agnello INFN-CNAF
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Chapter 21: Virtualization Technology and Security
Storage Virtualization
Chapter 1: Introduction
Virtual Memory Main memory can act as a cache for the secondary storage (disk) Advantages: illusion of having more physical memory program relocation protection.
Introduction to Operating Systems
Grid Canada Testbed using HEP applications
Chapter 22: Virtualization Security
What is Parallel and Distributed computing?
CASTOR: CERN’s data management system
What is an Operating System?
CSE 451: Operating Systems Autumn 2003 Lecture 12 Secondary Storage
Lecture Topics: 11/1 General Operating System Concepts Processes
Cloud Computing Architecture
CSE 451: Operating Systems Winter 2003 Lecture 12 Secondary Storage
CS246: Search-Engine Scale
CSE 451: Operating Systems Autumn 2004 Secondary Storage
Presentation transcript:

OffLine Physics Computing Simulation : CPU intensive, very low I/O, very long runs Reconstruction : CPU intensive, moderate I/O, long runs Data analysis : CPU intensive, High I/O, long runs Interactive Analysis : CPU intensive, High I/O, quick response, short runs @ 10 mins intervals Tape storage : 180000 3480 tapes in vault Staging model Traditionnally done on Mainframes (CRAY, IBM's) Simulation Recontruction Batch Data analysis Interative analysis

Applications Characteristics High data volumes : 180000 tapes in vault Staging model for Data analysis (i.e, xfer tape to disk, process it, xfer it out to tape) Standard I/O package (Zebra, EPIO) Little sharing, sometime for stage input Yes, FORTRAN ! High data volumes (disks, tapes, tape robots) Standard I/O packages Little file sharing Fortran

SHIFT Motivations Exploit cheap RISC technology spin-off RISC technology : 1988, Apollo DN1000 DN10000 : Mainframe CPU capacity 4-cpu, 1000 $/MIPS, 1 OS DN 10000: Faster than a mini, comparable to a mainframe SCSI disk today : > 2GB/disk, 1.8 MB/sec, < 4 K$ 3480 SCSI (STK): 2 MB/sec, 12 K$ HEP computing : always bigger needs in both data volumes CPU needs I/O bandwith SHIFT == cost effective computing SHIFT Motivations Exploit cheap RISC technology spin-off Exploit cheap SCSI technology Exploit HOPE success Answer to HEP computing always growing needs

SHIFT Goals Mainframe quality Scalable (down & up) Mianframe quality: stable, resilient Scale down (small experiments, external institutes - e.g. China) Scale up (higher demand, e.g. 40 GB at design, 250 now) Heterogeneous because of economics Integrated: present 1 single view to applications, all disk & tapes available everywhere System dev ; portable to minimize efforts, adaptable to new OS, devices, HW (i.e. Unix based) Mainframe quality Scalable (down & up) Heterogeneous (open) Integrated Minimal (portable) SW development

SHIFT Model Split in functional blocks Interconnect with a High speed backplane Simulation showed that for a 100 CERN Unit system : for 1 CU required 20 KB/s the backplane needed 18 MB/s + 3 MB/s Tape I/O per I/F > FDDI (at that time) & would require a lot of CPU time for protocol processing ==> UltraNet (1 Gb/s, 3-12 MB/s perf I/F) + Software !! CPU Server CPU Server CPU Server CPU Server CPU Server CPU Server Disk Server Disk Server High Speed Interconnect Tape Server Tape Server Tape Server

SHIFT Software Pseudo Distributed File System Tape Access Batch queues Pseudo distributed to present one single vies Tape access at the tape file level assuming a tape staging model Batch queues system wide available (actually load balanced) Pseudo Distributed File System Tape Access Batch queues

SHIFT Software Disk Pool Manager Remote Tape Copy DPM to organize file in pools of file systems (local & remote) RTCOPY = generalized remote tape staging NQS : load balancing RFIO: exploit netowrk I/O packages: Zebra, EPIO, Fatmen Developed in 4 month Currently 60000 lines of code Publicly available Disk Pool Manager Remote Tape Copy Clustered Batch Queues Remote File I/O Integration with I/O packages Unix Tape Control System Operator Interface Monitoring

CORE This is the environment CSF is the low I/O high CPU facility PIAF is being integrated Link to PARC Link to CHEOPS Link to private Simul farms (Opal) Intentions to expand to all LEP experiments and prepare LHC

CPU usage CORE is the main CPU provider IBM is still a big one (but $/mips) Cray is outragously expensive) IBM is still GP, CORE, SHIFT not Still have a lot to learn from mainframe world : Resilience (Disk failure rates) Maintainability (checkpoint/restart) Performance tuning Failure analysis Week 1-42 1992

SHIFT Stage statistics But after all, we do stage data, and not a little bit Tape drive usage is not bad Week 1-43 1992 Gigabytes SUN/STK stage data/drive/week : 48 GB

Tape mount statistics 10 % of all tape mounts ... but with 2 drives Week 1-43 1992 10 % of all tape mounts ... but with 2 drives We are increasing number of drives which are limiting data annalysis (jobs are waiting idle for tape mounts Tape mount/drive as good as mainframe IBM Mounts/Drive/Week 215 SUN/STK Mounts/Drive/Week 253

Conclusions RISC + SCSI cost effective CSF style easy to solve not tight to 1 manufacturer for CPU Network is not important apart speed, only the abstraction matters (sockets) Simulations : evry major HEP has such a project I/O: very few people have tried this, it caused us gey hairs Administration: machine number increase , config example, monitoring Scalable up (e.g. 40 GB to 250, DP) and down (Aleph, IHEP, SMC) Ready for LHC challenges (one order of magnitude) Only starting with // interactive data analysis Conclusions RISC + SCSI cost effective CSF style easy to solve (Remote) I/O is the major problem Administration is complex Really scalable