Systems in AMS02 AMS July 2003 Computing and Ground MIT Alexei Klimentov —

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
AMS TIM, CERN Jul 23, 2004 AMS Computing and Ground Centers Alexei Klimentov —
AMS COMPUTING VIII International Workshop Advanced Computing And Analysis Techniques in Physics Research Moscow, June 24-29, 2002 Vitali Choutko, Alexei.
 Changes to sources of funding for computing in the UK.  Past and present computing resources.  Future plans for computing developments. UK Status &
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
Building Large Scale Fabrics – A Summary Marcel Kunze, FZK.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
1999 Summer Student Lectures Computing at CERN Lecture 2 — Looking at Data Tony Cass —
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Hardware and Software Basics. Computer Hardware  Central Processing Unit - also called “The Chip”, a CPU, a processor, or a microprocessor  Memory (RAM)
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
AMS Computing Y2001-Y2002 AMS Technical Interchange Meeting MIT Jan 22-25, 2002 Vitali Choutko, Alexei Klimentov.
CERN 14/01/20021 Data Handling Scheme for the Italian Ground Segment (IGS), as part of AMS-02 Ground Segment (P.G. Rancoita) Functions of a “Regional Center”
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
AMS TIM, CERN Apr 12, 2005 AMS Computing and Ground Centers Status Report Alexei Klimentov —
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
OO Software and Data Handling in AMS Computing in High Energy and Nuclear Physics Beijing, September 3-7, 2001 Vitali Choutko, Alexei Klimentov MIT, ETHZ.
Alain Romeyer - 15/06/20041 CMS farm Mons Final goal : included in the GRID CMS framework To be involved in the CMS data processing scheme.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
AMS-02 Computing and Ground Data Handling CHEP 2004 September 29, Interlaken Alexei Klimentov — ETH Zurich and
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Status of AMS Regional Data Center The Second International Workshop on HEP Data Grid CHEP, KNU G. N. Kim, J. W. Shin, N. Tasneem, M. W.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
CHEP as an AMS Remote Data Center International HEP DataGrid Workshop CHEP, KNU G.N. Kim, H.B. Park, K.H. Cho, S. Ro, Y.D. Oh, D. Son (Kyungpook)
CHEP as an AMS Regional Center The Third International Workshop on HEP Data Grid CHEP, KNU G. N. Kim, J. W. Shin, N. Tasneem, M. W. Lee,
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
Sep 02 IPP Canada Remote Computing Plans Pekka K. Sinervo Department of Physics University of Toronto 4 Sep IPP Overview 2 Local Computing 3 Network.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
National HEP Data Grid Project in Korea Kihyeon Cho Center for High Energy Physics (CHEP) Kyungpook National University CDF CAF & Grid Meeting July 12,
RAL Site report John Gordon ITD October 1999
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
Introduction to the new mainframe © Copyright IBM Corp., All rights reserved. 1 Main Frame Computing Objectives Explain why data resides on mainframe.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Scientific Computing Facilities for CMS Simulation Shams Shahid Ayub CTC-CERN Computer Lab.
POCC activities AMS02 meeting at KSC October 11 th 2010.
January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 1 SIMULATION OF DAILY ACTIVITITIES AT REGIONAL CENTERS MONARC Collaboration Alexander Nazarenko.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, R. Brock,T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina,
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
AMS02 Data Volume, Staging and Archiving Issues AMS Computing Meeting CERN April 8, 2002 Alexei Klimentov.
1 AMS-02 POCC & SOC MSFC, 9-Apr-2009 Mike Capell Avionics & Operations Lead Senior Research Scientist ISS: 108x80m 420T 86KW 400km AMS: 5x4x3m 7T 2.3+KW.
AMS02 Software and Hardware Evaluation A.Eline. Outline  AMS SOC  AMS POC  AMS Gateway Computer  AMS Servers  AMS ProductionNodes  AMS Backup Solution.
E-USOC operations in Increments José Miguel Ezquerro Navarro E-USOC/UPM Jul 2010POIWG#28 - Huntsville, AL.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
WP18, High-speed data recording Krzysztof Wrona, European XFEL
PC Farms & Central Data Recording
SAM at CCIN2P3 configuration issues
Presentation transcript:

Systems in AMS02 AMS July 2003 Computing and Ground MIT Alexei Klimentov —

2Alexei Klimentov. AMS TIM. July Outline  AMS02 Data Flow  AMS02 Ground Centers  Science Operation Center Architecture choice of HW, cost estimation, implementation plan  Data Transmission SW  TReK SW

3Alexei Klimentov. AMS TIM. July AMS Science Operations Center ISS to Remote AMS Centers Data Flow Buffered data ACOP High Rate Frame MUX AMS Payload Operations Control Center POIC AMS GSC Monitoring & Science data Real-Time & “Dump” data Stored data RealTime,“Dump” & WhiteSands LOR playback External Communications Short-Term Long-Term Payload Data Service System Real-Time H&S data Marshall Space Flight Center, AL Real-Time & “Dump” data NearReal-Time &“Dump” data FileTransfer playback Payload Operation & Integration Center Payload Operation & Integration Center NASA’s Ground InfrastructureISS,,, AMS Regional Centers FileTransfer White Sand, NM Facility Event reconstruction, batch & interactive Physics analysis Data archiving Event reconstruction, batch & interactive Physics analysis Data archiving Commanding, Monitoring, Online analysis Commanding, Monitoring, Online analysis Buffering before transmission Buffering before transmission

4Alexei Klimentov. AMS TIM. July AMS Ground Centers (Ground Support Computers)  At Marshall Space Flight Center (MSFC), Huntsville Al  Receives monitoring and science data from NASA Payload Operation and Integration Center (POIC)  Buffers data until retransmission to the AMS Science Operation Center (SOC) and if necessary to AMS Payload Operations and Control Center (POCC)  Runs unattended 24h/day, 7 days/week  Must buffer about 600 GB (data for 2 weeks)

5Alexei Klimentov. AMS TIM. July AMS Ground Centers (Payload Operation and Control Center )  AMS02 “Counting Room”  Usual source of AMS commands  Receives H&S, monitoring, science and NASA data in real-time mode  Monitor the detector state and performance  Process about 10% of data in near real time mode to provide fast information to the shift taker  Video distribution “box”  Voice loops with NASA  Computing Facilities Primary and backup commanding stations Detector and subdetectors monitoring stations Stations for event display and subdetectors status displays Linux servers for online data processing and validation

6Alexei Klimentov. AMS TIM. July AMS Ground Centers ( Science Operation Center)  Receives the complete copy of ALL data  Data reconstruction, calibration, alignment and processing, generates event summary data and does event classification  Science analysis  Archive and record ALL raw, reconstructed and H&S data  Data distribution to AMS Universities and Laboratories

7Alexei Klimentov. AMS TIM. July (Regional Centers)  Analysis facility to support physicists from geographically closed AMS Universities and Laboratories;  Monte-Carlo Production;  Provide access to SOC data storage (event visualisation, detector and data production status, samples of data, video distribution);  Mirroring AMS DST/ESD. AMS Ground Centers

8Alexei Klimentov. AMS TIM. July AMS Data Volume (Tbytes) Data/ Year Total Raw ~64 ESD ~146 Tags Total ~212 MC ~206 Grand Total ~420 STS91 ISS

9Alexei Klimentov. AMS TIM. July Symmetric MultiProcessor Model Experiment Tape Storage TeraBytes of disks

10Alexei Klimentov. AMS TIM. July Scalable model Disk & Tape Storage TeraBytes of disks

11Alexei Klimentov. AMS TIM. July AMS02 Benchmarks Executive time of AMS “standard” job compare to CPU clock 1) V.Choutko, A.Klimentov AMS note ) Brand, CPU, Memory Intel PII dual-CPU 450 MHz, 512 MB RAM OS/Compiler RH Linux 6.2 / gcc 2.95 “Sim” 1 “Rec” 1 Intel PIII dual-CPU 933 MHz, 512 MB RAMRH Linux 6.2 / gcc Compaq, Quad α-ev MHz, 2 GB RAMRH Linux 6.2 / gcc AMD Athlon,1.2GHz, 256 MB RAMRH Linux 6.2 / gcc Intel Pentium IV 1.5GHz, 256 MB RAMRH Linux 6.2 / gcc Compaq dual-CPU PIV Xeon 1.7GHz, 2GB RAMRH Linux 6.2 / gcc Compaq dual α-ev68 866MHz, 2GB RAMTru64 Unix/ cxx Elonex Intel dual-CPU PIV Xeon 2GHz, 1GB RAMRH Linux 7.2 / gcc AMD Athlon 1800MP, dual-CPU 1.53GHz, 1GB RAMRH Linux 7.2 / gcc CPU SUN-Fire-880, 750MHz, 8GB RAMSolaris 5.8/C CPU Sun Ultrasparc-III+, 900MHz, 96GB RAMRH Linux 6.2 / gcc Compaq α-ev68 dual 866MHz, 2GB RAMRH Linux 7.1 / gcc

12Alexei Klimentov. AMS TIM. July AMS SOC (Data Production requirements)  Reliability – High (24h/day, 7days/week)  Performance goal – process data “quasi-online” (with typical delay < 1 day)  Disk Space – 12 months data “online”  Minimal human intervention (automatic data handling, job control and book-keeping)  System stability – months  Scalability  Price/Performance Complex system that consists of computing components including I/O nodes, worker nodes, data storage and networking switches. It should perform as a single system. Requirements :

13Alexei Klimentov. AMS TIM. July AMS Science Center Computing Facilities CERN/AMS Network AMS Physics Services N Central Data Services Shared Disk Servers 25 TeraByte disk 6 PC based servers 25 TeraByte disk 6 PC based servers tape robots tape drives LTO, DLT tape robots tape drives LTO, DLT Shared Tape Servers Home directories & registry consoles & monitors Production Facilities, Linux dual-CPU computers Linux, Intel and AMD Engineering Cluster 5 dual processor PCs Data Servers, Analysis Facilities (linux cluster) dual processor PCs 5 PC servers AMS regional Centers batch data processing batch data processing interactive physics analysis Interactive and Batch physics analysis

14Alexei Klimentov. AMS TIM. July AMS Computing facilities (disks and cpus projected characteristics) Components Intel/AMD PC Dual-CPU Intel PII, rated at 450 MHz, 512 MB RAM. 7.5 kUS$ Dual-CPU Intel, Rated at 2.2 GHz, 1GB RAM and RAID controller 7 kUS$ Dual-CPU rated at 8GHz, 2GB RAM and RAID controller 7 kUS$ Magnetic disk 18 GByte SCSI 80 US$/Gbyte SG 180 GByte SCSI 10 US$/Gbyte WD 200 GByte IDE 2 US$/Gbyte 700 Gbyte 1 US$/Gbyte Magnetic tape DLT 40 GB 3 US$/Gbyte SDLT and LTO 200 GB 0.8 US$/Gbyte ? 400 GB 0.3 US$/Gbyte

15Alexei Klimentov. AMS TIM. July AMS02 Computing Facilities Y (cost estimate) FunctionComputerQtyDisks (Tbytes) Cost kUS$ Sun, Intel, dual-CPU, 1.5+GHz2 2x1TB Raid-Array 55 POCCx2 Intel and AMD, dual-CPU, 2.4+GHz201TB Raid-Array150 Production FarmIntel and AMD, dual-CPU, 2.4+GHz5010 TB Raid-Array350 Database Serversdual-CPU 2.0+ GHz Intel or Sun SMP20.5TB50 Event Storage and Archiving DiskServers dual-CPU Intel 2.0+Ghz625 Tbyte RaidArray200 Interactive and Batch Analysis SMP computer, 4GB RAM, 300 Specint95 or Linux farm 2/101 Tbyte Raid Array55 Sub. Total860 Running Cost 150 Grand Total 1010

16Alexei Klimentov. AMS TIM. July AMS Computing facilities (implementation plan) Q End 2003 Choice of server and processing node architecture, setup 10% prototype of AMS production farm. Evaluation of archiving system 40% prototype of AMS production farm Q End 2004 Evaluation SMP vs distribudet computing, finalize the architecture of 60% prototype of AMS production farm, purchase and setup final configuration make choice of “analysis” computer, archiving and storage system Q Q AMS GSC prototype Al, data transmission tests tests between MSFC and CERN, MSFC and MIT Disk server and processor architecture evaluation Beg 2005 Mid 2005 End 2005 Purchase disks to setup dsik pool, purchase POCC computers. Purchase “analysis computer”, setup production farm in final configuration Setup final configuration of production farm and analysis computer

17Alexei Klimentov. AMS TIM. July CERN’s Network Connections CERN RENATER C-IXP IN2P3 TEN-155 KPNQwest (US) SWITCH 39/155 Mb/s 155Mb/s 2Mb/s 1Gb/s 2x255Mb/s 1Gb/s National Research Networks Mission Oriented Link Public Commercial WHO TEN-155: Trans- European Network at 155Mb/s 45Mb/s

18Alexei Klimentov. AMS TIM. July CERN’s Network Traffic CERN 40 Mb/s Out 38 Mb/s In KPNQwest (US) RENATERTEN-155IN2P3SWITCH 100Mb/s 2Mb/s 2x255Mb/s 40Mb/s 45Mb/s Link Bandwidth 5.2Mb/s 20Mb/s 4.7Mb/s 14Mb/s 0.1Mb/s 5.5Mb/s 25Mb/s 0.1Mb/s CERN : ~36 TB/month in/out AMS Raw data 0.66 TB/month = 2 Mb/s 1Mb/s = 11GB/day Incoming data rate Outgoing data rate

19Alexei Klimentov. AMS TIM. July Data Transmission  Will AMS need a dedicated line to send data from MSFC to ground centers or the public Internet can be used ?  What Software (SW) must be used for a bulk data transfer and how reliable is it ?  What data transfer performance can be achieved ? High Rate Data Transfer between MSFC Al and POCC/SOC, POCC and SOC, SOC and Regional centers will become a paramount importance

20Alexei Klimentov. AMS TIM. July Data Transmission SW Why not FileTransferProtocol (ftp) or ncftp, etc ? to speed up data transfer to encrypt sensitive data and not encrypt bulk data to run in batch mode with automatic retry in case of failure  … starting to look around and came up with bbftp in September 2001 (still looking for a good network monitoring tools) (bbftp developed in BaBar and used to transmit data from SLAC to adapted it for AMS, wrote service and control programs 1) 1) A.Elin, A.Klimentov AMS note ) P.Fisher, A.Klimentov AMS Note

21Alexei Klimentov. AMS TIM. July Data Transmission SW (tests) sourcedestination Test Duration (hours) Nominal Bandwidth (Mbit/sec) Iperf (Mbit/sec) Bbftp (Mbit/sec) CERN ICERN II CERN ICERN II CERN IIMIT12x3100[255] CERN IIMSFC Al24x2100[255] MSFC AlCERN II24x2100[255]

22Alexei Klimentov. AMS TIM. July Data Transmission Tests (conclusions)  In its current configuration Internet provides sufficient bandwidth to transmit AMS data from MSFC Al to AMS ground centers at rate approaching 9.5 Mbit/sec  bbftp is able to transfer and store data on a high end PC reliably with no data loss  bbftp performance is comparable of what achieved with network monitoring tools  bbftp can be used to transmit data simultaneously to multiple cites