Farm Completion Beat Jost and Niko Neufeld LHCb Week St. Petersburg June 2010.

Slides:



Advertisements
Similar presentations
PowerEdge T20 Customer Presentation. Product overview Customer benefits Use cases Summary PowerEdge T20 Overview 2 PowerEdge T20 mini tower server.
Advertisements

PowerEdge T20 Channel NDA presentation Dell Confidential – NDA Required.
Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.
CIT 470: Advanced Network and System AdministrationSlide #1 CIT 470: Advanced Network and System Administration Servers.
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
V IRTUALIZATION. V IRTUALIZATION I S E VERYWHERE Developers use it to test software on different operating systems IT professionals use it to test different.
The LHCb Event-Builder Markus Frank, Jean-Christophe Garnier, Clara Gaspar, Richard Jacobson, Beat Jost, Guoming Liu, Niko Neufeld, CERN/PH 17 th Real-Time.
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
5 th LHCb Computing Workshop, May 19 th 2015 Niko Neufeld, CERN/PH-Department
Performance benchmark of LHCb code on state-of-the-art x86 architectures Daniel Hugo Campora Perez, Niko Neufled, Rainer Schwemmer CHEP Okinawa.
LHCb readout infrastructure NA62 TDAQ WG Meeting April 1 st, 2009 Niko Neufeld, PH/LBC.
Project Cysera Hardware Configuration Drafted by Zoebir Bong.
HS06 on the last generation of CPU for HEP server farm Michele Michelotto 1.
Niko Neufeld CERN PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
5.3 HS23 Blade Server. The HS23 blade server is a dual CPU socket blade running Intel´s new Xeon® processor, the E5-2600, and is the first IBM BladeCenter.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Computing/Tier 3 Status at Panjab S. Gautam, V. Bhatnagar India-CMS Meeting, Sept 27-28, 2007 Delhi University, Delhi Centre of Advanced Study in Physics,
Bob Thome, Senior Director of Product Management, Oracle SIMPLIFYING YOUR HIGH AVAILABILITY DATABASE.
Microsoft Virtual Academy. Microsoft Virtual Academy.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
Objective  CEO of a small company  Create a small office network  $10,000 and $20,000 Budget  Three servers (workstations)  Firewall device  Switch.
© 2012 IBM Corporation IBM Flex System™ The elements of an IBM PureFlex System.
The Central Processing Unit
Status and plans for online installation LHCb Installation Review April, 12 th 2005 Niko Neufeld for the LHCb Online team.
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
Lessons from HLT benchmarking (For the new Farm) Rainer Schwemmer, LHCb Computing Workshop 2014.
Wright Technology Corp. Minh Duong Tina Mendoza Tina Mendoza Mark Rivera.
PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.
Virtualization for the LHCb Online system CHEP Taipei Dedicato a Zio Renato Enrico Bonaccorsi, (CERN)
A Combat Support Agency Defense Information Systems Agency Virtualization 17 August 2011.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
APAN SIP SERVER Hosted at the APAN Tokyo XP Thanks to  Prof. Konishi for organizing this  Takatoshi Ikeda/ KDDI for mounting the server at APAN TokyoXP.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Niko Neufeld PH/LBC. Detector front-end electronics Eventbuilder network Eventbuilder PCs (software LLT) Eventfilter Farm up to 4000 servers Eventfilter.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
News from Alberto et al. Fibers document separated from the rest of the computing resources
Weekly Report By: Devin Trejo Week of June 21, 2015-> June 28, 2015.
Why it might be interesting to look at ARM Ben Couturier, Vijay Kartik Niko Neufeld, PH-LBC SFT Technical Group Meeting 08/10/2012.
Online System Status LHCb Week Beat Jost / Cern 9 June 2015.
MULTICORE PROCESSOR TECHNOLOGY.  Introduction  history  Why multi-core ?  What do you mean by multicore?  Multi core architecture  Comparison of.
Virtualization Supplemental Material beyond the textbook.
Niko Neufeld, CERN/PH. Online data filtering and processing (quasi-) realtime data reduction for high-rate detectors High bandwidth networking for data.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
CSUDH Fall 2015 Instructor: Robert Spengler
Future experiment specific needs for LHCb OpenFabrics/Infiniband Workshop at CERN Monday June 26 Sai Suman Cherukuwada Sai Suman Cherukuwada and Niko Neufeld.
Common meeting of CERN DAQ teams CERN May 3 rd 2006 Niko Neufeld PH/LBC for the LHCb Online team.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Niko Neufeld HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Chap 4: Processors Mainly manufactured by Intel and AMD Important features of Processors: Processor Speed (900MHz, 3.2 GHz) Multiprocessing Capabilities.
Hardware Status Online CRRC08 prep meeting Niko Neufeld.
Moore vs. Moore Rainer Schwemmer, LHCb Computing Workshop 2015.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
LHCb and InfiniBand on FPGA
Title of the Poster Supervised By: Prof.*********
Challenges in ALICE and LHCb in LHC Run3
Personal Computers A Research and Reverse Engineering
Slow Control Servers C. Irmler SVD 27th B2GM, KEK 21 June 2017.
Ingredients 24 x 1Gbit port switch with 2 x 10 Gbit uplinks  KCHF
Enrico Bonaccorsi, (CERN) Loic Brarda, (CERN) Gary Moine, (CERN)
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
Oxford Site Report HEPSYSMAN
The LHCb Event Building Strategy
Power couple. Dell EMC servers powered by Intel® Xeon® processors and running Windows Server* 2016, ready to securely handle dynamic business workloads.
Event Building With Smart NICs
High-Performance Storage System for the LHCb Experiment
Network Processors for a 1 MHz Trigger-DAQ System
Presentation transcript:

Farm Completion Beat Jost and Niko Neufeld LHCb Week St. Petersburg June 2010

Filling the farm Thanks for interesting and useful discussions to –Loic Barda, Rolf Lindner, Laurent Roy and Eric Thomas Thanks for measurements and plots to –Juan Caicedo and Patrick Robbe Farm Completion St. Petersburg 06/ Niko Neufeld 2

The three limits: Power, Cooling, Money Power: 550 kW available (105 kW used) Cooling: nominally available 525 kW Rack-space: 1700 Us (plenty) Money: xx MCHF Farm Completion St. Petersburg 06/ Niko Neufeld 3

Event Filter Farm Level 1: –100 SuperMicro Twin servers (2 servers in a single 1U chassis with shared power- supply), Intel Harpertown CPU 5420 (2.5 GHz) 4 cores / socket, 1 GB RAM /core Level 2: –350 DELL Bladeservers (up to 16 blades in a 10 U chassis), Intel Harpertown CPU 5420 (2.5 GHz) 4 cores / socket, 2 GB RAM /core Farm Completion St. Petersburg 06/ Niko Neufeld 4

The new farm-node Both Intel and AMD have brought out new processors: with up to 12 cores / chip and (Intel) hyper-threads (a.k.a. virtual CPUs) Memory has (again) become faster and cheaper (DDR-3) and each processor has 3 memory channels (  “good” memory configuration = 3 * n, where n = 2, 4, 8, 16 Both processors are now NUMA (non-uniform memory access) –Study program ongoing to take profit from this Farm Completion St. Petersburg 06/ Niko Neufeld 5

How many jobs / server Farm Completion St. Petersburg 06/ Niko Neufeld 6

How fast? Farm Completion St. Petersburg 06/ Niko Neufeld 7

Server specifications 1 GB RAM per hardware thread == virtual core 1 Power supply failure should not affect more than 2 units 2 Gigabit Ethernet ports No constraints on power-consumption CPU (AMD 61xx / Intel 56xx) chosen such as to optimise the Moore/CHF Farm Completion St. Petersburg 06/ Niko Neufeld 8

A likely candidate 1.2 kW –redundant PS 4 servers with each –12 cores –24 GB (up to 96) RAM –1 HDD –2 x Gigabit Ethernet 21 kCHF list-price Farm Completion St. Petersburg 06/ Niko Neufeld 9

Conclusions We will run with 16 Moore jobs / server (twice as many as today) Each server will be 2 to 2.5 x faster than the current HLT node Each Moore instance can use up to 1.5 GB RAM –If really need more RAM 1.Reduce number of jobs 2.Increase (double) memory Farm Completion St. Petersburg 06/ Niko Neufeld 10

Procedure / planning StepDuration Decision to buy (day X)0 Technical specifications to firms1 week Firms reply (with offer) / validation of sample server 4 weeks Adjudication (negotiation)1 week Delivery (in batches if possible installation starts as soon as delivered) 6 weeks Finishing installation1 week Farm Level 3 in production13 weeks after initial decision Farm Completion St. Petersburg 06/ Niko Neufeld 11

To-do list Hardware Unpacking (surface SX8  need a lot of space and friendly volunteers) Installation in D1 –Power, network Burn-in (3 days) Exchange faulty servers / parts Software Install OS, verify OS tuning (NIC, memory arrangement etc…) Integrate in software- management (Quattor) Add to farm-control Farm Completion St. Petersburg 06/ Niko Neufeld 12

DETAILS Farm Completion St. Petersburg 06/ Niko Neufeld 13

Farm Completion St. Petersburg 06/ Niko Neufeld 14

How fast? (Moore v9r2 HLT1 only) DAQ & electronics upgrade - Niko Neufeld 15