openLab Technical Manager

Slides:



Advertisements
Similar presentations
Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
Advertisements

Computing Infrastructure
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Bernd Panzer-Steindel, CERN/IT 2 * 50 Itanium Server (dual 1.3/1.5 GHz Itanium2, 2 GB mem) High Througput Prototype (openlab + LCG prototype) (specific.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
ALICE Data Challenge V P. VANDE VYVRE – CERN/PH LCG PEB - CERN March 2004.
RLS Tier-1 Deployment James Casey, PPARC-LCG Fellow, CERN 10 th GridPP Meeting, CERN, 3 rd June 2004.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
May 2004Sverre Jarp1 Preparing the computing solutions for the Large Hadron Collider (LHC) at CERN Sverre Jarp, openlab CTO IT Department, CERN.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
CERN Physics Database Services and Plans Maria Girone, CERN-IT
NL Service Challenge Plans Kors Bos, Sander Klous, Davide Salomoni (NIKHEF) Pieter de Boer, Mark van de Sanden, Huub Stoffers, Ron Trompert, Jules Wolfrat.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
SJ – Mar The “opencluster” in “openlab” A technical overview Sverre Jarp IT Division CERN.
CERN openlab II (2006 – 2008) Grid-related activities Sverre Jarp CERN openlab CTO sverre.jarp at cern.ch.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
SJ – Nov CERN’s openlab Project Sverre Jarp, Wolfgang von Rüden IT Division CERN 29 November 2002.
BNL Oracle database services status and future plans Carlos Fernando Gamboa, John DeStefano, Dantong Yu Grid Group, RACF Facility Brookhaven National Lab,
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
CERN News on Grid and openlab François Fluckiger, Manager, CERN openlab for DataGrid Applications.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
SJ – June CERN openlab for DataGrid applications Sverre Jarp CERN openlab CTO IT Department, CERN.
CASTOR: possible evolution into the LHC era
ALICE Computing Data Challenge VI
OpenCluster Planning Sverre Jarp IT Division CERN October 2002.
The Beijing Tier 2: status and plans
“A Data Movement Service for the LHC”
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Dirk Duellmann CERN IT/PSS and 3D
OpenLab Enterasys Meeting
NL Service Challenge Plans
IT-DB Physics Services Planning for LHC start-up
LCG 3D Distributed Deployment of Databases
CERN openlab for DataGrid applications Programme of Work Overview F
Service Challenge 3 CERN
Bernd Panzer-Steindel, CERN/IT
Cluster Active Archive
The INFN TIER1 Regional Centre
NGS Oracle Service.
Conditions Data access using FroNTier Squid cache Server
Workshop Summary Dirk Duellmann.
The INFN Tier-1 Storage Implementation
The CERN openlab for DataGrid Applications François Fluckiger
Oracle Storage Performance Studies
Grid Canada Testbed using HEP applications
Web Server Administration
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Cluster Computers.
Presentation transcript:

openLab Technical Manager openlab overview CHEP2004 Sverre Jarp openLab Technical Manager IT Department CERN September 2004 Sverre Jarp

In partnership with September 2004 Sverre Jarp

CERN openlab Department’s main R&D focus Framework for collaboration with industry Evaluation, integration, validation of cutting-edge technologies that can serve LCG Initially a 3-year lifetime As of 1.1.2003 Later: Annual prolongations LCG CERN openlab 02 03 04 05 06 07 08 September 2004 Sverre Jarp

openlab participation 5 current partners Enterasys: 10 GbE core routers HP: Integrity servers (103 * 2-ways, 2 * 4-ways) Two post-doc positions IBM: Storage Tank file system (SAN FS) w/metadata servers and data servers (currently with 28 TB) Intel: Large number of 64-bit Itanium processors & 10 Gbps NICs 64-bit Nocona system w/PCI-Express Oracle: 10g Database software w/add-on’s One contributor Voltaire 96-way Infiniband switch and necessary HCAs Because of the time limit of this talk, only certain highlights of our activities can be covered! September 2004 Sverre Jarp

Full integration with the LCG testbed 10GE WAN connection 4 * GE connections to the backbone 4 *ENTERASYS N7 10 GE Switches 2 * Enterasys “new” Series 36 Disk Server (dual P4, IDE disks, ~ 1TB disk space each) 10 GE per node 10 GE per node 2 * 100 IA32 CPU Server (dual 2.4 GHz P4, 1 GB mem.) 10GE 1 GE per node 2 * 50 Itanium Server (dual 1.3/1.5 GHz Itanium2, 2 GB mem) 28 TB , IBM StorageTank 12 Tape Server STK 9940B High Througput Prototype (openlab + LCG prototype) (specific layout, October 2004) September 2004 Sverre Jarp

Voltaire Infiniband is a new technology with multiple strengths: Low latency High throughput Low processor overhead Connectivity to other protocols In particular: 1 Gb Ethernet (10 GbE later) Fibre-channel disks September 2004 Sverre Jarp

Enterasys September 2004 Sverre Jarp

Alice Data Challenge HW Architecture 4 x GE 10GE 10 Tape Server 3COM 4900 3COM 4900 16 x Gbit 32 x GE 32 IA64 HP-rx2600 Servers 2 x 1 GHz Itanium-2 2 GB RAM Broadcom NetXtrem BCM5701 (tg3) RedHat Advanced Workstation 2.1 6.4 GB/s to memory, 4.0 GB/s to I/O Enterasys E1 OAS 12 Gbit, 1 x 10 Gbit GDCs LDCs Enterasys ER16 16 slots 4/8 x Gbit or 1 x 10 Gbit/slot 4 x 7 Disk servers 2 x 2.0 GHz Xeon 1 GB RAM Intel 82544GC 3COM 4900 ~ 80 CPU servers 2 x 2.4 GHz Xeon, 1 GB RAM, Intel 8254EM Gigabit in PCI-X 133 (Intel PRO/1000), CERN Linux 7.3.3 September 2004 Sverre Jarp

Achievements (as seen by Alice) Sustained bandwidth to tape: Peak 350 MB/s Reached production-quality level only last week of testing Sustained 280 MB/s over 1 day but with interventions [goal was 300] Itanium systems from openlab successfully integrated in the ADC Very satisfactory stability Goal for ADC VI: 450 MB/s September 2004 Sverre Jarp

New generations In openlab we are currently using N7 “mid-range” routers Used successfully in tests with IBM StorageTank Recently, Enterasys has provided us with their most recent systems See Enterasys’s presentations and documentation here at CHEP2004 for details September 2004 Sverre Jarp

HP/Intel September 2004 Sverre Jarp

Use of Itanium systems Why Itanium? Choice made in already in 2002 Neither Intel EM64T or AMD64 available Pure 64-bit approach forces “complete conversion” to new mode Ported: ROOT, CLHEP, GEANT4, ALIROOT, LCG2, etc. HP Itanium servers Have excellent stability and I/O capabilities We use standard “Scientific Linux CERN 3” Intel and GNU compilers Very good performance monitoring tools For both application and system performance SPECint performance is adequate ~1300 SPECint ~1100 ROOTmarks with “stress” Eagerly awaiting Dual-core “Montecito” processors next year Intel is also very “bullish” on evolution towards 2007 September 2004 Sverre Jarp

Gridification A good success story: Starting point: The software chosen for LCG (VDT + EDG) had been developed only with IA32 (and specific Red Hat versions) in mind Consequence: Configure-files and make-files not prepared for multiple architectures. Source files not available in distributions (often not even locatable) Stephen Eccles, Andreas Unterkircher worked for many months to complete the porting of LCG-2 Result: All major components now work on Itanium/Linux: Worker Nodes, Compute Elements, Storage Elements, User Interface, etc. Well tested inside the Test Grid Code, available via Web-site, transferred to HP sites (Initially Puerto Rico and Bristol) Changes given back to developers VDT now built also for Itanium systems Porting experience summarized in white paper (on the Web) From now on the Grid is heterogeneous! September 2004 Sverre Jarp

Next generation disk servers Based on state-of-the-art equipment: 4-way Itanium server (RX4640) Two full-speed PCI-X slots 10 GbE and/or Infiniband 24 * S-ATA disks with 74 GB WD740 “Raptor” @ 10k rpm Burst speed of 100 MB/s Two 3ware 9500 RAID controllers In excess of 770 MB/s RAID-5 read speed Single stream, sequential Only 340 MB/s for write w/RAID 5 Read speed with RFIO is 650 MB/s (multiple streams) September 2004 Sverre Jarp

10 Gbps WAN tests (between CERN and CalTech) Initial breakthrough during Telecom-2003 with IPv4 (single/multiple) streams: 5.44 Gbps Linux, Itanium-2 (RX 2600), Intel 10Gbps NIC Also IPv6 (single/multiple) streams In June Again IPv4, and single stream (Datatag/Openlab): 6.55 Gbps with Linux, Itanium-2 (RX4640), S2IO NIC In September: Same conditions as before: 7.29 Gbps But SuNET with a much longer lightpath has just grabbed the record, even if they only reach 4.3 Gbps. We will be back! September 2004 Sverre Jarp

Data export to LCG Tier-1/-2 MSU RAL IC IN2P3 Data distribution ~70 Gbits/sec IFCA UB FNAL Cambridge Tests (initially) between CERN and Fermilab + BNL +NIKHEF (and others) Multiple HP Itanium servers with dual 1 Gb NICs Disk to disk transfers via GridFTP Each server: 100 MB/s IN + 100 MB/s OUT Aggregation of multiple streams across 10 GbE link Similar tuning as Internet2 tests Goal: Achieve 5 Gbps sustained by Dec. 04 Budapest CNAF Prague FZK Taipei PIC TRIUMF ICEPP BNL Legnaro CSCS Rome CIEMAT USC Krakow NIKHEF “Service Data Challenge” Stability is paramount – no longer just “raw” speed September 2004 Sverre Jarp

IBM September 2004 Sverre Jarp

StorageTank Collaboration directly between CERN and IBM Almaden CERN expressed the desire to use Linux clients and iSCSI protocol right from the start Not available in initial SAN FS product Random Access test CMS pileup “look-alike” developed by Rainer Többicke Scenario: 100 GB dataset, randomly accessed in ~50kB blocks 1 – 100 2 GHz P4-class clients, running 3 – 10000 “jobs” Hardware 4 IBM x335 metadata servers 8 IBM 200i controllers, 336 SCSI disks Added 2 IBM x345 servers as disk controllers after the test Results (after one week’s running) Peak data rate: 484 MB/s (with 9855 simultaneous “jobs”) After the test, special tuning, 10 servers, smaller number of clients: 705 MB/s September 2004 Sverre Jarp

StorageTank Plans Plans for Phase 2 Alice Computing Data Challenge VI Scenario ~50 input streams, ~50 output streams Data staged from input through disk subsystem to tape Goal is 450 MB/s “end-to-end” for 1 week (or more) Sizing the system Reasonable data block size, e.g. 256 KB Note that 450 MB/s end-to-end means 450 in + 450 out! Sufficient margin for dead time during tape mounts, tape file switching, etc. 1.5 GB/s disk bandwidth may be required September 2004 Sverre Jarp

Oracle September 2004 Sverre Jarp

Oracle Activities to date Two Oracle-funded research fellows since early 2004 Work focuses on prototyping new Oracle technologies, based on the needs of critical services, such as RLS (Replica Location Service) First results obtained on DataGuard Maintains a hot backup of the catalogue which can be used while the main instance is not available for operating system scheduled interventions Streams Allows asynchronous database replication 10g installed on Itanium cluster node September 2004 Sverre Jarp

Future Activities Significant activity expected in the area of Oracle Clusters Application Server Clusters + DB Clusters (RAC) Further exploitation of RAC, which is a technology that services a number of purposes: High availability - proven technology used in production at CERN to avoid single points of failure with transparent application failover Consolidation – allows a smallish number (1 – 8?, 1 – 64?) of database services to be consolidated into a single management entity Scalability – SAN-based infrastructure means CPU nodes and storage can be added to meet demand Many opportunities Further development of Streams Organize a “challenge” within DB group Move to Oracle-managed storage (ASM) ? Move to “the Oracle Grid” for IAS/DB infrastructure? Exploit many other 10g features? “Big file table-space” for Ultra-Large DBs (ULDBs) September 2004 Sverre Jarp

Conclusions CERN openlab: Solid collaboration with our industrial partners Encouraging results in multiple domains We believe partners are getting good “ROI” But only they can really confirm it No risk of running short of R&D IT Technology is still moving at an incredible pace Vital for LCG that the “right” pieces of technology are available for deployment Performance, cost, resilience, etc. Likely ingredients identified for LCG (so far): 64-bit programming, iSCSI, next generation I/O (10 Gb Ethernet, Infiniband, etc.) During summer: addition of 6 summer students September 2004 Sverre Jarp