D0 Taking Stock 11/2004 - 06/2005 Calibration Database Servers.

Slides:



Advertisements
Similar presentations
Module 13: Implementing ISA Server 2004 Enterprise Edition: Site-to-Site VPN Scenario.
Advertisements

Deploying GMP Applications Scott Fry, Director of Professional Services.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
Adding scalability to legacy PHP web applications Overview Mario A. Valdez-Ramirez.
An Example of IPv6 Necessity in the Greek School Network Athanassios Liakopoulos Greek Research & Technology Network.
1.1 Installing Windows Server 2008 Windows Server 2008 Editions Windows Server 2008 Installation Requirements X64 Installation Considerations Preparing.
DATABASE MANAGEMENT SYSTEMS 2 ANGELITO I. CUNANAN JR.
11 SERVER CLUSTERING Chapter 6. Chapter 6: SERVER CLUSTERING2 OVERVIEW  List the types of server clusters.  Determine which type of cluster to use for.
VMware vCenter Server Module 4.
Gareth Smith RAL PPD HEP Sysman. April 2003 RAL Particle Physics Department Site Report.

The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
Module 12: Designing an AD LDS Implementation. AD LDS Usage AD LDS is most commonly used as a solution to the following requirements: Providing an LDAP-based.
Working Group Meeting (McGrew/Toki) Discussion Items (focused on R&D software section) Goals of Simulations Simulation detector models to try Simulation.
Module 12: Designing High Availability in Windows Server ® 2008.
Web Migration Project Update for ATC April 28, 2009.
Windows 2000 Advanced Server and Clustering Prepared by: Tetsu Nagayama Russ Smith Dale Pena.
UNIT 1 INFRASTRUCTURE AND APPLICATION SUPPORT. UNIT OBJECTIVES Name the different ROSS application platforms. Describe the difference between client hardware,
Module 1: Installing and Upgrading to Exchange Server 2003.
The SLAC Cluster Chuck Boeheim Assistant Director, SLAC Computing Services.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
LAL Site Report Michel Jouvin LAL / IN2P3
Indiana University’s Name for its Sakai Implementation Oncourse CL (Collaborative Learning) Active Users = 112,341 Sites.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Open Search Office Web Services Database Doc Mgt Sys Pipeline Index Geospatial Analysis Text Search Faceting Caching Query parsing Clustering Synonyms.
History of Microprocessor MPIntroductionData BusAddress Bus
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
DONE-08 Sizing and Performance Tuning N-Tier Applications Mike Furgal Performance Manager Progress Software
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
June 6, 2007TeraGrid '071 Clustering the Reliable File Transfer Service Jim Basney and Patrick Duda NCSA, University of Illinois This material is based.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
Testing… Testing… 1, 2, 3.x... Performance Testing of Pi on NT George Krc Mead Paper.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
26SEP03 2 nd SAR Workshop Oklahoma University Dick Greenwood Louisiana Tech University LaTech IAC Site Report.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
19th September 2003Tim Adye1 RAL Tier A Status Tim Adye Rutherford Appleton Laboratory BaBar UK Collaboration Meeting Royal Holloway 19 th September 2003.
Server Performance, Scaling, Reliability and Configuration Norman White.
SPW1 Database Architecture Networked (DAN) January 29, 2003 Stephen P. White – CEPA/DBS.
System Manager February 11, What is System Manager System Manager unites Inter-Tel’s diverse product line into a family of products that can be.
Disk Farms at Jefferson Lab Bryan Hess
A B A B AR InterGrid Testbed Proposal for discussion Robin Middleton/Roger Barlow Rome: October 2001.
MULTICORE PROCESSOR TECHNOLOGY.  Introduction  history  Why multi-core ?  What do you mean by multicore?  Multi core architecture  Comparison of.
Partner Logo A Tier1 Centre at RAL and more John Gordon eScience Centre CLRC-RAL HEPiX/HEPNT - Catania 19th April 2002.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
3D Testing and Monitoring Lee Lueking LCG 3D Meeting Sept. 15, 2005.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
Lisa Giacchetti AFS: What is everyone doing? LISA GIACCHETTI Operating Systems Support.
Presented By: Kasey Campbell John Geer.  Hermes Company Transfer will allow the passing of files, large or small, between companies.  All files are.
CPU on the farms Yen-Chu Chen, Roman Lysak, Miro Siket, Stephen Wolbers June 4, 2003.
PIC port d’informació científica DateText1 November 2009 (Elena Planas) PIC Site review.
Tier1A Status Martin Bly 28 April CPU Farm Older hardware: –108 dual processors (450, 600 and 1GHz) –156 dual processor 1400MHz PIII Recent delivery:
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
The Storage Resource Broker and.
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
LHC Logging Cluster Nilo Segura IT/DB. Agenda ● Hardware Components ● Software Components ● Transparent Application Failover ● Service definition.
FroNtier Stress Tests at Tier-0 Status report Luis Ramos LCG3D Workshop – September 13, 2006.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
CEG 2400 FALL 2012 Windows Servers Network Operating Systems.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Database Requirements Updates from LHC Experiments WLCG Grid Deployment Board Meeting CERN, Geneva, Switzerland February 7, 2007 Alexandre Vaniachine (Argonne)
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
Distributed Databases
NGS Oracle Service.
Conditions Data access using FroNTier Squid cache Server
Remote SAM Initiative (RSI) – Proposed Work Plan
Presentation transcript:

D0 Taking Stock 11/ /2005 Calibration Database Servers

Overview Overview Successfully running during the last calendar year Successfully running during the last calendar year No major problems have arisen No major problems have arisen Mature product with low maintenance requirements Mature product with low maintenance requirements 10 different types of Calibration dbservers currently running 10 different types of Calibration dbservers currently running Muon servers, farms and users, are available but are still not being used Muon servers, farms and users, are available but are still not being used Can d0reco be changed so these servers be can shut down? Can d0reco be changed so these servers be can shut down?

Deployment Remote Servers now exist at: Remote Servers now exist at: Karlsluhe, IN2P3, Imperal College, WestGrid, SAR Karlsluhe, IN2P3, Imperal College, WestGrid, SAR Remote sites beginning to run within the next few weeks: Remote sites beginning to run within the next few weeks: OSCER (this is mostly Univ. of Oklahoma) Prague Manchester Wisconsin CMS (at fnal) Lancaster (it may take bit for Lancaster to get going due to manpower limits) RAL (same comment as for Lancaster) OSCER (this is mostly Univ. of Oklahoma) Prague Manchester Wisconsin CMS (at fnal) Lancaster (it may take bit for Lancaster to get going due to manpower limits) RAL (same comment as for Lancaster) Possible new remote sites: Possible new remote sites: Michigan State, Brookhaven, SLAC Michigan State, Brookhaven, SLAC Remote Servers are processing data with p of d0reco. Remote Servers are processing data with p of d0reco.

Hardware Production nodes Production nodes Two nodes, with servers divided for farms/users Two nodes, with servers divided for farms/users Dual 1.5 GHZ AMD Athlon w/2 GB Memory Dual 1.5 GHZ AMD Athlon w/2 GB Memory Failover nodes Failover nodes Two nodes, one serving each production node Two nodes, one serving each production node 1, Dual 1.5 GHZ AMD Athlon w/2 GB Memory (farms) 1, Dual 1.5 GHZ AMD Athlon w/2 GB Memory (farms) 1, Dual 1.0 GHZ Pentium III w/1 GB Memory (users) 1, Dual 1.0 GHZ Pentium III w/1 GB Memory (users) Hardware is supporting current loads and is expected to support load increase. Hardware is supporting current loads and is expected to support load increase. Contingency plan is to add new nodes and servers. Contingency plan is to add new nodes and servers.

Concerns & Plans Increasing Farms Load Increasing Farms Load 160 Additional dual processor, and faster, CPUs being added 160 Additional dual processor, and faster, CPUs being added Db servers are expected to be able to handle the expected load increase. Db servers are expected to be able to handle the expected load increase. Contingency plan: Contingency plan: Added a new node with additional servers, split the farm requests across servers, develop a load manager. Added a new node with additional servers, split the farm requests across servers, develop a load manager.

Concerns & Plans Multi-Run Data Sets Multi-Run Data Sets For each distinct run the client will request an entire calibration set. For each distinct run the client will request an entire calibration set. Db servers may face difficulties trying to handle diverse multi-run sets of events from multiple users. Db servers may face difficulties trying to handle diverse multi-run sets of events from multiple users. Possible solutions: Possible solutions: Double the size of the cache from ~20 runs to ~40runs Double the size of the cache from ~20 runs to ~40runs May require additional Linux boxes. May require additional Linux boxes. Add additional Linux boxes with large cache servers. Develop a load manager for distribution of requests. Add additional Linux boxes with large cache servers. Develop a load manager for distribution of requests. Maintenance & Support Maintenance & Support Upgrade is underway to version 4.x of omniORBpy. Upgrade is underway to version 4.x of omniORBpy. Some minor bugs will be addressed during the upgrade. Some minor bugs will be addressed during the upgrade. Continue providing support for remote sites. Continue providing support for remote sites.