Vincenzo Vagnoni LHCb Real Time Trigger Challenge Meeting CERN, 24 th February 2005.

Slides:



Advertisements
Similar presentations
Operating System.
Advertisements

© 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Enigma Data’s SmartMove.
VMWare to Hyper-V FOR SERVER What we looked at before migration  Performance – Hyper-V performs at near native speeds.  OS Compatibility – Hyper-V.
Linux on commodity network H/W Josh Parsons LUGOD talk August 15 th 2005.
1 Emulab Security. 2 Current Security Model Threat model: No malicious authenticated users, Bad Guys are all “outside” –Protect against accidents on the.
14.1 © 2004 Pearson Education, Inc. Exam Planning, Implementing, and Maintaining a Microsoft Windows Server 2003 Active Directory Infrastructure.
Chapter 23: ARP, ICMP, DHCP IS333 Spring 2015.
L. Granado Cardoso, F. Varela, N. Neufeld, C. Gaspar, C. Haen, CERN, Geneva, Switzerland D. Galli, INFN, Bologna, Italy ICALEPCS, October 2011.
Installing software on personal computer
Control and monitoring of on-line trigger algorithms using a SCADA system Eric van Herwijnen Wednesday 15 th February 2006.
Automating Linux Installations at CERN G. Cancio, L. Cons, P. Defert, M. Olive, I. Reguero, C. Rossi IT/PDP, CERN presented by G. Cancio.
1 MASTERING (VIRTUAL) NETWORKS A Case Study of Virtualizing Internet Lab Avin Chen Borokhovich Michael Goldfeld Arik.
70-293: MCSE Guide to Planning a Microsoft Windows Server 2003 Network, Enhanced Chapter 14: Problem Recovery.
Methodologies, strategies and experiences Virtualization.
THE AFFORDABLE SUPERCOMPUTER HARRISON CARRANZA APARICIO CARRANZA JOSE REYES ALAMO CUNY – NEW YORK CITY COLLEGE OF TECHNOLOGY ECC Conference 2015 – June.
Computer security virus, hacking and backups. Computer viruses are small software programs that are designed to spread from one computer to another.
Installing and maintaining clusters of FreeBSD servers using PXE and Rsync Cor Bosman XS4ALL
User Management in LHCb Gary Moine, CERN 29/08/
COMP1321 Digital Infrastructure Richard Henson February 2012.
Configuration of Linux Terminal Server Group: LNS10A6 Thebe Laxmi, Sharma Prabhakar, Patrick Appiah.
Chapter Fourteen Windows XP Professional Fault Tolerance.
Distributed File Systems
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
Enabling Palacios PXE-Boot Chen Jin Bharath Pattabiraman Patrick Foley.
WIRELESS NETWORKS AT HOME Presentation 9 th December 2004 Simon G Fraser.
Installation Overview Lab#2 1Hanin Abdulrahman. Installing Ubuntu Linux is the process of copying operating system files from a CD, DVD, or USB flash.
Session objectives Discuss whether or not virtualization makes sense for Exchange 2013 Describe supportability of virtualization features Explain sizing.
Sandor Acs 05/07/
Scott Drucker, Systems Engineer Migrating to Microsoft Vista with WinINSTALL.
InstantGrid: A Framework for On- Demand Grid Point Construction R.S.C. Ho, K.K. Yin, D.C.M. Lee, D.H.F. Hung, C.L. Wang, and F.C.M. Lau Dept. of Computer.
Computing for LHCb-Italy Domenico Galli, Umberto Marconi and Vincenzo Vagnoni Genève, January 17, 2001.
Condor and DRBL Bruno Gonçalves & Stefan Boettcher Emory University.
Kickstart Installation
Infrastructure for the LHCb RTTC Artur Barczyk CERN/PH RTTC meeting,
Computer security virus, hacking and backups. Computer viruses are small software programs that are designed to spread from one computer to another.
ALICE Use of CMF (CC) for the installation of OS and basic S/W OPC servers and other special S/W installed and configured by hand PVSS project provided.
Status of Farm Monitor and Control CERN, February 24, 2005 Gianluca Peco, INFN Bologna.
RAL Site report John Gordon ITD October 1999
Minimalist’s Linux Cluster Changyoung Choi, Jeonghyun Kim, Seyong Kim Department of Physics Sejong University.
Online System Status LHCb Week Beat Jost / Cern 9 June 2015.
Status of the new NA60 “cluster” Objectives, implementation and utilization NA60 weekly meetings Pedro Martins 03/03/2005.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Construction methods and monitoring in meta-cluster systems LIT, JINR Korenkov V.V, Mitsyn V.V, Chkhaberidze D.V, Belyakov D.V.
Computer and Network Infrastructure for the LHCb RTTC Artur Barczyk CERN/PH-LBC RTTC meeting,
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
HOW TO INSTALL WINDOWS 7? This step-by-step guide demonstrates how to install Windows 7 Ultimate. The guide is similar for other versions of Windows 7.
INFSO-RI ETICS Local Setup Experiences A Case Study for Installation at Customers Location 4th. All Hands MeetingUwe Müller-Wilm VEGA Bologna, Nov.
Update on Farm Monitor and Control Domenico Galli, Bologna RTTC meeting Genève, 14 april 2004.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
DAQ & ConfDB Configuration DB workshop CERN September 21 st, 2005 Artur Barczyk & Niko Neufeld.
CS 283Computer Networks Spring 2013 Instructor: Yuan Xue.
Status report of the new NA60 “cluster” Our OpenMosix farm will increase our computing power, using the DAQ/monitoring computers. NA60 weekly meetings.
Active-HDL Server Farm Course 11. All materials updated on: September 30, 2004 Outline 1.Introduction 2.Advantages 3.Requirements 4.Installation 5.Architecture.
@Yuan Xue CS 283Computer Networks Spring 2011 Instructor: Yuan Xue.
Deterlab Tutorial CS 285 Network Security. What is Deterlab? Deterlab is a security-enhanced experimental infrastructure (based on Emulab) that supports.
1 Chapter 2: Operating-System Structures Services Interface provided to users & programmers –System calls (programmer access) –User level access to system.
Chapter 1 : Computer Networks. Lecture 1. Introduction to computer networks: Network definition : A network is a collection of computers and other devices.
Thousands of Linux Installations (and only one administrator) A Linux cluster client for the University of Manchester A V Le Blanc I T Services University.
 Each interface card that was detected correctly will be listed under the Network Devices section. Ethernet devices in Linux are named eth0, eth1, eth2,
1 Remote Installation Service Windows 2003 Server Prof. Abdul Hameed.
Updating FreeBSD Unix System Administration. Objectives At the end of this session you should be able to: 1. Understand the differences between the CURRENT.
Andrea Chierici Virtualization tutorial Catania 1-3 dicember 2010
SmartCenter for Pointsec - MI
Heterogeneous Computation Team HybriLIT
IM-pack: Software Installation Using Disk Images
How can a detector saturate a 10Gb link through a remote file system
Washington University
SUSE Linux Enterprise Desktop Administration
Quattor Advanced Tutorial, LAL
Bending Ironic for Big Iron
Presentation transcript:

Vincenzo Vagnoni LHCb Real Time Trigger Challenge Meeting CERN, 24 th February 2005

Vincenzo Vagnoni LHCb RTTC Meeting, 24 th February Diskless booting  4 ways (to my knowledge) to operate a linux diskless machine Removable device booting (e.g. a la Knoppix on cdrom) Not flexible enough, option discarded! High-reliability Mini-Drive or Disk-On-Chip booting Interesting… “firmware”-oriented approach, like modern x-terminals Never tried, adopting such a solution would depend on the availability of low cost devices of this kind Not a real option at the moment, but to be kept in mind… “Classic” network boot with “root over NFS” Used for example for old x-terminals, for CETIA motherboards, etc… Used in production for 4 years in Bologna for the data analysis farm and also for two years for the Bologna MC production farm Works fine Network boot with root filesystem on ramdisk I’m not aware of other people using it apart us in the Bologna L1&HLT testbed Root filesystem downloaded at boot time together with kernel via network Application software directories mounted via NFS (or other network filesystem protocols) Works fine

Vincenzo Vagnoni LHCb RTTC Meeting, 24 th February Classic network boot  Requires few basic services PXE, DHCP, TFTP servers on a control PC Requires root directories to be exported by a control PC Each machine will have its root filesystem as a specific directory on the control PC Installation of a new node just requires the update of the DHCP configuration, the copy of a template directory and export of the copied directory via the network filesystem  One drawback Reliability depends on the reliability of the network filesystem In case the network filesystem hangs, the root filesystem is frozen, the machine is not reachable anymore in no way (the kernel frezees and waits for the root filesystem to come back online) A potential problem? For example, NFS at CNAF has shown some serious problems: in a complex network environment with large (unwanted!) variable latencies, deadlocks of the (linux implementation) protocol had shown up. However, we should’n t have such non-controlled latencies in our online network, otherwise the trigger is dead…

Vincenzo Vagnoni LHCb RTTC Meeting, 24 th February Ramdisk network boot  Requires the same services as the “root over NFS” PXE, DHCP, TFTP servers on a control PC Installation of a new node just requires the update of the DHCP configuration Kernel and ramdisk with root filesystem image is downloaded at boot time The root filesystem is memory resident  Advantage The machine is always operative and reachable (unless the memory doesn’t break or a bit flip is triggered by a cosmic ray… but ECC memories are protected against single bit flip ) Unwanted corruption (mistakes) of the filesystem files is automatically restored at reboot (changes to the filesystem are just temporary and lost)  Drawbacks The ramdisk eats memory, typically order of 200 MB for a “normal” root filesystem (not a real problem however to loose just 200 MB) Requires recompiled kernel with large ramdisk size (not a real problem anyway) Application software is too large in any case and should be mounted via network

Vincenzo Vagnoni LHCb RTTC Meeting, 24 th February Control PC  Scalability of the network boot and of the NFS exports shouldn’t be an issue A control PC serves just a few subfarms However, it is better the control PC doesn’t live too far (“networkly” speaking) the served subfarms  An issue will be to keep up-to-date and syncronized all the Control PCs operating systems, the application software served, the operating system(s) served, etc. The control PCs are “regular” “disked” machines, and have on their disks the core of the system The way these PCs are managed is a core business for an efficient- flawless-costless operation of the farm

Vincenzo Vagnoni LHCb RTTC Meeting, 24 th February Preparation for the RTTC  Well in time for the RTTC, we should sit together and configure a testbed farm at CERN Administrator(s) should make experience with this configuration  Several other issues should be addressed VLANs, e.g., might interfere with network boot for example in case of using unmanaged switches which learn dynamically the VLANs from the nodes… the node doesn’t know anything about the VLAN until it is booted with an “intelligent” operating system Just an example, as we won’t use unmanaged switches to my knowledge, but however all the details with occasional problems and solutions should be settled down in time Thus, we should sit together and define in more details the hardware/software infrastructure of the RTTC farm

Vincenzo Vagnoni LHCb RTTC Meeting, 24 th February Relevant tools  All the operations to prepare a new node can be done in principle “by hand” Of course, not feasible for a 1800 PC farm  But, feasible for the RTTC as few nodes will be involved However (see Gianluca’s talk) it would be nice to have already for the RTTC a GUI that automatically triggers the work to be done (update of the configuration files on the control PC, e.g. DHCP, etc…) Needs some strict interaction between administrator(s) and PVSS GUIs developers  Remote control of electrical power Not really necessary for the RTTC, but still nice to have it in time IPMI solution (see Gianluca’s) controlled via PVSS GUIs Alternatively, remotely ethernet-controlled power switches (e.g. those used at CNAF) controlled via PVSS GUIs Of course, essential issue for the final online farm