Control System Virtualization for the LHCb Online System ICALEPCS – San Francisco Enrico Bonaccorsi, (CERN)

Slides:



Advertisements
Similar presentations
Netbus: A Transparent Mechanism for Remote Device Access in Virtualized Systems Sanjay Kumar PhD Student Advisor: Prof. Karsten Schwan.
Advertisements

Virtual Switching Without a Hypervisor for a More Secure Cloud Xin Jin Princeton University Joint work with Eric Keller(UPenn) and Jennifer Rexford(Princeton)
What’s New: Windows Server 2012 R2 Tim Vander Kooi Systems Architect
Overview.
© 2009 IBM Corporation Statements of IBM future plans and directions are provided for information purposes only. Plans and direction are subject to change.
Profit from the cloud TM Parallels Dynamic Infrastructure AndOpenStack.
© 2010 VMware Inc. All rights reserved Confidential Performance Tuning for Windows Guest OS IT Pro Camp Presented by: Matthew Mitchell.
VSphere vs. Hyper-V Metron Performance Showdown. Objectives Architecture Available metrics Challenges in virtual environments Test environment and methods.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
VMware Infrastructure Alex Dementsov Tao Yang Clarkson University Feb 28, 2007.
PALM-3000 ATST/BBSO Visit Stephen Guiwits P3K System Hardware 126 Cahill February 11, 2010.
DatacenterMicrosoft Azure Consistency Connectivity Code.
Lesson 9: Creating and Configuring Virtual Networks
Do MUCH More with Less Presented by: Jon Farley 2W Technologies.
L. Granado Cardoso, F. Varela, N. Neufeld, C. Gaspar, C. Haen, CERN, Geneva, Switzerland D. Galli, INFN, Bologna, Italy ICALEPCS, October 2011.
Introduction to DoC Private Cloud
European Organization for Nuclear Research Virtualization Review and Discussion Omer Khalid 17 th June 2010.
Virtualization – to cluster or not to cluster 4/17/2012.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Virtualization Performance H. Reza Taheri Senior Staff Eng. VMware.
Virtualization A way To Begin with Virtual Reality… - Rahul Khanwani.
SUNY IT Master's Project Using Open Source Virtualization Technology In Computer Education By: Ronny L. Bull Advised By: Geethapriya Thamilarasu, Ph.D.
1. Outline Introduction Virtualization Platform - Hypervisor High-level NAS Functions Applications Supported NAS models 2.
Module 9 PS-M4110 Overview <Place supporting graphic here>
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Windows Server 2012 VSP Windows Kernel Applications Non-Hypervisor Aware OS Windows Server 2008, 2012 Windows Kernel VSC VMBus Emulation “Designed for.
Corporate Partner Overview and Update September 27, 2007 Gary Crane SURA Director IT Initiatives.
Andrey Zakharov Technical Sales Consultant, St. Petersburg office LinxCloud global team.
CERN IT Department CH-1211 Geneva 23 Switzerland t Experience with NetApp at CERN IT/DB Giacomo Tenaglia on behalf of Eric Grancher Ruben.
SAIGONTECH COPPERATIVE EDUCATION NETWORKING Spring 2009 Seminar #1 VIRTUALIZATION EVERYWHERE.
Sydney Region IT School Support Term Smaller Servers available on Contract.
Module 7: Hyper-V. Module Overview List the new features of Hyper-V Configure Hyper-V virtual machines.
VMware Infrastructure 3 The Next Generation in Virtualization.
Planning and Designing Server Virtualisation.
Network Plus Virtualization Concepts. Virtualization Overview Virtualization is the emulation of a computer environment called a Virtual Machine. A Hypervisor.
A study of introduction of the virtualization technology into operator consoles T.Ohata, M.Ishii / SPring-8 ICALEPCS 2005, October 10-14, 2005 Geneva,
MClock: Handling Throughput Variability for Hypervisor IO Scheduling in USENIX conference on Operating Systems Design and Implementation (OSDI ) 2010.
CERN IT Department CH-1211 Genève 23 Switzerland t Evolution of virtual infrastructure with Hyper-V Juraj Sucik, Slavomir Kubacka Internet.
CERN IT Department CH-1211 Geneva 23 Switzerland t IT/DB Tests and evolution SSD as flash cache.
Windows Server 2012 Hyper-V Networking
Virtualization for the LHCb Online system CHEP Taipei Dedicato a Zio Renato Enrico Bonaccorsi, (CERN)
Microsoft Virtual Academy. Microsoft Virtual Academy First HalfSecond Half (01) Introduction to Microsoft Virtualization(05) Hyper-V Management (02) Hyper-V.
Running Linux on Hyper-V and in Azure Anurag Gupta M357.
The DCS lab. Computer infrastructure Peter Chochula.
Consolidation and Optimization Best Practices: SQL Server 2008 and Hyper-V Dandy Weyn | Microsoft Corp. Antwerp, March
VMware vSphere Configuration and Management v6
Queensland University of Technology CRICOS No J VMware as implemented by the ITS department, QUT Scott Brewster 7 December 2006.
Rick Claus Sr. Technical Evangelist,
Copyright © 2005 VMware, Inc. All rights reserved. How virtualization can enable your business Richard Allen, IBM Alliance, VMware
Arne Wiebalck -- VM Performance: I/O
Management of the LHCb Online Network Based on SCADA System Guoming Liu * †, Niko Neufeld † * University of Ferrara, Italy † CERN, Geneva, Switzerland.
Microsoft Windows Server 2012 R2. What’s NEW in Windows Server 2012 R2.
Jérôme Jaussaud, Senior Product Manager
| Basel Fabric Management with Virtual Machine Manager Philipp Witschi – Cloud Architect & Microsoft vTSP Thomas Maurer – Cloud Architect & Microsoft MVP.
© 2015 VMware Inc. All rights reserved. Software-Defined Data Center Module 2.
Workload Active directory BizTalk server DHCP DNS Dynamics Exchange server Fax server IIS Lync server RDS SharePoint server SQL System Center Visual.
1 Enrico Bonaccorsi on behalf of ALICE, ATLAS, CMS, LHCb Virtualisation Enrico Bonaccorsi on behalf of ALICE, ATLAS, CMS, LHCb Virtualisation  Introduction.
Deploying disk deduplication for Hyper-v 3.0 Žigmund Maťašovský.
E2800 Marco Deveronico All Flash or Hybrid system
Power Systems with POWER8 Technical Sales Skills V1
Video Security Design Workshop:
RHEV Platform at LHCb Red Hat at CERN 17-18/1/17
Wilson Trailer Approach to Disaster Recovery
Virtualization OVERVIEW
Enrico Bonaccorsi, (CERN) Loic Brarda, (CERN) Gary Moine, (CERN)
Статус ГРИД-кластера ИЯФ СО РАН.
Overview Introduction VPS Understanding VPS Architecture
Conditions leading to the rise of virtual machines
Hyper-Converged Technology a User's Perspective
Microsoft Virtual Academy
Presentation transcript:

Control System Virtualization for the LHCb Online System ICALEPCS – San Francisco Enrico Bonaccorsi, (CERN) Luis Granado Cardoso, Niko Neufeld (CERN, Geneva) Francesco Sborzacchi (INFN/LNF, Frascati (Roma))

LHCb & Virtualization 2 Completely isolated network –Data Acquisition System –Experiment Control System Why do we virtualize –Improve manageability –High Availability –Hardware usability Better usage of hardware resources Move away from the model “one server = one application” – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

What we are virtualizing Around 200 control PCs running WinCC OA o 150 linux Red Hat / CentOS / Scientific Linux 6 o 50 windows Windows 2008 R Web Servers Gateways o Linux SSH and NX o Windows terminal services Common infrastructure servers o DHCP, DNS, Domain Controllers, … 3 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

Current virtualization infrastructure 20 blade servers distributed in two chassis in different racks 4 x 10 Gb/s Ethernet switches 4 x 8 Gb/s Fiber channel (FC) switches 2 x NetApp 3270 accessible via FC and iSCSI o Hybrid storage pool: SSD + SATA 2 independent clusters o General Purpose Cluster (DNS, DHCP, Web services,..) o Control Cluster (Dedicated to the control system) 4 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

Shared Storage Crucial component of the virtualization infrastructure Required for high availability Performance is a key point We want to guarantee a minimum of 40 random IOPS per VM o The equivalent experience of using a laptop with a 5400 RPM HDD 5 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

2x 8Gb Multi-switch interconnection Netapp FAS3270 Controller B Controller A RAID 6 Synchronous Mirroring RHEV-General RHEV-Control RHEV-GeneralRHEV-Control FC-Switch A1 FC-Switch B1 FC-Switch A2 FC-Switch B2 Mezzanine dual port FC Mezzanine dual port FC 8Gb FC Storage area network 7 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

Stack 80 Gb/s LHCb Experiment Network Force 10 Internal switches LACP 20Gb/s RHEV-General RHEV-Control FC-Switch B1 Mezzanine dual port 10Gb LACP 20Gb/s Mezzanine dual port 10Gb Internal switches Ethernet network 8 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

Resources optimization High performance hardware is expensive Storage deduplication o Eliminates duplicates of repeated data o Currently saving 67% of used storage space o Provides improvements in terms of IOPS (less data to cache!) Kernel Shared Memory o Maximize the usage of memory o Merge the same memory pages allowing overcommitting of memory without swapping 8 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

 Blade Poweredge M610  2 x 2.4GHz (8 real cores + Hyper Threading)  3 x 8 GB = 24GB RAM  2 x 10Gb network interfaces  2 X 1Gb network interfaces  2 X 8Gb fiber channel interfaces  Storage  4 X 8Gb Fiber channel switches  SSD pool + SATA  Deduplication ON  Network  4 X 10Gb Ethernet switches  4 X 1Gb Ethernet switches  Limits:  Average of 15 VM per Server Storage (random) o IOPS=45K Throughput=153MB/s Latency= ~10ms Storage (random) IOPS=45K Throughput=153MB/s writing, 300MB/s reading Latency= ~10ms Network Throughput = 5.37 Gb/s Latency = 0.15 ms for 1400B Benchmarks Netapp 3270, random Reading 4k + random Writing 4k, 215VMs, 200MB/VM Comparison during normal operation and during deduplication and takeover 9 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

150 WinCC OA Projects (WINCC001.. WINCC150) o 1 project per VM o Each project is connected to other 5 projects The two previous and after projects (according to the numbering The master project o Each project has 1000 datapoints created for writing o Each project performs dpSets locally and on the connected projects o Number of DPs to be set and rate are settable Each period the dps are selected randomly from the 1000 dps pool and set Testing the infrastructure: Pessimistic SCADA workloads 10 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

1 Master Project (WINCC001) o This project connects to all other projects o Has System Overview installed for easier control of the whole system FW version for PVSS 3.8 – produces a couple of errors but the PMON communication with the other projects works just fine Testing the infrastructure: Pessimistic SCADA workloads (2) 11 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

DateLocal Rate*Remote Rate*Total*CPU (%)Comment All OK All OK All OK Problems with 1 project (multiple disconnections/connections)** Decreased for live migration tests *dpSets per Second Results Summary Unattended long term test: – all VMs and projects performed stably – One instance had to be live migrated to solve some issues related to the real server We run the same tests that have been done for real machines by the CERN industrial control group (EN/ICE) and we obtained very similar results ** WINCC006, after some period, started disconnecting/connecting to WINCC005 and WINCC007 indefinitely. Problem was fixed by restarting the projects WINCC004 and WINCC008 which also connect to WINCC006. At the end of each “run” period, logs are collected and analysed for problems – PVSS_II.log, WCCOActrlNN.log are “grepped” for possible issues (“disconnect”, ”connect”, “queue”, “pending”, “lost”, …) Plots are also produced by calculating the rate from the dpSets timestamp (only local dpSets) 12 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

13 – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

Summary and outlook Virtualization of LHCb ECS o Reduce hardware o Achieving High Availability Storage the key component of the infrastructure Realistic SCADA workload emulator o Indispensable in the evaluation of many commercial storage systems Resources optimizations Performance results Migration of all Control PCs to VMs should be completed by Q – Control System Virtualization for LHCb – ICALEPCS 2013 – San Francisco

Backup slides 15

Hypervisors Essentially 3 choices: o Kernel based Virtual Machine (KVM) Currently used in LHCb Open source o VMWare: Most advanced even if closed source Too expensive for us o Hyper-V Core R2 and System Center Virtual Machine Manager (SCVMM) Almost for free (license needed for SCVMM) 16 – Control System Virtualization for the LHCb online system – ICALEPCS 2013 – San Francisco

Capacity planning 17 – Control System Virtualization for the LHCb online system – ICALEPCS 2013 – San Francisco