FermiCloud Infrastructure as a Service (IaaS) Cloud Computing In Support of the Fermilab Scientific Program OSG All Hands Meeting 2012 Steven Timm

Slides:



Advertisements
Similar presentations
Remus: High Availability via Asynchronous Virtual Machine Replication
Advertisements

Virtual Machine Technology Dr. Gregor von Laszewski Dr. Lizhe Wang.
Fermilab, the Grid & Cloud Computing Department and the KISTI Collaboration GSDC - KISTI Workshop Jangsan-ri, Nov 4, 2011 Gabriele Garzoglio Grid & Cloud.
FermiGrid and FermiCloud Updates Keith Chadwick Grid & Cloud Computing Department Fermilab Work supported by the U.S. Department of Energy under contract.
Profit from the cloud TM Parallels Dynamic Infrastructure AndOpenStack.
FermiCloud Update EGI Technical Forum 2012 Keith Chadwick Grid & Cloud Computing Department Fermilab Work supported by the U.S. Department of Energy under.
An Approach to Secure Cloud Computing Architectures By Y. Serge Joseph FAU security Group February 24th, 2011.
FermiCloud Dynamic Resource Provisioning Condor Week 2012 Steven Timm Fermilab Grid & Cloud Computing Dept. For FermiCloud team:
Idle virtual machine detection in FermiCloud Giovanni Franzini September 21, 2012 Scientific Computing Division Grid and Cloud Computing Department.
Implementing Failover Clustering with Hyper-V
VMs at a Tier-1 site EGEE’09, Sander Klous, Nikhef.
FermiCloud K. Chadwick, T. Hesselroth, F. Lowe, S. Timm, D. R. Yocum Grid And Cloud Computing Department Fermilab ISGC2011 Work supported by the U.S. Department.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Integration and Commissioning of a Prototype Federated Cloud for Scientific Workflows Steven C. Timm, Gabriele Garzoglio, Fermilab Seo-Young Noh, Haeng-Jing.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 7 2/23/2015.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Virtualization and Cloud Computing in Support of Science at Fermilab Keith Chadwick Grid & Cloud Computing Department Fermilab Work supported by the U.S.
+ CS 325: CS Hardware and Software Organization and Architecture Cloud Architectures.
Virtualization at Fermilab Keith Chadwick Fermilab Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Grid Developers’ use of FermiCloud (to be integrated with master slides)
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
608D CloudStack 3.0 Omer Palo Readiness Specialist, WW Tech Support Readiness May 8, 2012.
The High-Availability FermiCloud Infrastructure-as-a-Service Facility HEPIX Meeting Beijing Oct Steven C. Timm FermiCloud Project Leader Grid & Cloud.
Virtualization within FermiGrid Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Virtualization within FermiGrid Keith Chadwick Fermilab Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
The Grid & Cloud Computing Department at Fermilab and the KISTI Collaboration Meeting with KISTI Nov 1, 2011 Gabriele Garzoglio Grid & Cloud Computing.
Virtualization for the LHCb Online system CHEP Taipei Dedicato a Zio Renato Enrico Bonaccorsi, (CERN)
Fermilab Site Report Spring 2012 HEPiX Keith Chadwick Fermilab Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Visual Studio Windows Azure Portal Rest APIs / PS Cmdlets US-North Central Region FC TOR PDU Servers TOR PDU Servers TOR PDU Servers TOR PDU.
Windows Azure Virtual Machines Anton Boyko. A Continuous Offering From Private to Public Cloud.
GLIDEINWMS - PARAG MHASHILKAR Department Meeting, August 07, 2013.
Efficient Live Checkpointing Mechanisms for computation and memory-intensive VMs in a data center Kasidit Chanchio Vasabilab Dept of Computer Science,
FermiCloud: Enabling Scientific Workflows with Federation and Interoperability Steven C. Timm FermiCloud Project Lead Grid & Cloud Computing Department.
Full and Para Virtualization
Ian Gable HEPiX Spring 2009, Umeå 1 VM CPU Benchmarking the HEPiX Way Manfred Alef, Ian Gable FZK Karlsruhe University of Victoria May 28, 2009.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Eileen Berman. Condor in the Fermilab Grid FacilitiesApril 30, 2008  Fermi National Accelerator Laboratory is a high energy physics laboratory outside.
An Introduction to Campus Grids 19-Apr-2010 Keith Chadwick & Steve Timm.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
FermiGrid Keith Chadwick. Overall Deployment Summary 5 Racks in FCC:  3 Dell Racks on FCC1 –Can be relocated to FCC2 in FY2009. –Would prefer a location.
FermiCloud Project Overview and Progress Keith Chadwick Grid & Cloud Computing Department Head Fermilab Work supported by the U.S. Department of Energy.
Auxiliary services Web page Secrets repository RSV Nagios Monitoring Ganglia NIS server Syslog Forward FermiCloud: A private cloud to support Fermilab.
Fermilab / FermiGrid / FermiCloud Security Update Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359 Keith Chadwick Grid.
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
The FermiCloud Infrastructure- as-a-Service Facility Steven C. Timm FermiCloud Project Leader Grid & Cloud Computing Department Fermilab Work supported.
FermiCloud Status Report Fall 2010 Keith Chadwick Grid & Cloud Computing Department Head Fermilab Work supported by the U.S. Department.
Authentication, Authorization, and Contextualization in FermiCloud S. Timm, D. Yocum, F. Lowe, K. Chadwick, G. Garzoglio, D. Strain, D. Dykstra, T. Hesselroth.
© 2012 Eucalyptus Systems, Inc. Cloud Computing Introduction Eucalyptus Education Services 2.
Instituto de Biocomputación y Física de Sistemas Complejos Cloud resources and BIFI activities in JRA2 Reunión JRU Española.
FermiCloud Project: Integration, Development, Futures Gabriele Garzoglio Associate Head Grid & Cloud Computing Department Fermilab Work supported by the.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
April 18, 2006FermiGrid Project1 FermiGrid Project Status April 18, 2006 Keith Chadwick.
Hao Wu, Shangping Ren, Gabriele Garzoglio, Steven Timm, Gerard Bernabeu, Hyun Woo Kim, Keith Chadwick, Seo-Young Noh A Reference Model for Virtual Machine.
FermiCloud Update ISGC 2013 Keith Chadwick Grid & Cloud Computing Department Fermilab Work supported by the U.S. Department of Energy under contract No.
GPCF* Update Present status as a series of questions / answers related to decisions made / yet to be made * General Physics Computing Facility (GPCF) is.
Parag Mhashilkar (Fermi National Accelerator Laboratory)
FermiCloud Review Response to Questions Keith Chadwick Steve Timm Gabriele Garzoglio Work supported by the U.S. Department of Energy under contract No.
FermiGrid The Fermilab Campus Grid 28-Oct-2010 Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Virtualization within FermiGrid Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
July 18, 2011S. Timm FermiCloud Enabling Scientific Computing with Integrated Private Cloud Infrastructures Steven Timm.
High Throughput & Resilient Fabric Deployments on FermiCloud Steven C. Timm Grid & Cloud Services Department Fermilab Work supported by the U.S. Department.
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
StratusLab Final Periodic Review
StratusLab Final Periodic Review
f f FermiGrid – Site AuthoriZation (SAZ) Service
Enrico Bonaccorsi, (CERN) Loic Brarda, (CERN) Gary Moine, (CERN)
FermiGrid and FermiCloud Updates
Versatile HPC: Comet Virtual Clusters for the Long Tail of Science SC17 Denver Colorado Comet Virtualization Team: Trevor Cooper, Dmitry Mishin, Christopher.
Presentation transcript:

FermiCloud Infrastructure as a Service (IaaS) Cloud Computing In Support of the Fermilab Scientific Program OSG All Hands Meeting 2012 Steven Timm Fermilab Grid & Cloud Computing Dept. For FermiCloud team: K. Chadwick, D. Yocum, F. Lowe, G. Garzoglio, T. Levshina, P. Mhashilkar, H. Kim Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359

FermiCloud Introduction As part of the FY2010 activities, the (then) Grid Facilities Department established a project to implement an initial “FermiCloud” capability. GOAL: Deliver production-capable Infrastucture-as-a-service to support Fermilab Scientific Program Reuse what we learned from Grid High Availability, Authentication/Authorization, Virtualization FermiCloud Phase I—Completed Nov. 2010: –Specify, acquire and deploy the FermiCloud hardware, –Establish initial FermiCloud requirements and select the “best” open source cloud computing framework that best met these requirements (OpenNebula). –Deploy capabilities to meet the needs of the stakeholders (JDEM analysis development, Grid Developers and Integration test stands, Storage/dCache Developers, LQCD testbed). –Replaced six old racks of integration/test nodes with one rack. 20-Mar-2012FermiCloud

FermiCloud – Current Activities FermiCloud Phase II: –Implement x509 based authentication (patches contributed back to OpenNebula project and are generally available in OpenNebula V3.2), perform secure contexualization of virtual machines at launch. –Implement monitoring and accounting, –Target “small” low-cpu-load servers such as Grid gatekeepers, forwarding nodes, small databases, monitoring, etc. –Begin the hardware deployment of a distributed SAN, FermiCloud Phase III: –Select and deploy a true multi-user filesystem on top of a distributed & replicated SAN, –Deploy 24x7 production services –Live migration becomes important for this phase. 20-Mar-20122FermiCloud

FermiCloud – Hardware Specifications Currently 23 systems split across FCC-3 and GCC-B: 2 x 2.67 GHz Intel “Westmere” 4 core CPU –Total 8 physical cores, potentially 16 cores with Hyper Threading (HT), 24 GBytes of memory (we are considering an upgrade to 48), 2 x 1GBit Ethernet interface (1 public, 1 private), 8 port Raid Controller, 2 x 300 GBytes of high speed local disk (15K RPM SAS), 6 x 2 TBytes = 12 TB raw of RAID SATA disk = ~10 TB formatted, InfiniBand SysConnect II DDR HBA, Brocade FibreChannel HBA (added in Fall 2011), 2U SuperMicro chassis with redundant power supplies 20-Mar-20123FermiCloud

FermiCloud—Software Stack Current production –Scientific Linux 5.7 host, SLF5 and SLF6 guest –KVM hypervisor (Xen available on request). –OpenNebula 2.0 with command-line launch –Virtual machines distributed via SCP Coming soon –Scientific Linux 6.1, SLF5 and SLF6 guests –KVM hypervisor –OpenNebula 3.2 with X.509 authentication Command line, SunStone Web UI, EC2 emulation, OCCI interface, Condor-G –Persistent virtual machines stored on SAN (GFS). All Open Source 20-Mar-2012FermiCloud

FermiCloud Typical VM Specifications Unit: –1 Virtual CPU [2.67 GHz “core” with Hyper Threading (HT)], –2 GBytes of memory, –10-20 GBytes of of SAN based “VM Image” storage, –Additional ~20-50 GBytes of “transient” local storage. Additional CPU “cores”, memory and storage are available for “purchase”: –Based on the (Draft) FermiCloud Economic Model, –Raw VM costs are competitive with Amazon EC2, –FermiCloud VMs can be custom configured per “client”, –Access to Fermilab science datasets is much better than Amazon EC2. 20-Mar-20125FermiCloud

FermiCloud – Monitoring 20-Mar-20126FermiCloud Temporary FermiCloud Usage Monitor: – –Data collection dynamically “ping-pongs” across systems deployed in FCC and GCC to offer redundancy, –See plot on next page. FermiCloud Redundant Ganglia Servers: – – Preliminary RSV based monitoring pilot: –

FermiCloud Capacity # of Units Nominal (1 physical core = 1 VM) % over subscription % over subscription (1 HT core = 1 VM) % over subscription552 Note – FermiGrid Production Services are operated at 100% to 200% “oversubscription” FermiCloud Target VM states as reported by “virsh list” 20-Mar-20127FermiCloud

Description of Virtual Machine States Reported by “virsh list” Command StateDescription runningThe domain is currently running on a CPU. Note – KVM based VMs show up in this state even when they are “idle”. idleThe domain is idle, and not running or runnable. This can be caused because the domain is waiting on I/O (a traditional wait state) or has gone to sleep because there was nothing else for it to do. Note – Xen based VMs typically show up in this state even when they are “running”. pausedThe domain has been paused, usually occurring through the administrator running virsh suspend. When in a paused state the domain will still consume allocated resources like memory, but will not be eligible for scheduling by the hypervisor. shutdownThe domain is in the process of shutting down, i.e. the guest operating system has been notified and should be in the process of stopping its operations gracefully. shut offThe domain has been shut down. When in a shut off state the domain does not consume resources. crashedThe domain has crashed. Usually this state can only occur if the domain has been configured not to restart on crash. dyingThe domain is in process of dying, but hasn't completely shutdown or crashed. 20-Mar-20128FermiCloud

FermiCloud – Monitoring Requirements & Goals Need to monitor to assure that: –All hardware is available (both in FCC3 and GCC-B), –All necessary and required OpenNebula services are running, –All Virtual Machine hosts are healthy –All “24x7” & “9x5” virtual machines (VMs) are running, –If a building is “lost”, then automatically relaunch “24x7” VMs on surviving infrastructure, then relaunch “9x5” VMs if there is sufficient remaining capacity, –Perform notification (via Service-Now) when exceptions are detected. We plan to replace the temporary monitoring with an infrastructure based on either Nagios or Zabbix during CY2012. –Possibly utilizing the OSG Resource Service Validation (RSV) scripts. –This work will likely be performed in collaboration with KISTI. Goal is to identify really idle virtual machines and suspend if necessary. –Can’t trust hypervisor VM state output on this—Need rule-based definition –In times of resource need, we want the ability to suspend or “shelve” the really idle VMs in order to free up resources for higher priority usage. –Shelving of “9x5” and “opportunistic” VMs will allow us to use FermiCloud resources for Grid worker node VMs during nights and weekends (this is part of the draft economic model). 20-Mar-20129FermiCloud

Currently have two “probes” based on the Gratia accounting framework used by Fermilab and the Open Science Grid Standard Process Accounting (“psacct”) Probe: –Installed and runs within the virtual machine image, –Reports to standard gratia-fermi-psacct.fnal.gov. Open Nebula Gratia Accounting Probe: –Runs on the OpenNebula management node and collects data from ONE logs, emits standard Gratia usage records, –Reports to the “virtualization” Gratia collector, –The “virtualization” Gratia collector runs existing standard Gratia collector software (no development was required), –The development of the Open Nebula Gratia accounting probe was performed by Tanya Levshina and Parag Mhashilkar. Additional Gratia accounting probes could be developed: –Commercial – OracleVM, VMware, --- –Open Source – Nimbus, Eucalyptus, OpenStack, … FermiCloud - Accounting 20-Mar FermiCloud

Open Nebula Gratia Accounting Probe Fermilab Gratia Collector/Reporter MySQL Database Collector gratia-fermi-transfergr10x3 gratia-fermi- itb gr10x4 Collector gr10x2gratia-fermi- qcd Collector gratia-fermi-psacctgr10x1 Collector gr10x0gratia-fermi- osg MySQL Database Collector gratia-fermi-transfergr11x3 gratia-fermi- itb gr11x4 Collector gr11x2gratia-fermi- qcd Collector gratia-fermi-psacctgr11x1 Reporter gr11x0gratia-fermi- osg ONE DB onevm_query ONE API gratia_onevm Gratia API onevm data Probe Config Probe Config osg user map osg user map New Code Standard Usage Records 20-Mar FermiCloud

FermiCloud – Gratia Accounting Reports Here are the preliminary results of “replaying” the previous year of the OpenNebula “OneVM” data into the new accounting probe: 20-Mar-2012FermiCloud

vCluster Deployable on demand virtual cluster using hybrid cloud computing resources. –Head nodes launched on virtual machines within the FermiCloud private cloud. –Worker nodes launched on virtual machines within the Amazon EC2 public cloud and on FermiCloud –Work done by Dr. Seo-Young Noh, KISTI-GSDC visitor at Fermilab, summer 2011 –Look ahead at Condor queue to see what kind of virtual machines are needed and submit as needed. 20-Mar FermiCloud

MPI on FermiCloud (Note 1) Configuration#Host Systems#VM/host#CPU Total Physical CPU HPL Benchmark (Gflops) Bare Metal without pinning Bare Metal with pinning (Note 2) VM without pinning (Notes 2,3) 281 vCPU168.2 VM with pinning (Notes 2,3) 281 vCPU VM+SRIOV with pinning (Notes 2,4) 272 vCPU Notes:(1) Work performed by Dr. Hyunwoo Kim of KISTI in collaboration with Dr. Steven Timm of Fermilab. (2) Process/Virtual Machine “pinned” to CPU and associated NUMA memory via use of numactl. (3) Software Bridged Virtual Network using IP over IB (seen by Virtual Machine as a virtual Ethernet). (4) SRIOV driver presents native InfiniBand to virtual machine(s), 2 nd virtual CPU is required to start SRIOV, but is only a virtual CPU, not an actual physical CPU. 20-Mar FermiCloud

FermiCloud Facilities FCC GCC The FCC and GCC buildings are separated by approximately 1 mile (1.6 km). FCC has UPS and Generator. GCC has UPS. 20-Mar FermiCloud

FermiCloud – Network & SAN “Today” Private Ethernet over dedicated fiber Fibre Channel 20-Mar FermiCloud Nexus 7010 Nexus 7010 Nexus 7010 Nexus 7010 Nexus 7010 Nexus 7010 FCC-2GCC-A FCC-3GCC-B Nexus 7010 Nexus 7010 fcl316 To fcl323 fcl316 To fcl323 FCC-3 Brocade Satabeast Brocade fcl001 To fcl015 fcl001 To fcl015 GCC-B Brocade Satabeast Brocade

FermiCloud – Network & SAN (Possible Future – FY2013/2014) 20-Mar-2012FermiCloud Fibre Channel FCC-2 Nexus 7010 Nexus 7010 Nexus 7010 Nexus 7010 GCC-A Nexus 7010 Nexus 7010 FCC-3 Nexus 7010 Nexus 7010 GCC-B fcl316 To fcl330 fcl316 To fcl330 FCC-3 Brocade Satabeast Brocade fcl0yy To fcl0zz fcl0yy To fcl0zz FCC-2 Brocade fcl0xx To fcl0yy fcl0xx To fcl0yy GCC-A Brocade fcl001 To fcl015 fcl001 To fcl015 GCC-B Brocade Satabeast Brocade

FermiCloud – Support for Science (direct and indirect) Grid middleware developers and integrators as well as CDF, D0 and the Cosmic and Intensity Frontiers. Detailed analysis of filesystem performance for Grid applications –(Lustre vs. Hadoop vs. OrangeFS vs. Bluearc). Tuning and stress tests of GUMS-XACML and SAZ-XACML dCache NFS v4.1 interface Fabric testbeds—IPv6 for Grid and Cloud applications OSG Software Team integration and testing of OSG 3.0 RPMS Scientific control and communication system tests for JDEM and NOVA Production services have been deployed on FermiCloud, including: –Specialized GridFTP virtual machines for the Intensity Frontier experiments, –NFS Administration virtual machines to allow Intensity Frontier experiments to perform file ownership management of their BlueArc volumes. –SAMGrid forwarding nodes to send D0 jobs to WLCG sites via GlideinWMS 20-Mar FermiCloud

Recent FermiCloud Uses Application/ProjectStakeholder JDEM Demo Data Processing SystemLSST OSG Software Team build and testOSG GEANT4 validation serverCET GlobusOnline test work (CEDPS)OSG & CEDPS CVMFS testingIF Test GUMS and SAZ serversFermiGrid Extenci Lustre-over-WANOSG Vcluster demo (grid cluster in the cloud)GCC & KISTI Expt-specific gridftp servers (MINOS, NOVA) IF SAMGrid forwarding nodesD0 dCache NFS4.1 testingREX & DMS Scientist surveyDirectorate MCASGCC OSG-XSEDE interoperabilityOSG, XSEDE 20-Mar-2012FermiCloud

FermiCloud Conclusions The existing (temporary) FermiCloud usage monitoring shows that the peak FermiCloud usage is ~100% of the nominal capacity and ~50% of the expected oversubscription capacity. The FermiCloud collaboration with KISTI has leveraged the resources and expertise of both institutions to achieve significant benefits. FermiCloud has plans to implement both monitoring and accounting by extension of existing tools in CY2012. Using SRIOV drivers on FermiCloud virtual machines, MPI performance has been demonstrated to be >96% of the native “bare metal” performance. –Note that this HPL benchmark performance measurement was accomplished using 2 fewer physical CPUs than the corresponding “bare metal” performance measurement! FermiCloud personnel are working to implement a SAN storage deployment that will offer a true multi-user filesystem on top of a distributed & replicated SAN. Science is directly and indirectly benefiting from FermiCloud. 20-Mar FermiCloud

Summary FermiCloud operates at the forefront of delivering cloud computing capabilities to support physics research: –By starting small, developing a list of requirements, building on existing Grid knowledge and infrastructure to address those requirements, FermiCloud has managed to deliver an Infrastructure as a Service cloud computing capability that supports science at Fermilab. –The Open Science Grid software team is using FermiCloud resources to support their RPM “refactoring”. Thanks to: –Open Source collaborators—OpenNebula and Scientific Linux –Fermilab Networking and Facilities team –Collaborators from KISTI-GSDC We welcome collaborators and visitors! Keep an eye on Fermilab’s employment website for opening (hopefully) coming soon. 20-Mar FermiCloud

Extra Slides

FermiCloud vs. Magellan FermiCloudMagellan Bottom up requirements and design.Top down requirements and mission. Funded out of existing budget ($230K+$128K).Funded via ARRA ($32M). Multi phase project, with each phase building on knowledge gained during previous phases. Fixed term project without ongoing funding. Evaluated available open source cloud computing frameworks (Eucalyptus, Nimbus, OpenNebula) against requirements and selected OpenNebula. We plan to “circle back” and evaluate OpenStack this year. Spent a lot of time trying to get the open source version of Eucalyptus to work at scale, eventually switched to a combination of Nimbus and OpenStack late in the project. Approached cloud computing from a Grid and high throughput computing (HTC) perspective. Approached cloud computing from a high performance computing (HPC) perspective. Significant prior experience delivering production Grid services via open source virtualization (Xen and KVM). Unclear. Have SRIOV drivers for InfiniBand.Did not have SRIOV drivers for InfiniBand before the end of the project. Actively sought collaboration (OpenNebula, KISTI).Project was sited at NERSC and Argonne. 20-Mar-2012FermiCloud